var/home/core/zuul-output/0000755000175000017500000000000015110265675014535 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015110310653015464 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000004064144115110310643017673 0ustar rootrootNov 22 07:56:23 crc systemd[1]: Starting Kubernetes Kubelet... Nov 22 07:56:23 crc restorecon[4753]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 22 07:56:23 crc restorecon[4753]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:24 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 22 07:56:25 crc restorecon[4753]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 22 07:56:25 crc restorecon[4753]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Nov 22 07:56:27 crc kubenswrapper[4789]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 22 07:56:27 crc kubenswrapper[4789]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Nov 22 07:56:27 crc kubenswrapper[4789]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 22 07:56:27 crc kubenswrapper[4789]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 22 07:56:27 crc kubenswrapper[4789]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Nov 22 07:56:27 crc kubenswrapper[4789]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.155557 4789 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.158866 4789 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.158882 4789 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.158887 4789 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.158892 4789 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.158896 4789 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.158901 4789 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.158905 4789 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.158909 4789 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.158913 4789 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.158918 4789 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.158924 4789 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.158929 4789 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.158935 4789 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.158940 4789 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.158945 4789 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.158950 4789 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.158955 4789 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.158959 4789 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.158964 4789 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.158968 4789 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.158973 4789 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.158977 4789 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.158981 4789 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.158994 4789 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.158998 4789 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.159004 4789 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.159010 4789 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.159014 4789 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.159019 4789 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.159024 4789 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.159028 4789 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.159033 4789 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.159037 4789 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.159042 4789 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.159046 4789 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.159051 4789 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.159056 4789 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.159061 4789 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.159066 4789 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.159070 4789 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.159075 4789 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.159081 4789 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.159086 4789 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.159091 4789 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.159095 4789 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.159099 4789 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.159104 4789 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.159108 4789 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.159112 4789 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.159116 4789 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.159120 4789 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.159126 4789 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.159131 4789 feature_gate.go:330] unrecognized feature gate: Example Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.159135 4789 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.159139 4789 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.159143 4789 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.159147 4789 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.159151 4789 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.159156 4789 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.159161 4789 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.159165 4789 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.159169 4789 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.159173 4789 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.159176 4789 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.159181 4789 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.159185 4789 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.159189 4789 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.159193 4789 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.159197 4789 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.159200 4789 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.159204 4789 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160452 4789 flags.go:64] FLAG: --address="0.0.0.0" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160466 4789 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160474 4789 flags.go:64] FLAG: --anonymous-auth="true" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160479 4789 flags.go:64] FLAG: --application-metrics-count-limit="100" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160485 4789 flags.go:64] FLAG: --authentication-token-webhook="false" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160489 4789 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160496 4789 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160502 4789 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160506 4789 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160510 4789 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160514 4789 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160519 4789 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160523 4789 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160527 4789 flags.go:64] FLAG: --cgroup-root="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160531 4789 flags.go:64] FLAG: --cgroups-per-qos="true" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160535 4789 flags.go:64] FLAG: --client-ca-file="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160539 4789 flags.go:64] FLAG: --cloud-config="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160543 4789 flags.go:64] FLAG: --cloud-provider="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160547 4789 flags.go:64] FLAG: --cluster-dns="[]" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160553 4789 flags.go:64] FLAG: --cluster-domain="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160557 4789 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160561 4789 flags.go:64] FLAG: --config-dir="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160565 4789 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160570 4789 flags.go:64] FLAG: --container-log-max-files="5" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160576 4789 flags.go:64] FLAG: --container-log-max-size="10Mi" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160580 4789 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160585 4789 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160589 4789 flags.go:64] FLAG: --containerd-namespace="k8s.io" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160594 4789 flags.go:64] FLAG: --contention-profiling="false" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160598 4789 flags.go:64] FLAG: --cpu-cfs-quota="true" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160602 4789 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160607 4789 flags.go:64] FLAG: --cpu-manager-policy="none" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160611 4789 flags.go:64] FLAG: --cpu-manager-policy-options="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160617 4789 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160621 4789 flags.go:64] FLAG: --enable-controller-attach-detach="true" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160625 4789 flags.go:64] FLAG: --enable-debugging-handlers="true" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160629 4789 flags.go:64] FLAG: --enable-load-reader="false" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160634 4789 flags.go:64] FLAG: --enable-server="true" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160638 4789 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160643 4789 flags.go:64] FLAG: --event-burst="100" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160648 4789 flags.go:64] FLAG: --event-qps="50" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160652 4789 flags.go:64] FLAG: --event-storage-age-limit="default=0" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160656 4789 flags.go:64] FLAG: --event-storage-event-limit="default=0" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160661 4789 flags.go:64] FLAG: --eviction-hard="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160665 4789 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160670 4789 flags.go:64] FLAG: --eviction-minimum-reclaim="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160674 4789 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160678 4789 flags.go:64] FLAG: --eviction-soft="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160682 4789 flags.go:64] FLAG: --eviction-soft-grace-period="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160686 4789 flags.go:64] FLAG: --exit-on-lock-contention="false" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160690 4789 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160694 4789 flags.go:64] FLAG: --experimental-mounter-path="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160699 4789 flags.go:64] FLAG: --fail-cgroupv1="false" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160703 4789 flags.go:64] FLAG: --fail-swap-on="true" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160707 4789 flags.go:64] FLAG: --feature-gates="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160713 4789 flags.go:64] FLAG: --file-check-frequency="20s" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160718 4789 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160722 4789 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160726 4789 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160730 4789 flags.go:64] FLAG: --healthz-port="10248" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160734 4789 flags.go:64] FLAG: --help="false" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160738 4789 flags.go:64] FLAG: --hostname-override="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160743 4789 flags.go:64] FLAG: --housekeeping-interval="10s" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160760 4789 flags.go:64] FLAG: --http-check-frequency="20s" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160766 4789 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160770 4789 flags.go:64] FLAG: --image-credential-provider-config="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160774 4789 flags.go:64] FLAG: --image-gc-high-threshold="85" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160778 4789 flags.go:64] FLAG: --image-gc-low-threshold="80" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160782 4789 flags.go:64] FLAG: --image-service-endpoint="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160786 4789 flags.go:64] FLAG: --kernel-memcg-notification="false" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160789 4789 flags.go:64] FLAG: --kube-api-burst="100" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160794 4789 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160798 4789 flags.go:64] FLAG: --kube-api-qps="50" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160802 4789 flags.go:64] FLAG: --kube-reserved="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160806 4789 flags.go:64] FLAG: --kube-reserved-cgroup="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160810 4789 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160815 4789 flags.go:64] FLAG: --kubelet-cgroups="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160819 4789 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160823 4789 flags.go:64] FLAG: --lock-file="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160827 4789 flags.go:64] FLAG: --log-cadvisor-usage="false" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160831 4789 flags.go:64] FLAG: --log-flush-frequency="5s" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160835 4789 flags.go:64] FLAG: --log-json-info-buffer-size="0" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160841 4789 flags.go:64] FLAG: --log-json-split-stream="false" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160845 4789 flags.go:64] FLAG: --log-text-info-buffer-size="0" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160850 4789 flags.go:64] FLAG: --log-text-split-stream="false" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160854 4789 flags.go:64] FLAG: --logging-format="text" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160859 4789 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160864 4789 flags.go:64] FLAG: --make-iptables-util-chains="true" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160868 4789 flags.go:64] FLAG: --manifest-url="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160873 4789 flags.go:64] FLAG: --manifest-url-header="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160878 4789 flags.go:64] FLAG: --max-housekeeping-interval="15s" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160883 4789 flags.go:64] FLAG: --max-open-files="1000000" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160888 4789 flags.go:64] FLAG: --max-pods="110" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160893 4789 flags.go:64] FLAG: --maximum-dead-containers="-1" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160897 4789 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160901 4789 flags.go:64] FLAG: --memory-manager-policy="None" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160905 4789 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160910 4789 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160914 4789 flags.go:64] FLAG: --node-ip="192.168.126.11" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160918 4789 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160929 4789 flags.go:64] FLAG: --node-status-max-images="50" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160933 4789 flags.go:64] FLAG: --node-status-update-frequency="10s" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160937 4789 flags.go:64] FLAG: --oom-score-adj="-999" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160940 4789 flags.go:64] FLAG: --pod-cidr="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160944 4789 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160952 4789 flags.go:64] FLAG: --pod-manifest-path="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160956 4789 flags.go:64] FLAG: --pod-max-pids="-1" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160960 4789 flags.go:64] FLAG: --pods-per-core="0" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160964 4789 flags.go:64] FLAG: --port="10250" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160969 4789 flags.go:64] FLAG: --protect-kernel-defaults="false" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160973 4789 flags.go:64] FLAG: --provider-id="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160977 4789 flags.go:64] FLAG: --qos-reserved="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160983 4789 flags.go:64] FLAG: --read-only-port="10255" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160988 4789 flags.go:64] FLAG: --register-node="true" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160992 4789 flags.go:64] FLAG: --register-schedulable="true" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.160996 4789 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.161004 4789 flags.go:64] FLAG: --registry-burst="10" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.161008 4789 flags.go:64] FLAG: --registry-qps="5" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.161012 4789 flags.go:64] FLAG: --reserved-cpus="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.161016 4789 flags.go:64] FLAG: --reserved-memory="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.161022 4789 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.161026 4789 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.161031 4789 flags.go:64] FLAG: --rotate-certificates="false" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.161035 4789 flags.go:64] FLAG: --rotate-server-certificates="false" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.161039 4789 flags.go:64] FLAG: --runonce="false" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.161044 4789 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.161049 4789 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.161053 4789 flags.go:64] FLAG: --seccomp-default="false" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.161057 4789 flags.go:64] FLAG: --serialize-image-pulls="true" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.161061 4789 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.161066 4789 flags.go:64] FLAG: --storage-driver-db="cadvisor" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.161070 4789 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.161074 4789 flags.go:64] FLAG: --storage-driver-password="root" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.161078 4789 flags.go:64] FLAG: --storage-driver-secure="false" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.161082 4789 flags.go:64] FLAG: --storage-driver-table="stats" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.161086 4789 flags.go:64] FLAG: --storage-driver-user="root" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.161090 4789 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.161094 4789 flags.go:64] FLAG: --sync-frequency="1m0s" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.161098 4789 flags.go:64] FLAG: --system-cgroups="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.161103 4789 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.161109 4789 flags.go:64] FLAG: --system-reserved-cgroup="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.161112 4789 flags.go:64] FLAG: --tls-cert-file="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.161117 4789 flags.go:64] FLAG: --tls-cipher-suites="[]" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.161123 4789 flags.go:64] FLAG: --tls-min-version="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.161127 4789 flags.go:64] FLAG: --tls-private-key-file="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.161131 4789 flags.go:64] FLAG: --topology-manager-policy="none" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.161135 4789 flags.go:64] FLAG: --topology-manager-policy-options="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.161140 4789 flags.go:64] FLAG: --topology-manager-scope="container" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.161145 4789 flags.go:64] FLAG: --v="2" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.161152 4789 flags.go:64] FLAG: --version="false" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.161157 4789 flags.go:64] FLAG: --vmodule="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.161162 4789 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.161167 4789 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.161267 4789 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.161271 4789 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.161275 4789 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.161279 4789 feature_gate.go:330] unrecognized feature gate: Example Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.161282 4789 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.161286 4789 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.161290 4789 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.161294 4789 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.161298 4789 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.161302 4789 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.161306 4789 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.161309 4789 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.161313 4789 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.161317 4789 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.161320 4789 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.161324 4789 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.161327 4789 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.161331 4789 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.161335 4789 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.161339 4789 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.161343 4789 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.161348 4789 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.161353 4789 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.161357 4789 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.161361 4789 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.161364 4789 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.161370 4789 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.161375 4789 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.161380 4789 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.161384 4789 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.161388 4789 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.161393 4789 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.161397 4789 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.161401 4789 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.161404 4789 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.161409 4789 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.161413 4789 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.161417 4789 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.161421 4789 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.161425 4789 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.161428 4789 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.161432 4789 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.161435 4789 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.161439 4789 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.161442 4789 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.161445 4789 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.161449 4789 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.161452 4789 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.161456 4789 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.161459 4789 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.161462 4789 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.161466 4789 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.161469 4789 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.161473 4789 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.161477 4789 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.161480 4789 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.161483 4789 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.161487 4789 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.161491 4789 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.161494 4789 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.161501 4789 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.161504 4789 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.161508 4789 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.161511 4789 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.161515 4789 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.161519 4789 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.161523 4789 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.161528 4789 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.161532 4789 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.161536 4789 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.161540 4789 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.161553 4789 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.225407 4789 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.225473 4789 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.225612 4789 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.225629 4789 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.225641 4789 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.225652 4789 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.225661 4789 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.225670 4789 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.225679 4789 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.225689 4789 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.225702 4789 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.225712 4789 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.225723 4789 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.225733 4789 feature_gate.go:330] unrecognized feature gate: Example Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.225791 4789 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.225810 4789 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.225823 4789 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.225836 4789 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.225845 4789 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.225854 4789 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.225862 4789 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.225871 4789 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.225879 4789 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.225888 4789 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.225897 4789 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.225905 4789 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.225914 4789 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.225922 4789 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.225930 4789 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.225938 4789 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.225947 4789 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.225958 4789 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.225968 4789 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.225977 4789 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.225986 4789 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.225997 4789 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.226010 4789 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.226019 4789 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.226027 4789 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.226036 4789 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.226045 4789 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.226054 4789 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.226062 4789 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.226071 4789 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.226079 4789 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.226088 4789 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.226096 4789 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.226104 4789 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.226113 4789 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.226121 4789 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.226129 4789 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.226137 4789 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.226146 4789 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.226154 4789 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.226162 4789 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.226171 4789 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.226180 4789 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.226188 4789 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.226196 4789 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.226205 4789 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.226213 4789 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.226221 4789 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.226230 4789 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.226238 4789 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.226246 4789 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.226255 4789 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.226263 4789 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.226272 4789 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.226280 4789 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.226289 4789 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.226297 4789 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.226308 4789 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.226318 4789 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.226333 4789 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.226568 4789 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.226582 4789 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.226592 4789 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.226602 4789 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.226612 4789 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.226621 4789 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.226629 4789 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.226637 4789 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.226646 4789 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.226656 4789 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.226664 4789 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.226672 4789 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.226680 4789 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.226688 4789 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.226697 4789 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.226705 4789 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.226713 4789 feature_gate.go:330] unrecognized feature gate: Example Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.226722 4789 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.226730 4789 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.226739 4789 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.226776 4789 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.226784 4789 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.226793 4789 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.226802 4789 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.226811 4789 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.226820 4789 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.226830 4789 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.226839 4789 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.226847 4789 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.226856 4789 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.226864 4789 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.226872 4789 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.226880 4789 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.226889 4789 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.226898 4789 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.226907 4789 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.226915 4789 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.226924 4789 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.226933 4789 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.226941 4789 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.226950 4789 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.226959 4789 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.226970 4789 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.226980 4789 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.226989 4789 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.226999 4789 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.227008 4789 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.227016 4789 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.227025 4789 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.227033 4789 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.227042 4789 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.227050 4789 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.227059 4789 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.227067 4789 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.227078 4789 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.227087 4789 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.227096 4789 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.227106 4789 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.227115 4789 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.227124 4789 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.227133 4789 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.227142 4789 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.227150 4789 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.227161 4789 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.227172 4789 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.227181 4789 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.227191 4789 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.227199 4789 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.227210 4789 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.227242 4789 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.227252 4789 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.227267 4789 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.228294 4789 server.go:940] "Client rotation is on, will bootstrap in background" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.237928 4789 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.238051 4789 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.240962 4789 server.go:997] "Starting client certificate rotation" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.241014 4789 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.242699 4789 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2026-01-14 00:25:15.521983111 +0000 UTC Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.242827 4789 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 1264h28m48.2791618s for next certificate rotation Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.327977 4789 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.331972 4789 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.427051 4789 log.go:25] "Validated CRI v1 runtime API" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.561194 4789 log.go:25] "Validated CRI v1 image API" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.563522 4789 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.579447 4789 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-11-22-07-51-26-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.580005 4789 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:41 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.608194 4789 manager.go:217] Machine: {Timestamp:2025-11-22 07:56:27.60540846 +0000 UTC m=+1.839808813 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2799998 MemoryCapacity:33654120448 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:fd8856dd-ccc1-49f3-a39b-e4a6f9389fc3 BootID:06b9c1b2-d063-4af7-880e-0654164bb896 Filesystems:[{Device:/run/user/1000 DeviceMajor:0 DeviceMinor:41 Capacity:3365408768 Type:vfs Inodes:821633 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108169 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827060224 Type:vfs Inodes:4108169 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827060224 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:e0:1e:3a Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:e0:1e:3a Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:f4:b1:34 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:51:b1:27 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:ea:34:34 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:1b:02:e9 Speed:-1 Mtu:1496} {Name:ens7.23 MacAddress:52:54:00:87:96:e4 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:72:aa:df:44:90:ce Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:92:ff:3f:b0:23:70 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654120448 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.608621 4789 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.608953 4789 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.616232 4789 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.617500 4789 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.617563 4789 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.617877 4789 topology_manager.go:138] "Creating topology manager with none policy" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.617896 4789 container_manager_linux.go:303] "Creating device plugin manager" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.618609 4789 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.618653 4789 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.619067 4789 state_mem.go:36] "Initialized new in-memory state store" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.619198 4789 server.go:1245] "Using root directory" path="/var/lib/kubelet" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.715453 4789 kubelet.go:418] "Attempting to sync node with API server" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.715516 4789 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.715547 4789 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.715569 4789 kubelet.go:324] "Adding apiserver pod source" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.715588 4789 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.723681 4789 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.723925 4789 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.129.56.56:6443: connect: connection refused Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.723957 4789 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.129.56.56:6443: connect: connection refused Nov 22 07:56:27 crc kubenswrapper[4789]: E1122 07:56:27.724064 4789 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.129.56.56:6443: connect: connection refused" logger="UnhandledError" Nov 22 07:56:27 crc kubenswrapper[4789]: E1122 07:56:27.724064 4789 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.129.56.56:6443: connect: connection refused" logger="UnhandledError" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.724940 4789 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.726821 4789 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.728644 4789 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.728685 4789 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.728700 4789 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.728714 4789 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.728734 4789 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.728747 4789 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.728809 4789 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.728830 4789 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.728845 4789 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.728858 4789 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.728885 4789 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.728899 4789 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.729787 4789 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.730453 4789 server.go:1280] "Started kubelet" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.730787 4789 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.730857 4789 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.731472 4789 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.56:6443: connect: connection refused Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.731706 4789 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Nov 22 07:56:27 crc systemd[1]: Started Kubernetes Kubelet. Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.740790 4789 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.740853 4789 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.741115 4789 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-13 04:46:24.977545143 +0000 UTC Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.741287 4789 server.go:460] "Adding debug handlers to kubelet server" Nov 22 07:56:27 crc kubenswrapper[4789]: E1122 07:56:27.741294 4789 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.742318 4789 volume_manager.go:287] "The desired_state_of_world populator starts" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.742341 4789 volume_manager.go:289] "Starting Kubelet Volume Manager" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.742386 4789 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Nov 22 07:56:27 crc kubenswrapper[4789]: E1122 07:56:27.751125 4789 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.56:6443: connect: connection refused" interval="200ms" Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.751303 4789 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.129.56.56:6443: connect: connection refused Nov 22 07:56:27 crc kubenswrapper[4789]: E1122 07:56:27.751370 4789 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.129.56.56:6443: connect: connection refused" logger="UnhandledError" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.751657 4789 factory.go:153] Registering CRI-O factory Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.751673 4789 factory.go:221] Registration of the crio container factory successfully Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.751736 4789 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.751766 4789 factory.go:55] Registering systemd factory Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.751776 4789 factory.go:221] Registration of the systemd container factory successfully Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.751799 4789 factory.go:103] Registering Raw factory Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.751817 4789 manager.go:1196] Started watching for new ooms in manager Nov 22 07:56:27 crc kubenswrapper[4789]: E1122 07:56:27.751199 4789 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.129.56.56:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187a451b5326199a default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-22 07:56:27.730409882 +0000 UTC m=+1.964810195,LastTimestamp:2025-11-22 07:56:27.730409882 +0000 UTC m=+1.964810195,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.752889 4789 manager.go:319] Starting recovery of all containers Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.765953 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.766667 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.766709 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.766721 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.766734 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.766783 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.766795 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.766805 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.766819 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.766831 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.766867 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.766879 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.766893 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.766906 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.766916 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.766937 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.766948 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.766961 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.766974 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.766985 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.766997 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.767010 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.767021 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.767033 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.767043 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.767057 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.767070 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.767082 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.767095 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.767107 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.767117 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.767128 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.767139 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.767150 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.767161 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.767174 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.767186 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.767196 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.767206 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.767216 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.767226 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.767235 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.767245 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.767258 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.767269 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.767279 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.767289 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.767299 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.767310 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.767320 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.767329 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.767339 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.767353 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.767363 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.767376 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.767387 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.767398 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.767408 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.767418 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.767429 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.767438 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.767447 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.767456 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.767468 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.767477 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.767487 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.767498 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.767506 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.767516 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.767525 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.767534 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.767543 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.767552 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.767561 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.767570 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.767579 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.767588 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.767597 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.767605 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.767619 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.767629 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.767638 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.767648 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.767659 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.767670 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.767678 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.767687 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.767695 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.767714 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.767723 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.767732 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.767742 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.767763 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.767771 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.767781 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.767789 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.767799 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.767810 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.767819 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.767829 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.767837 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.767846 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.767855 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.767864 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.767877 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.767888 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.767897 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.767908 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.767918 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.767929 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.767974 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.769668 4789 manager.go:324] Recovery completed Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.771089 4789 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.771141 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.771164 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.771179 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.771191 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.771202 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.771212 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.771223 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.771234 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.771244 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.771255 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.771267 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.771278 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.771289 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.771302 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.771315 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.771327 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.771338 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.771348 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.771357 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.771370 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.771383 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.771398 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.771409 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.771418 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.771427 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.771438 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.771449 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.771461 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.771470 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.771481 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.771490 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.771500 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.771510 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.771522 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.771532 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.771550 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.771562 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.771576 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.771585 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.771595 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.771604 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.771614 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.771628 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.771660 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.771669 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.771678 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.771688 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.771697 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.771705 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.771714 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.771724 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.771732 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.771741 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.771767 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.771777 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.771787 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.771797 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.771806 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.771815 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.771824 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.771833 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.771842 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.771851 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.771860 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.771870 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.771883 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.771891 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.771899 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.771908 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.771917 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.771927 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.771935 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.771943 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.771952 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.771960 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.771968 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.771976 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.771985 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.771995 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.772006 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.772016 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.772026 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.772035 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.772045 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.772053 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.772061 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.772069 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.772078 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.772086 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.772095 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.772103 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.772111 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.772120 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.772129 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.772137 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.772148 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.772156 4789 reconstruct.go:97] "Volume reconstruction finished" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.772165 4789 reconciler.go:26] "Reconciler: start to sync state" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.782993 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.784972 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.785046 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.785066 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.786177 4789 cpu_manager.go:225] "Starting CPU manager" policy="none" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.786194 4789 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.786211 4789 state_mem.go:36] "Initialized new in-memory state store" Nov 22 07:56:27 crc kubenswrapper[4789]: E1122 07:56:27.842566 4789 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 22 07:56:27 crc kubenswrapper[4789]: E1122 07:56:27.943223 4789 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 22 07:56:27 crc kubenswrapper[4789]: E1122 07:56:27.952218 4789 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.56:6443: connect: connection refused" interval="400ms" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.960854 4789 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.963833 4789 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.963871 4789 status_manager.go:217] "Starting to sync pod status with apiserver" Nov 22 07:56:27 crc kubenswrapper[4789]: I1122 07:56:27.963900 4789 kubelet.go:2335] "Starting kubelet main sync loop" Nov 22 07:56:27 crc kubenswrapper[4789]: E1122 07:56:27.963950 4789 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Nov 22 07:56:27 crc kubenswrapper[4789]: W1122 07:56:27.966151 4789 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.129.56.56:6443: connect: connection refused Nov 22 07:56:27 crc kubenswrapper[4789]: E1122 07:56:27.966224 4789 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.129.56.56:6443: connect: connection refused" logger="UnhandledError" Nov 22 07:56:28 crc kubenswrapper[4789]: E1122 07:56:28.044294 4789 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 22 07:56:28 crc kubenswrapper[4789]: E1122 07:56:28.064945 4789 kubelet.go:2359] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Nov 22 07:56:28 crc kubenswrapper[4789]: E1122 07:56:28.145103 4789 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 22 07:56:28 crc kubenswrapper[4789]: E1122 07:56:28.245707 4789 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 22 07:56:28 crc kubenswrapper[4789]: E1122 07:56:28.265047 4789 kubelet.go:2359] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Nov 22 07:56:28 crc kubenswrapper[4789]: E1122 07:56:28.346333 4789 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 22 07:56:28 crc kubenswrapper[4789]: E1122 07:56:28.353546 4789 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.56:6443: connect: connection refused" interval="800ms" Nov 22 07:56:28 crc kubenswrapper[4789]: E1122 07:56:28.447444 4789 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 22 07:56:28 crc kubenswrapper[4789]: I1122 07:56:28.487365 4789 policy_none.go:49] "None policy: Start" Nov 22 07:56:28 crc kubenswrapper[4789]: I1122 07:56:28.489243 4789 memory_manager.go:170] "Starting memorymanager" policy="None" Nov 22 07:56:28 crc kubenswrapper[4789]: I1122 07:56:28.489317 4789 state_mem.go:35] "Initializing new in-memory state store" Nov 22 07:56:28 crc kubenswrapper[4789]: E1122 07:56:28.526300 4789 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.129.56.56:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187a451b5326199a default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-22 07:56:27.730409882 +0000 UTC m=+1.964810195,LastTimestamp:2025-11-22 07:56:27.730409882 +0000 UTC m=+1.964810195,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 22 07:56:28 crc kubenswrapper[4789]: E1122 07:56:28.547800 4789 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 22 07:56:28 crc kubenswrapper[4789]: E1122 07:56:28.648877 4789 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 22 07:56:28 crc kubenswrapper[4789]: E1122 07:56:28.665583 4789 kubelet.go:2359] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Nov 22 07:56:28 crc kubenswrapper[4789]: I1122 07:56:28.733037 4789 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.56:6443: connect: connection refused Nov 22 07:56:28 crc kubenswrapper[4789]: I1122 07:56:28.742129 4789 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-02 16:49:06.417639985 +0000 UTC Nov 22 07:56:28 crc kubenswrapper[4789]: I1122 07:56:28.742242 4789 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 248h52m37.675401612s for next certificate rotation Nov 22 07:56:28 crc kubenswrapper[4789]: E1122 07:56:28.749372 4789 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 22 07:56:28 crc kubenswrapper[4789]: W1122 07:56:28.826861 4789 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.129.56.56:6443: connect: connection refused Nov 22 07:56:28 crc kubenswrapper[4789]: E1122 07:56:28.826950 4789 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.129.56.56:6443: connect: connection refused" logger="UnhandledError" Nov 22 07:56:28 crc kubenswrapper[4789]: E1122 07:56:28.849493 4789 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 22 07:56:28 crc kubenswrapper[4789]: I1122 07:56:28.856240 4789 manager.go:334] "Starting Device Plugin manager" Nov 22 07:56:28 crc kubenswrapper[4789]: I1122 07:56:28.856306 4789 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Nov 22 07:56:28 crc kubenswrapper[4789]: I1122 07:56:28.856321 4789 server.go:79] "Starting device plugin registration server" Nov 22 07:56:28 crc kubenswrapper[4789]: I1122 07:56:28.856893 4789 eviction_manager.go:189] "Eviction manager: starting control loop" Nov 22 07:56:28 crc kubenswrapper[4789]: I1122 07:56:28.856913 4789 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Nov 22 07:56:28 crc kubenswrapper[4789]: I1122 07:56:28.857342 4789 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Nov 22 07:56:28 crc kubenswrapper[4789]: I1122 07:56:28.857939 4789 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Nov 22 07:56:28 crc kubenswrapper[4789]: I1122 07:56:28.857964 4789 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Nov 22 07:56:28 crc kubenswrapper[4789]: E1122 07:56:28.865413 4789 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Nov 22 07:56:28 crc kubenswrapper[4789]: I1122 07:56:28.957390 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 07:56:28 crc kubenswrapper[4789]: I1122 07:56:28.958666 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:28 crc kubenswrapper[4789]: I1122 07:56:28.958695 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:28 crc kubenswrapper[4789]: I1122 07:56:28.958708 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:28 crc kubenswrapper[4789]: I1122 07:56:28.958733 4789 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 22 07:56:28 crc kubenswrapper[4789]: E1122 07:56:28.959306 4789 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.56:6443: connect: connection refused" node="crc" Nov 22 07:56:28 crc kubenswrapper[4789]: W1122 07:56:28.987695 4789 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.129.56.56:6443: connect: connection refused Nov 22 07:56:28 crc kubenswrapper[4789]: E1122 07:56:28.987800 4789 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.129.56.56:6443: connect: connection refused" logger="UnhandledError" Nov 22 07:56:29 crc kubenswrapper[4789]: W1122 07:56:29.139258 4789 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.129.56.56:6443: connect: connection refused Nov 22 07:56:29 crc kubenswrapper[4789]: E1122 07:56:29.139329 4789 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.129.56.56:6443: connect: connection refused" logger="UnhandledError" Nov 22 07:56:29 crc kubenswrapper[4789]: E1122 07:56:29.154536 4789 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.56:6443: connect: connection refused" interval="1.6s" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.159582 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.160879 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.160923 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.160936 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.160960 4789 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 22 07:56:29 crc kubenswrapper[4789]: E1122 07:56:29.161474 4789 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.56:6443: connect: connection refused" node="crc" Nov 22 07:56:29 crc kubenswrapper[4789]: W1122 07:56:29.381363 4789 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.129.56.56:6443: connect: connection refused Nov 22 07:56:29 crc kubenswrapper[4789]: E1122 07:56:29.381426 4789 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.129.56.56:6443: connect: connection refused" logger="UnhandledError" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.466060 4789 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc"] Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.466191 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.467377 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.467452 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.467472 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.467711 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.467943 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.468009 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.469348 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.469365 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.469375 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.469403 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.469412 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.469421 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.469550 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.469888 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.469951 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.470557 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.470605 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.470615 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.470741 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.470884 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.470957 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.471141 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.471178 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.471194 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.471979 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.472034 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.472053 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.472057 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.472104 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.472120 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.472259 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.472345 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.472394 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.473154 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.473206 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.473225 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.473283 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.473307 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.473320 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.473528 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.473574 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.474646 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.474682 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.474697 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.562207 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.563190 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.563269 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.563285 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.563310 4789 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 22 07:56:29 crc kubenswrapper[4789]: E1122 07:56:29.563831 4789 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.56:6443: connect: connection refused" node="crc" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.606225 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.606347 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.606395 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.606428 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.606467 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.606531 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.606586 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.606646 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.606700 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.606740 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.606853 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.606900 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.606948 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.606983 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.607013 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.708331 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.708432 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.708481 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.708527 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.708577 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.708618 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.708644 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.708678 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.708661 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.708735 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.708827 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.708859 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.708885 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.708933 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.708932 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.708983 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.708778 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.708862 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.708950 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.709038 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.708732 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.709070 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.709090 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.709108 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.709002 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.709177 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.709188 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.709233 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.709281 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.709322 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.732374 4789 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.56:6443: connect: connection refused Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.793306 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.820594 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.831086 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.854202 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 22 07:56:29 crc kubenswrapper[4789]: I1122 07:56:29.858602 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Nov 22 07:56:29 crc kubenswrapper[4789]: W1122 07:56:29.969482 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-675e59689ce661a4f356ad4999b5cf0ae5386951c47ea6f6feed7e582d6facbf WatchSource:0}: Error finding container 675e59689ce661a4f356ad4999b5cf0ae5386951c47ea6f6feed7e582d6facbf: Status 404 returned error can't find the container with id 675e59689ce661a4f356ad4999b5cf0ae5386951c47ea6f6feed7e582d6facbf Nov 22 07:56:29 crc kubenswrapper[4789]: W1122 07:56:29.970147 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-3cbb16062b009700ed9a038a0ee9d5fea5f08434faf4b155e577fe2922ce5d64 WatchSource:0}: Error finding container 3cbb16062b009700ed9a038a0ee9d5fea5f08434faf4b155e577fe2922ce5d64: Status 404 returned error can't find the container with id 3cbb16062b009700ed9a038a0ee9d5fea5f08434faf4b155e577fe2922ce5d64 Nov 22 07:56:29 crc kubenswrapper[4789]: W1122 07:56:29.970503 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-dae542fd7c4761f763a5c9fc07ff845e3e7fefa22e42204ce7ccaf8b31663c7b WatchSource:0}: Error finding container dae542fd7c4761f763a5c9fc07ff845e3e7fefa22e42204ce7ccaf8b31663c7b: Status 404 returned error can't find the container with id dae542fd7c4761f763a5c9fc07ff845e3e7fefa22e42204ce7ccaf8b31663c7b Nov 22 07:56:29 crc kubenswrapper[4789]: W1122 07:56:29.971823 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-9724af8a01e23e696c6ca15f0b45668bc73fd4eff4498e92c8f716cf345eaa78 WatchSource:0}: Error finding container 9724af8a01e23e696c6ca15f0b45668bc73fd4eff4498e92c8f716cf345eaa78: Status 404 returned error can't find the container with id 9724af8a01e23e696c6ca15f0b45668bc73fd4eff4498e92c8f716cf345eaa78 Nov 22 07:56:29 crc kubenswrapper[4789]: W1122 07:56:29.975051 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-6596fafb90d2b691516937d96e2caa29940d1cf66cc67fecfbbdafb3b9a5a50f WatchSource:0}: Error finding container 6596fafb90d2b691516937d96e2caa29940d1cf66cc67fecfbbdafb3b9a5a50f: Status 404 returned error can't find the container with id 6596fafb90d2b691516937d96e2caa29940d1cf66cc67fecfbbdafb3b9a5a50f Nov 22 07:56:30 crc kubenswrapper[4789]: I1122 07:56:30.364314 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 07:56:30 crc kubenswrapper[4789]: I1122 07:56:30.365889 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:30 crc kubenswrapper[4789]: I1122 07:56:30.365958 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:30 crc kubenswrapper[4789]: I1122 07:56:30.365977 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:30 crc kubenswrapper[4789]: I1122 07:56:30.366009 4789 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 22 07:56:30 crc kubenswrapper[4789]: E1122 07:56:30.366620 4789 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.56:6443: connect: connection refused" node="crc" Nov 22 07:56:30 crc kubenswrapper[4789]: I1122 07:56:30.732818 4789 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.56:6443: connect: connection refused Nov 22 07:56:30 crc kubenswrapper[4789]: E1122 07:56:30.756110 4789 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.56:6443: connect: connection refused" interval="3.2s" Nov 22 07:56:30 crc kubenswrapper[4789]: W1122 07:56:30.941011 4789 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.129.56.56:6443: connect: connection refused Nov 22 07:56:30 crc kubenswrapper[4789]: E1122 07:56:30.941143 4789 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.129.56.56:6443: connect: connection refused" logger="UnhandledError" Nov 22 07:56:30 crc kubenswrapper[4789]: I1122 07:56:30.972325 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"675e59689ce661a4f356ad4999b5cf0ae5386951c47ea6f6feed7e582d6facbf"} Nov 22 07:56:30 crc kubenswrapper[4789]: I1122 07:56:30.973827 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"dae542fd7c4761f763a5c9fc07ff845e3e7fefa22e42204ce7ccaf8b31663c7b"} Nov 22 07:56:30 crc kubenswrapper[4789]: I1122 07:56:30.974997 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"9724af8a01e23e696c6ca15f0b45668bc73fd4eff4498e92c8f716cf345eaa78"} Nov 22 07:56:30 crc kubenswrapper[4789]: I1122 07:56:30.976201 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"3cbb16062b009700ed9a038a0ee9d5fea5f08434faf4b155e577fe2922ce5d64"} Nov 22 07:56:30 crc kubenswrapper[4789]: I1122 07:56:30.977288 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"6596fafb90d2b691516937d96e2caa29940d1cf66cc67fecfbbdafb3b9a5a50f"} Nov 22 07:56:31 crc kubenswrapper[4789]: W1122 07:56:31.364640 4789 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.129.56.56:6443: connect: connection refused Nov 22 07:56:31 crc kubenswrapper[4789]: E1122 07:56:31.365666 4789 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.129.56.56:6443: connect: connection refused" logger="UnhandledError" Nov 22 07:56:31 crc kubenswrapper[4789]: I1122 07:56:31.733221 4789 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.56:6443: connect: connection refused Nov 22 07:56:31 crc kubenswrapper[4789]: I1122 07:56:31.967642 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 07:56:31 crc kubenswrapper[4789]: I1122 07:56:31.969843 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:31 crc kubenswrapper[4789]: I1122 07:56:31.969893 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:31 crc kubenswrapper[4789]: I1122 07:56:31.969909 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:31 crc kubenswrapper[4789]: I1122 07:56:31.969941 4789 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 22 07:56:31 crc kubenswrapper[4789]: E1122 07:56:31.970490 4789 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.56:6443: connect: connection refused" node="crc" Nov 22 07:56:31 crc kubenswrapper[4789]: I1122 07:56:31.985428 4789 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="ecc1356a3d72607e2210df5b2cc91ae62e54e796a4b311c0d98cbc2e4376d5ae" exitCode=0 Nov 22 07:56:31 crc kubenswrapper[4789]: I1122 07:56:31.985514 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"ecc1356a3d72607e2210df5b2cc91ae62e54e796a4b311c0d98cbc2e4376d5ae"} Nov 22 07:56:31 crc kubenswrapper[4789]: I1122 07:56:31.985682 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 07:56:31 crc kubenswrapper[4789]: I1122 07:56:31.987339 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:31 crc kubenswrapper[4789]: I1122 07:56:31.987396 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:31 crc kubenswrapper[4789]: I1122 07:56:31.987415 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:31 crc kubenswrapper[4789]: I1122 07:56:31.988455 4789 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="b06327bb060078f5598ca94bfe5db80187cdbf6b5c8b6385f1b97b102ac97185" exitCode=0 Nov 22 07:56:31 crc kubenswrapper[4789]: I1122 07:56:31.988510 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"b06327bb060078f5598ca94bfe5db80187cdbf6b5c8b6385f1b97b102ac97185"} Nov 22 07:56:31 crc kubenswrapper[4789]: I1122 07:56:31.988533 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 07:56:31 crc kubenswrapper[4789]: I1122 07:56:31.989957 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:31 crc kubenswrapper[4789]: I1122 07:56:31.989995 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:31 crc kubenswrapper[4789]: I1122 07:56:31.990009 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:31 crc kubenswrapper[4789]: I1122 07:56:31.991346 4789 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="ac2b567fcebc517d3fb4366aebce03351183ea03ce62f86c2abdf5b95e20b34a" exitCode=0 Nov 22 07:56:31 crc kubenswrapper[4789]: I1122 07:56:31.991384 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"ac2b567fcebc517d3fb4366aebce03351183ea03ce62f86c2abdf5b95e20b34a"} Nov 22 07:56:31 crc kubenswrapper[4789]: I1122 07:56:31.991495 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 07:56:31 crc kubenswrapper[4789]: I1122 07:56:31.992836 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:31 crc kubenswrapper[4789]: I1122 07:56:31.992874 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:31 crc kubenswrapper[4789]: I1122 07:56:31.992886 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:31 crc kubenswrapper[4789]: I1122 07:56:31.994725 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 07:56:31 crc kubenswrapper[4789]: I1122 07:56:31.994834 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"ca5af01b2e23b2bcac23ed767bc7dcd2ac0be1010ee5e675b3b44333ee337180"} Nov 22 07:56:31 crc kubenswrapper[4789]: I1122 07:56:31.994869 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"43d9e10254883b6491bef617cf9f0b319171bc1fbc49c7770ed9630a8a1ae2cd"} Nov 22 07:56:31 crc kubenswrapper[4789]: I1122 07:56:31.994884 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"223c2300bb6afb8812a72eb7605015b0473f59046565cdf39fb9b9f0cf6634bb"} Nov 22 07:56:31 crc kubenswrapper[4789]: I1122 07:56:31.995643 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:31 crc kubenswrapper[4789]: I1122 07:56:31.995678 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:31 crc kubenswrapper[4789]: I1122 07:56:31.995692 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:31 crc kubenswrapper[4789]: I1122 07:56:31.997714 4789 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="0755dd1f1e7b1fda23c8973b56521ab15b62b9eb76f9b5604110e04e2f25b422" exitCode=0 Nov 22 07:56:31 crc kubenswrapper[4789]: I1122 07:56:31.997762 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"0755dd1f1e7b1fda23c8973b56521ab15b62b9eb76f9b5604110e04e2f25b422"} Nov 22 07:56:31 crc kubenswrapper[4789]: I1122 07:56:31.997879 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 07:56:31 crc kubenswrapper[4789]: I1122 07:56:31.998498 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:31 crc kubenswrapper[4789]: I1122 07:56:31.998529 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:31 crc kubenswrapper[4789]: I1122 07:56:31.998541 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:32 crc kubenswrapper[4789]: W1122 07:56:32.031411 4789 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.129.56.56:6443: connect: connection refused Nov 22 07:56:32 crc kubenswrapper[4789]: E1122 07:56:32.031482 4789 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.129.56.56:6443: connect: connection refused" logger="UnhandledError" Nov 22 07:56:32 crc kubenswrapper[4789]: W1122 07:56:32.153067 4789 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.129.56.56:6443: connect: connection refused Nov 22 07:56:32 crc kubenswrapper[4789]: E1122 07:56:32.153148 4789 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.129.56.56:6443: connect: connection refused" logger="UnhandledError" Nov 22 07:56:32 crc kubenswrapper[4789]: I1122 07:56:32.732552 4789 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.56:6443: connect: connection refused Nov 22 07:56:33 crc kubenswrapper[4789]: I1122 07:56:33.004344 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"b8e79bfb6172c894868a8e1b47ac701ba4fdaca1fc6b1582f7e70807aadfdf1b"} Nov 22 07:56:33 crc kubenswrapper[4789]: I1122 07:56:33.004515 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 07:56:33 crc kubenswrapper[4789]: I1122 07:56:33.005659 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:33 crc kubenswrapper[4789]: I1122 07:56:33.005689 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:33 crc kubenswrapper[4789]: I1122 07:56:33.005698 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:33 crc kubenswrapper[4789]: I1122 07:56:33.008567 4789 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="680b49efb88db68b75efc6ca47cdfa449c1cf688d277b3e1eaa75cb466e90480" exitCode=0 Nov 22 07:56:33 crc kubenswrapper[4789]: I1122 07:56:33.008688 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 07:56:33 crc kubenswrapper[4789]: I1122 07:56:33.008710 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"680b49efb88db68b75efc6ca47cdfa449c1cf688d277b3e1eaa75cb466e90480"} Nov 22 07:56:33 crc kubenswrapper[4789]: I1122 07:56:33.011774 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:33 crc kubenswrapper[4789]: I1122 07:56:33.011814 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:33 crc kubenswrapper[4789]: I1122 07:56:33.011825 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:33 crc kubenswrapper[4789]: I1122 07:56:33.015279 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"80aaecb38dfcb4eee3fb4794072b15316da66929b2497cd72122ff4b4df3d030"} Nov 22 07:56:33 crc kubenswrapper[4789]: I1122 07:56:33.015325 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 07:56:33 crc kubenswrapper[4789]: I1122 07:56:33.015335 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"640c7b83f299315f76e81e97aa0e2d1fcac82097d725a1aef442220edf2bc629"} Nov 22 07:56:33 crc kubenswrapper[4789]: I1122 07:56:33.015354 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"1be9a55547d17dad2a3aec3dd6bd36d2f176654e0c9fa76987addd070937b158"} Nov 22 07:56:33 crc kubenswrapper[4789]: I1122 07:56:33.020171 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:33 crc kubenswrapper[4789]: I1122 07:56:33.020218 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:33 crc kubenswrapper[4789]: I1122 07:56:33.020232 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:33 crc kubenswrapper[4789]: I1122 07:56:33.022342 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 07:56:33 crc kubenswrapper[4789]: I1122 07:56:33.022334 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"6846ff27c129ade97811d4af6d7cfb5449cd41d2f6e302821d2c5726a487e6a4"} Nov 22 07:56:33 crc kubenswrapper[4789]: I1122 07:56:33.024819 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:33 crc kubenswrapper[4789]: I1122 07:56:33.024866 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:33 crc kubenswrapper[4789]: I1122 07:56:33.024885 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:33 crc kubenswrapper[4789]: I1122 07:56:33.029387 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"263dce4c3f4f3fe388b21d4bb55a9fc7d956c3cca7da34ee00f470d88ef501f1"} Nov 22 07:56:33 crc kubenswrapper[4789]: I1122 07:56:33.029457 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"3b6666e0ce42a46f669505659298a19931b1bf7c253ae6a97c6acfd80a9e717f"} Nov 22 07:56:33 crc kubenswrapper[4789]: I1122 07:56:33.029475 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"251e72e1cd0f3e0ee195bacf33ab38765d1dd943adf4395f9d6dd282bda024cb"} Nov 22 07:56:33 crc kubenswrapper[4789]: I1122 07:56:33.029487 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"ba91466edcf49f73751622f98632b3040a046ccd8a42ba8ecee7ab08688a8c24"} Nov 22 07:56:33 crc kubenswrapper[4789]: I1122 07:56:33.732619 4789 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.56:6443: connect: connection refused Nov 22 07:56:33 crc kubenswrapper[4789]: E1122 07:56:33.957433 4789 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.56:6443: connect: connection refused" interval="6.4s" Nov 22 07:56:34 crc kubenswrapper[4789]: I1122 07:56:34.035029 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"b0233c87ba1f1a4ee0ab1214fc45db43ad6122e128c5890cf29134b0e0217ce6"} Nov 22 07:56:34 crc kubenswrapper[4789]: I1122 07:56:34.036026 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 07:56:34 crc kubenswrapper[4789]: I1122 07:56:34.038015 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:34 crc kubenswrapper[4789]: I1122 07:56:34.038215 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:34 crc kubenswrapper[4789]: I1122 07:56:34.038344 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:34 crc kubenswrapper[4789]: I1122 07:56:34.039651 4789 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="c5b1cf28577e60931184499a22559f30b0bdf5dab0b6cb542c948fec5d9c54cc" exitCode=0 Nov 22 07:56:34 crc kubenswrapper[4789]: I1122 07:56:34.039862 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 07:56:34 crc kubenswrapper[4789]: I1122 07:56:34.040081 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"c5b1cf28577e60931184499a22559f30b0bdf5dab0b6cb542c948fec5d9c54cc"} Nov 22 07:56:34 crc kubenswrapper[4789]: I1122 07:56:34.040228 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 07:56:34 crc kubenswrapper[4789]: I1122 07:56:34.040246 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 22 07:56:34 crc kubenswrapper[4789]: I1122 07:56:34.040293 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 07:56:34 crc kubenswrapper[4789]: I1122 07:56:34.041174 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 07:56:34 crc kubenswrapper[4789]: I1122 07:56:34.052111 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:34 crc kubenswrapper[4789]: I1122 07:56:34.052207 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:34 crc kubenswrapper[4789]: I1122 07:56:34.052246 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:34 crc kubenswrapper[4789]: I1122 07:56:34.052125 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:34 crc kubenswrapper[4789]: I1122 07:56:34.052392 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:34 crc kubenswrapper[4789]: I1122 07:56:34.052426 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:34 crc kubenswrapper[4789]: I1122 07:56:34.054770 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:34 crc kubenswrapper[4789]: I1122 07:56:34.054819 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:34 crc kubenswrapper[4789]: I1122 07:56:34.054833 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:34 crc kubenswrapper[4789]: I1122 07:56:34.054890 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:34 crc kubenswrapper[4789]: I1122 07:56:34.054923 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:34 crc kubenswrapper[4789]: I1122 07:56:34.054941 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:35 crc kubenswrapper[4789]: I1122 07:56:35.046634 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"a1d47e3e0a1939863a1ae9e1f52262d8fad029cd90552c21e73a98cdf0a803ba"} Nov 22 07:56:35 crc kubenswrapper[4789]: I1122 07:56:35.046699 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"65813cc305e2105a598eb0b701dd079f1f26db8cc1b2e67679e7086c2ba3630f"} Nov 22 07:56:35 crc kubenswrapper[4789]: I1122 07:56:35.046708 4789 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 22 07:56:35 crc kubenswrapper[4789]: I1122 07:56:35.046720 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"53953e73ae4a11a47b7a5332786b97d9f129d9576e66aa2664836ec5f3786683"} Nov 22 07:56:35 crc kubenswrapper[4789]: I1122 07:56:35.046728 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 07:56:35 crc kubenswrapper[4789]: I1122 07:56:35.046737 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"c410501c102d72819d29aa0036b6270b1ce9804bba84092161b2c2d8ebaa885a"} Nov 22 07:56:35 crc kubenswrapper[4789]: I1122 07:56:35.046783 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 07:56:35 crc kubenswrapper[4789]: I1122 07:56:35.047836 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:35 crc kubenswrapper[4789]: I1122 07:56:35.047864 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:35 crc kubenswrapper[4789]: I1122 07:56:35.047873 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:35 crc kubenswrapper[4789]: I1122 07:56:35.047867 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:35 crc kubenswrapper[4789]: I1122 07:56:35.047995 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:35 crc kubenswrapper[4789]: I1122 07:56:35.048014 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:35 crc kubenswrapper[4789]: I1122 07:56:35.171371 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 07:56:35 crc kubenswrapper[4789]: I1122 07:56:35.172681 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:35 crc kubenswrapper[4789]: I1122 07:56:35.172715 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:35 crc kubenswrapper[4789]: I1122 07:56:35.172725 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:35 crc kubenswrapper[4789]: I1122 07:56:35.172748 4789 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 22 07:56:35 crc kubenswrapper[4789]: I1122 07:56:35.591885 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 22 07:56:35 crc kubenswrapper[4789]: I1122 07:56:35.592635 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 07:56:35 crc kubenswrapper[4789]: I1122 07:56:35.594290 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:35 crc kubenswrapper[4789]: I1122 07:56:35.594363 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:35 crc kubenswrapper[4789]: I1122 07:56:35.594383 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:35 crc kubenswrapper[4789]: I1122 07:56:35.601057 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 22 07:56:36 crc kubenswrapper[4789]: I1122 07:56:36.056087 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 07:56:36 crc kubenswrapper[4789]: I1122 07:56:36.056155 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"c9d09f0deafe8d22df512d2acd120ce10eec182173daaf5029ab72f848ceecd5"} Nov 22 07:56:36 crc kubenswrapper[4789]: I1122 07:56:36.056195 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 07:56:36 crc kubenswrapper[4789]: I1122 07:56:36.057217 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:36 crc kubenswrapper[4789]: I1122 07:56:36.057289 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:36 crc kubenswrapper[4789]: I1122 07:56:36.057311 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:36 crc kubenswrapper[4789]: I1122 07:56:36.057565 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:36 crc kubenswrapper[4789]: I1122 07:56:36.057604 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:36 crc kubenswrapper[4789]: I1122 07:56:36.057615 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:36 crc kubenswrapper[4789]: I1122 07:56:36.782949 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 22 07:56:37 crc kubenswrapper[4789]: I1122 07:56:37.003468 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 22 07:56:37 crc kubenswrapper[4789]: I1122 07:56:37.003644 4789 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 22 07:56:37 crc kubenswrapper[4789]: I1122 07:56:37.003686 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 07:56:37 crc kubenswrapper[4789]: I1122 07:56:37.005068 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:37 crc kubenswrapper[4789]: I1122 07:56:37.005111 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:37 crc kubenswrapper[4789]: I1122 07:56:37.005124 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:37 crc kubenswrapper[4789]: I1122 07:56:37.058426 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 07:56:37 crc kubenswrapper[4789]: I1122 07:56:37.058426 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 07:56:37 crc kubenswrapper[4789]: I1122 07:56:37.059285 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:37 crc kubenswrapper[4789]: I1122 07:56:37.059320 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:37 crc kubenswrapper[4789]: I1122 07:56:37.059331 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:37 crc kubenswrapper[4789]: I1122 07:56:37.059468 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:37 crc kubenswrapper[4789]: I1122 07:56:37.059522 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:37 crc kubenswrapper[4789]: I1122 07:56:37.059531 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:37 crc kubenswrapper[4789]: I1122 07:56:37.384909 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 22 07:56:37 crc kubenswrapper[4789]: I1122 07:56:37.385070 4789 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 22 07:56:37 crc kubenswrapper[4789]: I1122 07:56:37.385111 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 07:56:37 crc kubenswrapper[4789]: I1122 07:56:37.386485 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:37 crc kubenswrapper[4789]: I1122 07:56:37.386553 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:37 crc kubenswrapper[4789]: I1122 07:56:37.386572 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:37 crc kubenswrapper[4789]: I1122 07:56:37.721343 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 22 07:56:37 crc kubenswrapper[4789]: I1122 07:56:37.812031 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Nov 22 07:56:38 crc kubenswrapper[4789]: I1122 07:56:38.060675 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 07:56:38 crc kubenswrapper[4789]: I1122 07:56:38.060714 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 07:56:38 crc kubenswrapper[4789]: I1122 07:56:38.065358 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:38 crc kubenswrapper[4789]: I1122 07:56:38.065402 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:38 crc kubenswrapper[4789]: I1122 07:56:38.065418 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:38 crc kubenswrapper[4789]: I1122 07:56:38.065424 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:38 crc kubenswrapper[4789]: I1122 07:56:38.065453 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:38 crc kubenswrapper[4789]: I1122 07:56:38.065462 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:38 crc kubenswrapper[4789]: E1122 07:56:38.865639 4789 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Nov 22 07:56:39 crc kubenswrapper[4789]: I1122 07:56:39.177672 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 22 07:56:39 crc kubenswrapper[4789]: I1122 07:56:39.177972 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 07:56:39 crc kubenswrapper[4789]: I1122 07:56:39.179802 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:39 crc kubenswrapper[4789]: I1122 07:56:39.179855 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:39 crc kubenswrapper[4789]: I1122 07:56:39.179877 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:42 crc kubenswrapper[4789]: I1122 07:56:42.365262 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 22 07:56:42 crc kubenswrapper[4789]: I1122 07:56:42.365450 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 07:56:42 crc kubenswrapper[4789]: I1122 07:56:42.367958 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:42 crc kubenswrapper[4789]: I1122 07:56:42.367992 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:42 crc kubenswrapper[4789]: I1122 07:56:42.368006 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:42 crc kubenswrapper[4789]: I1122 07:56:42.371162 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 22 07:56:42 crc kubenswrapper[4789]: I1122 07:56:42.966232 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Nov 22 07:56:42 crc kubenswrapper[4789]: I1122 07:56:42.966444 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 07:56:42 crc kubenswrapper[4789]: I1122 07:56:42.968382 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:42 crc kubenswrapper[4789]: I1122 07:56:42.968423 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:42 crc kubenswrapper[4789]: I1122 07:56:42.968437 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:43 crc kubenswrapper[4789]: I1122 07:56:43.073285 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 07:56:43 crc kubenswrapper[4789]: I1122 07:56:43.074263 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:43 crc kubenswrapper[4789]: I1122 07:56:43.074319 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:43 crc kubenswrapper[4789]: I1122 07:56:43.074331 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:44 crc kubenswrapper[4789]: I1122 07:56:44.213307 4789 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:47698->192.168.126.11:17697: read: connection reset by peer" start-of-body= Nov 22 07:56:44 crc kubenswrapper[4789]: I1122 07:56:44.213891 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:47698->192.168.126.11:17697: read: connection reset by peer" Nov 22 07:56:44 crc kubenswrapper[4789]: I1122 07:56:44.733644 4789 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Nov 22 07:56:45 crc kubenswrapper[4789]: I1122 07:56:45.079549 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 22 07:56:45 crc kubenswrapper[4789]: I1122 07:56:45.081298 4789 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="b0233c87ba1f1a4ee0ab1214fc45db43ad6122e128c5890cf29134b0e0217ce6" exitCode=255 Nov 22 07:56:45 crc kubenswrapper[4789]: I1122 07:56:45.081342 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"b0233c87ba1f1a4ee0ab1214fc45db43ad6122e128c5890cf29134b0e0217ce6"} Nov 22 07:56:45 crc kubenswrapper[4789]: I1122 07:56:45.081530 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 07:56:45 crc kubenswrapper[4789]: I1122 07:56:45.082353 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:45 crc kubenswrapper[4789]: I1122 07:56:45.082388 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:45 crc kubenswrapper[4789]: I1122 07:56:45.082401 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:45 crc kubenswrapper[4789]: I1122 07:56:45.082929 4789 scope.go:117] "RemoveContainer" containerID="b0233c87ba1f1a4ee0ab1214fc45db43ad6122e128c5890cf29134b0e0217ce6" Nov 22 07:56:45 crc kubenswrapper[4789]: E1122 07:56:45.174121 4789 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": net/http: TLS handshake timeout" node="crc" Nov 22 07:56:45 crc kubenswrapper[4789]: W1122 07:56:45.177376 4789 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout Nov 22 07:56:45 crc kubenswrapper[4789]: I1122 07:56:45.177460 4789 trace.go:236] Trace[1964541869]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (22-Nov-2025 07:56:35.176) (total time: 10001ms): Nov 22 07:56:45 crc kubenswrapper[4789]: Trace[1964541869]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10000ms (07:56:45.177) Nov 22 07:56:45 crc kubenswrapper[4789]: Trace[1964541869]: [10.001016807s] [10.001016807s] END Nov 22 07:56:45 crc kubenswrapper[4789]: E1122 07:56:45.177479 4789 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Nov 22 07:56:45 crc kubenswrapper[4789]: I1122 07:56:45.365436 4789 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 22 07:56:45 crc kubenswrapper[4789]: I1122 07:56:45.365504 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Nov 22 07:56:45 crc kubenswrapper[4789]: W1122 07:56:45.393814 4789 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout Nov 22 07:56:45 crc kubenswrapper[4789]: I1122 07:56:45.393905 4789 trace.go:236] Trace[2048100287]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (22-Nov-2025 07:56:35.392) (total time: 10001ms): Nov 22 07:56:45 crc kubenswrapper[4789]: Trace[2048100287]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (07:56:45.393) Nov 22 07:56:45 crc kubenswrapper[4789]: Trace[2048100287]: [10.001215582s] [10.001215582s] END Nov 22 07:56:45 crc kubenswrapper[4789]: E1122 07:56:45.393927 4789 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Nov 22 07:56:45 crc kubenswrapper[4789]: I1122 07:56:45.568395 4789 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Nov 22 07:56:45 crc kubenswrapper[4789]: I1122 07:56:45.568452 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Nov 22 07:56:45 crc kubenswrapper[4789]: I1122 07:56:45.574111 4789 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Nov 22 07:56:45 crc kubenswrapper[4789]: I1122 07:56:45.574179 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Nov 22 07:56:46 crc kubenswrapper[4789]: I1122 07:56:46.086732 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 22 07:56:46 crc kubenswrapper[4789]: I1122 07:56:46.089221 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"098c31b822e7c2351f1142af36b942d67ff8262896e28022c96a187d40bd0ddb"} Nov 22 07:56:46 crc kubenswrapper[4789]: I1122 07:56:46.089381 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 07:56:46 crc kubenswrapper[4789]: I1122 07:56:46.090295 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:46 crc kubenswrapper[4789]: I1122 07:56:46.090341 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:46 crc kubenswrapper[4789]: I1122 07:56:46.090354 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:47 crc kubenswrapper[4789]: I1122 07:56:47.390929 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 22 07:56:47 crc kubenswrapper[4789]: I1122 07:56:47.391148 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 07:56:47 crc kubenswrapper[4789]: I1122 07:56:47.391263 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 22 07:56:47 crc kubenswrapper[4789]: I1122 07:56:47.392439 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:47 crc kubenswrapper[4789]: I1122 07:56:47.392506 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:47 crc kubenswrapper[4789]: I1122 07:56:47.392528 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:47 crc kubenswrapper[4789]: I1122 07:56:47.400150 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 22 07:56:48 crc kubenswrapper[4789]: I1122 07:56:48.093616 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 07:56:48 crc kubenswrapper[4789]: I1122 07:56:48.094586 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:48 crc kubenswrapper[4789]: I1122 07:56:48.094617 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:48 crc kubenswrapper[4789]: I1122 07:56:48.094627 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:48 crc kubenswrapper[4789]: E1122 07:56:48.866445 4789 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Nov 22 07:56:49 crc kubenswrapper[4789]: I1122 07:56:49.095892 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 07:56:49 crc kubenswrapper[4789]: I1122 07:56:49.097613 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:49 crc kubenswrapper[4789]: I1122 07:56:49.097667 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:49 crc kubenswrapper[4789]: I1122 07:56:49.097682 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:50 crc kubenswrapper[4789]: E1122 07:56:50.567956 4789 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="7s" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.570833 4789 trace.go:236] Trace[1088445345]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (22-Nov-2025 07:56:36.656) (total time: 13914ms): Nov 22 07:56:50 crc kubenswrapper[4789]: Trace[1088445345]: ---"Objects listed" error: 13914ms (07:56:50.570) Nov 22 07:56:50 crc kubenswrapper[4789]: Trace[1088445345]: [13.914235703s] [13.914235703s] END Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.570871 4789 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.570988 4789 trace.go:236] Trace[1387249762]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (22-Nov-2025 07:56:38.262) (total time: 12308ms): Nov 22 07:56:50 crc kubenswrapper[4789]: Trace[1387249762]: ---"Objects listed" error: 12308ms (07:56:50.570) Nov 22 07:56:50 crc kubenswrapper[4789]: Trace[1387249762]: [12.308358769s] [12.308358769s] END Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.571016 4789 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.571512 4789 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.731811 4789 apiserver.go:52] "Watching apiserver" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.735729 4789 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.735944 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf"] Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.736232 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.736311 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.736402 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 07:56:50 crc kubenswrapper[4789]: E1122 07:56:50.736518 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.736587 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 22 07:56:50 crc kubenswrapper[4789]: E1122 07:56:50.736815 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.737107 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.737123 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 07:56:50 crc kubenswrapper[4789]: E1122 07:56:50.737189 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.737819 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.738277 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.738325 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.738699 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.740427 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.740558 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.740669 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.740686 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.740795 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.751587 4789 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.764663 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.773301 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.773339 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.773361 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.773379 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.773401 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.773422 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.773444 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.773465 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.773486 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.773506 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.773829 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.773835 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.773856 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.773841 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.773869 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.773894 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.773968 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.774003 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.774031 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.774041 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.774059 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.774087 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.774101 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.774109 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.774133 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.774158 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.774178 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.774203 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.774222 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.774246 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.774268 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.774292 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.774312 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.774334 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.774355 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.774376 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.774396 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.774417 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.774441 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.774464 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.774484 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.774501 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.774520 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.774539 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.774556 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.774575 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.774595 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.774614 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.774636 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.774661 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.774691 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.774712 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.774735 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.774792 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.774817 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.774240 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.774268 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.774428 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.774628 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.774772 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.774836 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.774921 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.774925 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.774989 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.775006 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.775021 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 22 07:56:50 crc kubenswrapper[4789]: E1122 07:56:50.775049 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 07:56:51.275019198 +0000 UTC m=+25.509419571 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.775075 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.775073 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.775103 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.775165 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.775178 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.775190 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.775213 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.775245 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.775272 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.775297 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.775321 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.775344 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.775366 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.775405 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.775431 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.775455 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.775482 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.775504 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.775527 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.775549 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.775576 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.775599 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.775623 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.775645 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.775666 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.775688 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.775709 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.775730 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.775766 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.775787 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.775811 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.775830 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.775851 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.775873 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.775894 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.775915 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.775941 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.775961 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.775983 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.776001 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.776023 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.776046 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.776066 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.776087 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.776112 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.776131 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.776154 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.776177 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.776201 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.776223 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.776243 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.776264 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.776286 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.776308 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.776331 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.776354 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.776377 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.776401 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.776424 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.776449 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.776473 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.776496 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.776520 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.776541 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.776567 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.776590 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.776614 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.776639 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.776664 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.776689 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.776711 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.776733 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.776881 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.776910 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.776934 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.776958 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.776982 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.777008 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.777031 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.777057 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.777086 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.777112 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.777138 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.777165 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.777191 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.777216 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.777242 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.777269 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.777296 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.777325 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.777352 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.777383 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.777409 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.777437 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.777463 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.777487 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.777515 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.777542 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.777583 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.777607 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.777631 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.777654 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.777679 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.777702 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.777775 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.777807 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.777833 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.777860 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.777883 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.777916 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.777942 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.777966 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.777991 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.778016 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.778044 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.778070 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.778096 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.778122 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.778148 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.778176 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.778204 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.778229 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.778255 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.778292 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.778321 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.778349 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.778375 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.778400 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.778428 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.778451 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.778479 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.778507 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.778534 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.778559 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.778582 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.778608 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.778633 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.778661 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.778686 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.778712 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.778736 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.779485 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.779600 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.779629 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.779660 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.779688 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.779761 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.779971 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.780356 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.780400 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.780443 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.780472 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.780501 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.780532 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.780563 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.780591 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.780689 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.780721 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.780751 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.780822 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.780917 4789 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.780933 4789 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.780947 4789 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.780961 4789 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.780975 4789 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.780990 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.781005 4789 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.781019 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.781033 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.781047 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.781061 4789 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.781075 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.781090 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.781105 4789 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.781121 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.781136 4789 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.781150 4789 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.784059 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.784605 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.784652 4789 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.786841 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.787007 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.775094 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.775267 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.775288 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.775395 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.775431 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.775501 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.775539 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.775564 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.775590 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.776117 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.776429 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.776435 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.776590 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.776603 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.776605 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.776813 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.777249 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.777258 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.776907 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.791714 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.777567 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.777598 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.777727 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.777813 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.777883 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.778004 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.778103 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.778293 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.778388 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.778426 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.778446 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.778470 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.778596 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.778566 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.778623 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.778637 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.778680 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.778809 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.778866 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.778991 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.780200 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.780306 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.780350 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.781005 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.781164 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.781708 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.782094 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.782117 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.782353 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.783194 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.783203 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.783382 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.783655 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: E1122 07:56:50.784252 4789 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 22 07:56:50 crc kubenswrapper[4789]: E1122 07:56:50.786225 4789 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.787878 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.787897 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.788352 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.788399 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.789196 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.789390 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.789439 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.789568 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.789730 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.789932 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.790062 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.790128 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.789956 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.790292 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.790423 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.790614 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.790701 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.790841 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.790878 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.790906 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.791109 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.791295 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.791475 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.791773 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.791944 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.792087 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.792169 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.792415 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.792969 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.793142 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.793793 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.794811 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.795110 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.795468 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.795984 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.796035 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.796340 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.796399 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.799405 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.800300 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.801021 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.801998 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.804232 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.804287 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.804451 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.799047 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 22 07:56:50 crc kubenswrapper[4789]: E1122 07:56:50.804679 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-22 07:56:51.304658777 +0000 UTC m=+25.539059130 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.802393 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.802660 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.803143 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.803156 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.803659 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.803685 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: E1122 07:56:50.804818 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-22 07:56:51.304804021 +0000 UTC m=+25.539204404 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.804791 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.805225 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.805278 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.805226 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.805431 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.805620 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: E1122 07:56:50.806024 4789 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 22 07:56:50 crc kubenswrapper[4789]: E1122 07:56:50.806050 4789 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 22 07:56:50 crc kubenswrapper[4789]: E1122 07:56:50.806063 4789 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 07:56:50 crc kubenswrapper[4789]: E1122 07:56:50.806110 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-22 07:56:51.306099106 +0000 UTC m=+25.540499449 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.806104 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.806911 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.807018 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.807023 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.807332 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.807648 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.807692 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.807961 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.808058 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: E1122 07:56:50.808018 4789 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 22 07:56:50 crc kubenswrapper[4789]: E1122 07:56:50.808111 4789 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 22 07:56:50 crc kubenswrapper[4789]: E1122 07:56:50.808125 4789 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 07:56:50 crc kubenswrapper[4789]: E1122 07:56:50.808192 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-22 07:56:51.308175872 +0000 UTC m=+25.542576235 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.808035 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.807983 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.808567 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.811128 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.812920 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.812965 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.813259 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.813953 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.821083 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.821213 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.821258 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.822085 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.822303 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.822393 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.822415 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.822729 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.822978 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.823099 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.823186 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.823410 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.823425 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.823488 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.823803 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.824147 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.824278 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.830651 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.830989 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.831078 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.831173 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.831229 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.831344 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.831382 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.832998 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.835630 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.835892 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.836724 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.836738 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.836840 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.836900 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.842009 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.842116 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.842319 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.844023 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.844952 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.844966 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.844951 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.845019 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.845070 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.845147 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.845207 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.845217 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.845381 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.846015 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.846607 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.847184 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.848154 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.849013 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.850964 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.852076 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.852724 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.852948 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.853770 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.858455 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.859738 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.861305 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.868703 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.870519 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.877937 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.881767 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.881840 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.881988 4789 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.882008 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.882020 4789 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.882032 4789 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.882044 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.882062 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.882074 4789 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.882087 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.882099 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.882116 4789 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.882127 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.882139 4789 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.882149 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.882165 4789 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.882176 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.882188 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.882204 4789 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.882215 4789 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.882227 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.882269 4789 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.882294 4789 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.882307 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.882325 4789 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.882336 4789 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.882346 4789 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.882361 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.882370 4789 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.882379 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.882388 4789 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.882403 4789 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.882432 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.882441 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.882450 4789 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.882463 4789 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.882475 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.882484 4789 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.882493 4789 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.882506 4789 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.882517 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.882527 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.882542 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.882596 4789 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.882607 4789 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.882616 4789 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.882630 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.882640 4789 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.882650 4789 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.882660 4789 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.882673 4789 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.882683 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.882693 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.882710 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.882720 4789 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.882731 4789 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.882740 4789 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.882770 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.882782 4789 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.882793 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.882802 4789 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.882815 4789 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.882826 4789 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.882836 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.882846 4789 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.882859 4789 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.882868 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.882878 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.882891 4789 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.882901 4789 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.882913 4789 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.882924 4789 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.882939 4789 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.882949 4789 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.882959 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.882968 4789 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.882980 4789 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.882991 4789 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.882999 4789 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.883009 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.883022 4789 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.883030 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.883038 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.883050 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.883058 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.883067 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.883076 4789 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.883091 4789 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.883102 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.883119 4789 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.883137 4789 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.883156 4789 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.883168 4789 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.883181 4789 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.883197 4789 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.883208 4789 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.883220 4789 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.883230 4789 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.883243 4789 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.883247 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.883253 4789 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.883289 4789 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.883302 4789 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.883314 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.883331 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.883344 4789 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.883355 4789 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.883366 4789 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.883382 4789 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.883397 4789 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.883411 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.883423 4789 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.883438 4789 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.883469 4789 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.883479 4789 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.883495 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.883506 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.883517 4789 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.883528 4789 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.883544 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.883556 4789 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.883568 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.883579 4789 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.883596 4789 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.883607 4789 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.883618 4789 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.883634 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.883651 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.883652 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.883663 4789 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.883679 4789 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.883696 4789 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.883707 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.883719 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.883729 4789 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.883743 4789 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.883780 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.883792 4789 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.883803 4789 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.883819 4789 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.883830 4789 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.883841 4789 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.883855 4789 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.883865 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.883878 4789 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.883888 4789 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.883902 4789 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.883915 4789 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.883926 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.883937 4789 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.883952 4789 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.883963 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.883973 4789 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.883989 4789 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.884001 4789 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.884012 4789 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.884022 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.884041 4789 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.884054 4789 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.884069 4789 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.884080 4789 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.884097 4789 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.884109 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.884120 4789 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.884130 4789 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.884144 4789 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.884154 4789 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.884165 4789 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.884179 4789 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.884189 4789 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.884200 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.884210 4789 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.884224 4789 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.884234 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.884245 4789 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.884255 4789 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.884270 4789 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.884280 4789 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.884291 4789 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.884302 4789 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.884318 4789 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Nov 22 07:56:50 crc kubenswrapper[4789]: I1122 07:56:50.888364 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.050103 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.056589 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.061086 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 22 07:56:51 crc kubenswrapper[4789]: W1122 07:56:51.068220 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-fee5ae2388b10b7f1ade25844ab584a23d46da19737aeb88e4e9622c7ad1e7f3 WatchSource:0}: Error finding container fee5ae2388b10b7f1ade25844ab584a23d46da19737aeb88e4e9622c7ad1e7f3: Status 404 returned error can't find the container with id fee5ae2388b10b7f1ade25844ab584a23d46da19737aeb88e4e9622c7ad1e7f3 Nov 22 07:56:51 crc kubenswrapper[4789]: E1122 07:56:51.070317 4789 kuberuntime_manager.go:1274] "Unhandled Error" err=< Nov 22 07:56:51 crc kubenswrapper[4789]: container &Container{Name:network-operator,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,Command:[/bin/bash -c #!/bin/bash Nov 22 07:56:51 crc kubenswrapper[4789]: set -o allexport Nov 22 07:56:51 crc kubenswrapper[4789]: if [[ -f /etc/kubernetes/apiserver-url.env ]]; then Nov 22 07:56:51 crc kubenswrapper[4789]: source /etc/kubernetes/apiserver-url.env Nov 22 07:56:51 crc kubenswrapper[4789]: else Nov 22 07:56:51 crc kubenswrapper[4789]: echo "Error: /etc/kubernetes/apiserver-url.env is missing" Nov 22 07:56:51 crc kubenswrapper[4789]: exit 1 Nov 22 07:56:51 crc kubenswrapper[4789]: fi Nov 22 07:56:51 crc kubenswrapper[4789]: exec /usr/bin/cluster-network-operator start --listen=0.0.0.0:9104 Nov 22 07:56:51 crc kubenswrapper[4789]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:cno,HostPort:9104,ContainerPort:9104,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:RELEASE_VERSION,Value:4.18.1,ValueFrom:nil,},EnvVar{Name:KUBE_PROXY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b97554198294bf544fbc116c94a0a1fb2ec8a4de0e926bf9d9e320135f0bee6f,ValueFrom:nil,},EnvVar{Name:KUBE_RBAC_PROXY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09,ValueFrom:nil,},EnvVar{Name:MULTUS_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26,ValueFrom:nil,},EnvVar{Name:MULTUS_ADMISSION_CONTROLLER_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317,ValueFrom:nil,},EnvVar{Name:CNI_PLUGINS_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc,ValueFrom:nil,},EnvVar{Name:BOND_CNI_PLUGIN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78,ValueFrom:nil,},EnvVar{Name:WHEREABOUTS_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4,ValueFrom:nil,},EnvVar{Name:ROUTE_OVERRRIDE_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa,ValueFrom:nil,},EnvVar{Name:MULTUS_NETWORKPOLICY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:23f833d3738d68706eb2f2868bd76bd71cee016cffa6faf5f045a60cc8c6eddd,ValueFrom:nil,},EnvVar{Name:OVN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,ValueFrom:nil,},EnvVar{Name:OVN_NB_RAFT_ELECTION_TIMER,Value:10,ValueFrom:nil,},EnvVar{Name:OVN_SB_RAFT_ELECTION_TIMER,Value:16,ValueFrom:nil,},EnvVar{Name:OVN_NORTHD_PROBE_INTERVAL,Value:10000,ValueFrom:nil,},EnvVar{Name:OVN_CONTROLLER_INACTIVITY_PROBE,Value:180000,ValueFrom:nil,},EnvVar{Name:OVN_NB_INACTIVITY_PROBE,Value:60000,ValueFrom:nil,},EnvVar{Name:EGRESS_ROUTER_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c,ValueFrom:nil,},EnvVar{Name:NETWORK_METRICS_DAEMON_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d,ValueFrom:nil,},EnvVar{Name:NETWORK_CHECK_SOURCE_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,ValueFrom:nil,},EnvVar{Name:NETWORK_CHECK_TARGET_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,ValueFrom:nil,},EnvVar{Name:NETWORK_OPERATOR_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,ValueFrom:nil,},EnvVar{Name:CLOUD_NETWORK_CONFIG_CONTROLLER_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8048f1cb0be521f09749c0a489503cd56d85b68c6ca93380e082cfd693cd97a8,ValueFrom:nil,},EnvVar{Name:CLI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2,ValueFrom:nil,},EnvVar{Name:FRR_K8S_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5dbf844e49bb46b78586930149e5e5f5dc121014c8afd10fe36f3651967cc256,ValueFrom:nil,},EnvVar{Name:NETWORKING_CONSOLE_PLUGIN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd,ValueFrom:nil,},EnvVar{Name:POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:host-etc-kube,ReadOnly:true,MountPath:/etc/kubernetes,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:metrics-tls,ReadOnly:false,MountPath:/var/run/secrets/serving-cert,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rdwmf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-operator-58b4c7f79c-55gtf_openshift-network-operator(37a5e44f-9a88-4405-be8a-b645485e7312): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Nov 22 07:56:51 crc kubenswrapper[4789]: > logger="UnhandledError" Nov 22 07:56:51 crc kubenswrapper[4789]: E1122 07:56:51.071739 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"network-operator\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" podUID="37a5e44f-9a88-4405-be8a-b645485e7312" Nov 22 07:56:51 crc kubenswrapper[4789]: W1122 07:56:51.077006 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-ebcf819aa9b8405ea57d406d6adcb348956a49a49cea32efce2eaf02ee7a2d3c WatchSource:0}: Error finding container ebcf819aa9b8405ea57d406d6adcb348956a49a49cea32efce2eaf02ee7a2d3c: Status 404 returned error can't find the container with id ebcf819aa9b8405ea57d406d6adcb348956a49a49cea32efce2eaf02ee7a2d3c Nov 22 07:56:51 crc kubenswrapper[4789]: E1122 07:56:51.078870 4789 kuberuntime_manager.go:1274] "Unhandled Error" err=< Nov 22 07:56:51 crc kubenswrapper[4789]: container &Container{Name:webhook,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,Command:[/bin/bash -c set -xe Nov 22 07:56:51 crc kubenswrapper[4789]: if [[ -f "/env/_master" ]]; then Nov 22 07:56:51 crc kubenswrapper[4789]: set -o allexport Nov 22 07:56:51 crc kubenswrapper[4789]: source "/env/_master" Nov 22 07:56:51 crc kubenswrapper[4789]: set +o allexport Nov 22 07:56:51 crc kubenswrapper[4789]: fi Nov 22 07:56:51 crc kubenswrapper[4789]: # OVN-K will try to remove hybrid overlay node annotations even when the hybrid overlay is not enabled. Nov 22 07:56:51 crc kubenswrapper[4789]: # https://github.com/ovn-org/ovn-kubernetes/blob/ac6820df0b338a246f10f412cd5ec903bd234694/go-controller/pkg/ovn/master.go#L791 Nov 22 07:56:51 crc kubenswrapper[4789]: ho_enable="--enable-hybrid-overlay" Nov 22 07:56:51 crc kubenswrapper[4789]: echo "I$(date "+%m%d %H:%M:%S.%N") - network-node-identity - start webhook" Nov 22 07:56:51 crc kubenswrapper[4789]: # extra-allowed-user: service account `ovn-kubernetes-control-plane` Nov 22 07:56:51 crc kubenswrapper[4789]: # sets pod annotations in multi-homing layer3 network controller (cluster-manager) Nov 22 07:56:51 crc kubenswrapper[4789]: exec /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.crc.testing:6443 \ Nov 22 07:56:51 crc kubenswrapper[4789]: --webhook-cert-dir="/etc/webhook-cert" \ Nov 22 07:56:51 crc kubenswrapper[4789]: --webhook-host=127.0.0.1 \ Nov 22 07:56:51 crc kubenswrapper[4789]: --webhook-port=9743 \ Nov 22 07:56:51 crc kubenswrapper[4789]: ${ho_enable} \ Nov 22 07:56:51 crc kubenswrapper[4789]: --enable-interconnect \ Nov 22 07:56:51 crc kubenswrapper[4789]: --disable-approver \ Nov 22 07:56:51 crc kubenswrapper[4789]: --extra-allowed-user="system:serviceaccount:openshift-ovn-kubernetes:ovn-kubernetes-control-plane" \ Nov 22 07:56:51 crc kubenswrapper[4789]: --wait-for-kubernetes-api=200s \ Nov 22 07:56:51 crc kubenswrapper[4789]: --pod-admission-conditions="/var/run/ovnkube-identity-config/additional-pod-admission-cond.json" \ Nov 22 07:56:51 crc kubenswrapper[4789]: --loglevel="${LOGLEVEL}" Nov 22 07:56:51 crc kubenswrapper[4789]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOGLEVEL,Value:2,ValueFrom:nil,},EnvVar{Name:KUBERNETES_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:webhook-cert,ReadOnly:false,MountPath:/etc/webhook-cert/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:env-overrides,ReadOnly:false,MountPath:/env,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovnkube-identity-cm,ReadOnly:false,MountPath:/var/run/ovnkube-identity-config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s2kz5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000470000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-node-identity-vrzqb_openshift-network-node-identity(ef543e1b-8068-4ea3-b32a-61027b32e95d): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Nov 22 07:56:51 crc kubenswrapper[4789]: > logger="UnhandledError" Nov 22 07:56:51 crc kubenswrapper[4789]: W1122 07:56:51.079171 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-e2d596dffb054102dab0e4b677138df4a733961307d2b729f4c2f07948a39536 WatchSource:0}: Error finding container e2d596dffb054102dab0e4b677138df4a733961307d2b729f4c2f07948a39536: Status 404 returned error can't find the container with id e2d596dffb054102dab0e4b677138df4a733961307d2b729f4c2f07948a39536 Nov 22 07:56:51 crc kubenswrapper[4789]: E1122 07:56:51.080782 4789 kuberuntime_manager.go:1274] "Unhandled Error" err=< Nov 22 07:56:51 crc kubenswrapper[4789]: container &Container{Name:approver,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,Command:[/bin/bash -c set -xe Nov 22 07:56:51 crc kubenswrapper[4789]: if [[ -f "/env/_master" ]]; then Nov 22 07:56:51 crc kubenswrapper[4789]: set -o allexport Nov 22 07:56:51 crc kubenswrapper[4789]: source "/env/_master" Nov 22 07:56:51 crc kubenswrapper[4789]: set +o allexport Nov 22 07:56:51 crc kubenswrapper[4789]: fi Nov 22 07:56:51 crc kubenswrapper[4789]: Nov 22 07:56:51 crc kubenswrapper[4789]: echo "I$(date "+%m%d %H:%M:%S.%N") - network-node-identity - start approver" Nov 22 07:56:51 crc kubenswrapper[4789]: exec /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.crc.testing:6443 \ Nov 22 07:56:51 crc kubenswrapper[4789]: --disable-webhook \ Nov 22 07:56:51 crc kubenswrapper[4789]: --csr-acceptance-conditions="/var/run/ovnkube-identity-config/additional-cert-acceptance-cond.json" \ Nov 22 07:56:51 crc kubenswrapper[4789]: --loglevel="${LOGLEVEL}" Nov 22 07:56:51 crc kubenswrapper[4789]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOGLEVEL,Value:4,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:env-overrides,ReadOnly:false,MountPath:/env,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovnkube-identity-cm,ReadOnly:false,MountPath:/var/run/ovnkube-identity-config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s2kz5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000470000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-node-identity-vrzqb_openshift-network-node-identity(ef543e1b-8068-4ea3-b32a-61027b32e95d): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Nov 22 07:56:51 crc kubenswrapper[4789]: > logger="UnhandledError" Nov 22 07:56:51 crc kubenswrapper[4789]: E1122 07:56:51.081111 4789 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:iptables-alerter,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2,Command:[/iptables-alerter/iptables-alerter.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONTAINER_RUNTIME_ENDPOINT,Value:unix:///run/crio/crio.sock,ValueFrom:nil,},EnvVar{Name:ALERTER_POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{68157440 0} {} 65Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:iptables-alerter-script,ReadOnly:false,MountPath:/iptables-alerter,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-slash,ReadOnly:true,MountPath:/host,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rczfb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod iptables-alerter-4ln5h_openshift-network-operator(d75a4c96-2883-4a0b-bab2-0fab2b6c0b49): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars" logger="UnhandledError" Nov 22 07:56:51 crc kubenswrapper[4789]: E1122 07:56:51.082316 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"webhook\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\", failed to \"StartContainer\" for \"approver\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"]" pod="openshift-network-node-identity/network-node-identity-vrzqb" podUID="ef543e1b-8068-4ea3-b32a-61027b32e95d" Nov 22 07:56:51 crc kubenswrapper[4789]: E1122 07:56:51.082366 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"iptables-alerter\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-network-operator/iptables-alerter-4ln5h" podUID="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.103102 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.103477 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.104610 4789 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="098c31b822e7c2351f1142af36b942d67ff8262896e28022c96a187d40bd0ddb" exitCode=255 Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.104666 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"098c31b822e7c2351f1142af36b942d67ff8262896e28022c96a187d40bd0ddb"} Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.104740 4789 scope.go:117] "RemoveContainer" containerID="b0233c87ba1f1a4ee0ab1214fc45db43ad6122e128c5890cf29134b0e0217ce6" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.105310 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"e2d596dffb054102dab0e4b677138df4a733961307d2b729f4c2f07948a39536"} Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.106550 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"ebcf819aa9b8405ea57d406d6adcb348956a49a49cea32efce2eaf02ee7a2d3c"} Nov 22 07:56:51 crc kubenswrapper[4789]: E1122 07:56:51.106695 4789 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:iptables-alerter,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2,Command:[/iptables-alerter/iptables-alerter.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONTAINER_RUNTIME_ENDPOINT,Value:unix:///run/crio/crio.sock,ValueFrom:nil,},EnvVar{Name:ALERTER_POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{68157440 0} {} 65Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:iptables-alerter-script,ReadOnly:false,MountPath:/iptables-alerter,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-slash,ReadOnly:true,MountPath:/host,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rczfb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod iptables-alerter-4ln5h_openshift-network-operator(d75a4c96-2883-4a0b-bab2-0fab2b6c0b49): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars" logger="UnhandledError" Nov 22 07:56:51 crc kubenswrapper[4789]: E1122 07:56:51.107827 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"iptables-alerter\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-network-operator/iptables-alerter-4ln5h" podUID="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.108613 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"fee5ae2388b10b7f1ade25844ab584a23d46da19737aeb88e4e9622c7ad1e7f3"} Nov 22 07:56:51 crc kubenswrapper[4789]: E1122 07:56:51.110559 4789 kuberuntime_manager.go:1274] "Unhandled Error" err=< Nov 22 07:56:51 crc kubenswrapper[4789]: container &Container{Name:webhook,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,Command:[/bin/bash -c set -xe Nov 22 07:56:51 crc kubenswrapper[4789]: if [[ -f "/env/_master" ]]; then Nov 22 07:56:51 crc kubenswrapper[4789]: set -o allexport Nov 22 07:56:51 crc kubenswrapper[4789]: source "/env/_master" Nov 22 07:56:51 crc kubenswrapper[4789]: set +o allexport Nov 22 07:56:51 crc kubenswrapper[4789]: fi Nov 22 07:56:51 crc kubenswrapper[4789]: # OVN-K will try to remove hybrid overlay node annotations even when the hybrid overlay is not enabled. Nov 22 07:56:51 crc kubenswrapper[4789]: # https://github.com/ovn-org/ovn-kubernetes/blob/ac6820df0b338a246f10f412cd5ec903bd234694/go-controller/pkg/ovn/master.go#L791 Nov 22 07:56:51 crc kubenswrapper[4789]: ho_enable="--enable-hybrid-overlay" Nov 22 07:56:51 crc kubenswrapper[4789]: echo "I$(date "+%m%d %H:%M:%S.%N") - network-node-identity - start webhook" Nov 22 07:56:51 crc kubenswrapper[4789]: # extra-allowed-user: service account `ovn-kubernetes-control-plane` Nov 22 07:56:51 crc kubenswrapper[4789]: # sets pod annotations in multi-homing layer3 network controller (cluster-manager) Nov 22 07:56:51 crc kubenswrapper[4789]: exec /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.crc.testing:6443 \ Nov 22 07:56:51 crc kubenswrapper[4789]: --webhook-cert-dir="/etc/webhook-cert" \ Nov 22 07:56:51 crc kubenswrapper[4789]: --webhook-host=127.0.0.1 \ Nov 22 07:56:51 crc kubenswrapper[4789]: --webhook-port=9743 \ Nov 22 07:56:51 crc kubenswrapper[4789]: ${ho_enable} \ Nov 22 07:56:51 crc kubenswrapper[4789]: --enable-interconnect \ Nov 22 07:56:51 crc kubenswrapper[4789]: --disable-approver \ Nov 22 07:56:51 crc kubenswrapper[4789]: --extra-allowed-user="system:serviceaccount:openshift-ovn-kubernetes:ovn-kubernetes-control-plane" \ Nov 22 07:56:51 crc kubenswrapper[4789]: --wait-for-kubernetes-api=200s \ Nov 22 07:56:51 crc kubenswrapper[4789]: --pod-admission-conditions="/var/run/ovnkube-identity-config/additional-pod-admission-cond.json" \ Nov 22 07:56:51 crc kubenswrapper[4789]: --loglevel="${LOGLEVEL}" Nov 22 07:56:51 crc kubenswrapper[4789]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOGLEVEL,Value:2,ValueFrom:nil,},EnvVar{Name:KUBERNETES_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:webhook-cert,ReadOnly:false,MountPath:/etc/webhook-cert/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:env-overrides,ReadOnly:false,MountPath:/env,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovnkube-identity-cm,ReadOnly:false,MountPath:/var/run/ovnkube-identity-config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s2kz5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000470000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-node-identity-vrzqb_openshift-network-node-identity(ef543e1b-8068-4ea3-b32a-61027b32e95d): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Nov 22 07:56:51 crc kubenswrapper[4789]: > logger="UnhandledError" Nov 22 07:56:51 crc kubenswrapper[4789]: E1122 07:56:51.110674 4789 kuberuntime_manager.go:1274] "Unhandled Error" err=< Nov 22 07:56:51 crc kubenswrapper[4789]: container &Container{Name:network-operator,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,Command:[/bin/bash -c #!/bin/bash Nov 22 07:56:51 crc kubenswrapper[4789]: set -o allexport Nov 22 07:56:51 crc kubenswrapper[4789]: if [[ -f /etc/kubernetes/apiserver-url.env ]]; then Nov 22 07:56:51 crc kubenswrapper[4789]: source /etc/kubernetes/apiserver-url.env Nov 22 07:56:51 crc kubenswrapper[4789]: else Nov 22 07:56:51 crc kubenswrapper[4789]: echo "Error: /etc/kubernetes/apiserver-url.env is missing" Nov 22 07:56:51 crc kubenswrapper[4789]: exit 1 Nov 22 07:56:51 crc kubenswrapper[4789]: fi Nov 22 07:56:51 crc kubenswrapper[4789]: exec /usr/bin/cluster-network-operator start --listen=0.0.0.0:9104 Nov 22 07:56:51 crc kubenswrapper[4789]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:cno,HostPort:9104,ContainerPort:9104,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:RELEASE_VERSION,Value:4.18.1,ValueFrom:nil,},EnvVar{Name:KUBE_PROXY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b97554198294bf544fbc116c94a0a1fb2ec8a4de0e926bf9d9e320135f0bee6f,ValueFrom:nil,},EnvVar{Name:KUBE_RBAC_PROXY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09,ValueFrom:nil,},EnvVar{Name:MULTUS_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26,ValueFrom:nil,},EnvVar{Name:MULTUS_ADMISSION_CONTROLLER_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317,ValueFrom:nil,},EnvVar{Name:CNI_PLUGINS_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc,ValueFrom:nil,},EnvVar{Name:BOND_CNI_PLUGIN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78,ValueFrom:nil,},EnvVar{Name:WHEREABOUTS_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4,ValueFrom:nil,},EnvVar{Name:ROUTE_OVERRRIDE_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa,ValueFrom:nil,},EnvVar{Name:MULTUS_NETWORKPOLICY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:23f833d3738d68706eb2f2868bd76bd71cee016cffa6faf5f045a60cc8c6eddd,ValueFrom:nil,},EnvVar{Name:OVN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,ValueFrom:nil,},EnvVar{Name:OVN_NB_RAFT_ELECTION_TIMER,Value:10,ValueFrom:nil,},EnvVar{Name:OVN_SB_RAFT_ELECTION_TIMER,Value:16,ValueFrom:nil,},EnvVar{Name:OVN_NORTHD_PROBE_INTERVAL,Value:10000,ValueFrom:nil,},EnvVar{Name:OVN_CONTROLLER_INACTIVITY_PROBE,Value:180000,ValueFrom:nil,},EnvVar{Name:OVN_NB_INACTIVITY_PROBE,Value:60000,ValueFrom:nil,},EnvVar{Name:EGRESS_ROUTER_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c,ValueFrom:nil,},EnvVar{Name:NETWORK_METRICS_DAEMON_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d,ValueFrom:nil,},EnvVar{Name:NETWORK_CHECK_SOURCE_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,ValueFrom:nil,},EnvVar{Name:NETWORK_CHECK_TARGET_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,ValueFrom:nil,},EnvVar{Name:NETWORK_OPERATOR_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,ValueFrom:nil,},EnvVar{Name:CLOUD_NETWORK_CONFIG_CONTROLLER_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8048f1cb0be521f09749c0a489503cd56d85b68c6ca93380e082cfd693cd97a8,ValueFrom:nil,},EnvVar{Name:CLI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2,ValueFrom:nil,},EnvVar{Name:FRR_K8S_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5dbf844e49bb46b78586930149e5e5f5dc121014c8afd10fe36f3651967cc256,ValueFrom:nil,},EnvVar{Name:NETWORKING_CONSOLE_PLUGIN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd,ValueFrom:nil,},EnvVar{Name:POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:host-etc-kube,ReadOnly:true,MountPath:/etc/kubernetes,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:metrics-tls,ReadOnly:false,MountPath:/var/run/secrets/serving-cert,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rdwmf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-operator-58b4c7f79c-55gtf_openshift-network-operator(37a5e44f-9a88-4405-be8a-b645485e7312): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Nov 22 07:56:51 crc kubenswrapper[4789]: > logger="UnhandledError" Nov 22 07:56:51 crc kubenswrapper[4789]: E1122 07:56:51.111949 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"network-operator\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" podUID="37a5e44f-9a88-4405-be8a-b645485e7312" Nov 22 07:56:51 crc kubenswrapper[4789]: E1122 07:56:51.112533 4789 kuberuntime_manager.go:1274] "Unhandled Error" err=< Nov 22 07:56:51 crc kubenswrapper[4789]: container &Container{Name:approver,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,Command:[/bin/bash -c set -xe Nov 22 07:56:51 crc kubenswrapper[4789]: if [[ -f "/env/_master" ]]; then Nov 22 07:56:51 crc kubenswrapper[4789]: set -o allexport Nov 22 07:56:51 crc kubenswrapper[4789]: source "/env/_master" Nov 22 07:56:51 crc kubenswrapper[4789]: set +o allexport Nov 22 07:56:51 crc kubenswrapper[4789]: fi Nov 22 07:56:51 crc kubenswrapper[4789]: Nov 22 07:56:51 crc kubenswrapper[4789]: echo "I$(date "+%m%d %H:%M:%S.%N") - network-node-identity - start approver" Nov 22 07:56:51 crc kubenswrapper[4789]: exec /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.crc.testing:6443 \ Nov 22 07:56:51 crc kubenswrapper[4789]: --disable-webhook \ Nov 22 07:56:51 crc kubenswrapper[4789]: --csr-acceptance-conditions="/var/run/ovnkube-identity-config/additional-cert-acceptance-cond.json" \ Nov 22 07:56:51 crc kubenswrapper[4789]: --loglevel="${LOGLEVEL}" Nov 22 07:56:51 crc kubenswrapper[4789]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOGLEVEL,Value:4,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:env-overrides,ReadOnly:false,MountPath:/env,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovnkube-identity-cm,ReadOnly:false,MountPath:/var/run/ovnkube-identity-config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s2kz5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000470000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-node-identity-vrzqb_openshift-network-node-identity(ef543e1b-8068-4ea3-b32a-61027b32e95d): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Nov 22 07:56:51 crc kubenswrapper[4789]: > logger="UnhandledError" Nov 22 07:56:51 crc kubenswrapper[4789]: E1122 07:56:51.113917 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"webhook\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\", failed to \"StartContainer\" for \"approver\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"]" pod="openshift-network-node-identity/network-node-identity-vrzqb" podUID="ef543e1b-8068-4ea3-b32a-61027b32e95d" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.119765 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.172993 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.192966 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.204560 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.213119 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.216640 4789 scope.go:117] "RemoveContainer" containerID="098c31b822e7c2351f1142af36b942d67ff8262896e28022c96a187d40bd0ddb" Nov 22 07:56:51 crc kubenswrapper[4789]: E1122 07:56:51.216849 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.219103 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.226180 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.238056 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93be48ce-7d87-448b-b03e-ab6736cfc34c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba91466edcf49f73751622f98632b3040a046ccd8a42ba8ecee7ab08688a8c24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b6666e0ce42a46f669505659298a19931b1bf7c253ae6a97c6acfd80a9e717f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://251e72e1cd0f3e0ee195bacf33ab38765d1dd943adf4395f9d6dd282bda024cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://098c31b822e7c2351f1142af36b942d67ff8262896e28022c96a187d40bd0ddb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b0233c87ba1f1a4ee0ab1214fc45db43ad6122e128c5890cf29134b0e0217ce6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T07:56:44Z\\\",\\\"message\\\":\\\"W1122 07:56:33.475056 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1122 07:56:33.475654 1 crypto.go:601] Generating new CA for check-endpoints-signer@1763798193 cert, and key in /tmp/serving-cert-1893313301/serving-signer.crt, /tmp/serving-cert-1893313301/serving-signer.key\\\\nI1122 07:56:33.744277 1 observer_polling.go:159] Starting file observer\\\\nW1122 07:56:33.747344 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1122 07:56:33.747899 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 07:56:33.749500 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1893313301/tls.crt::/tmp/serving-cert-1893313301/tls.key\\\\\\\"\\\\nF1122 07:56:44.205890 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T07:56:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://098c31b822e7c2351f1142af36b942d67ff8262896e28022c96a187d40bd0ddb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"message\\\":\\\"le observer\\\\nW1122 07:56:50.641735 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1122 07:56:50.642792 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 07:56:50.643945 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3588433834/tls.crt::/tmp/serving-cert-3588433834/tls.key\\\\\\\"\\\\nI1122 07:56:50.949040 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 07:56:50.952068 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 07:56:50.952092 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 07:56:50.952116 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 07:56:50.952121 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 07:56:50.958549 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1122 07:56:50.958565 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1122 07:56:50.958579 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 07:56:50.958585 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 07:56:50.958591 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 07:56:50.958596 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1122 07:56:50.958600 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 07:56:50.958605 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1122 07:56:50.962189 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T07:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://263dce4c3f4f3fe388b21d4bb55a9fc7d956c3cca7da34ee00f470d88ef501f1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac2b567fcebc517d3fb4366aebce03351183ea03ce62f86c2abdf5b95e20b34a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac2b567fcebc517d3fb4366aebce03351183ea03ce62f86c2abdf5b95e20b34a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T07:56:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T07:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.248971 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.258221 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.267884 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.276973 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.285587 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.288485 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 07:56:51 crc kubenswrapper[4789]: E1122 07:56:51.288642 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 07:56:52.28860995 +0000 UTC m=+26.523010223 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.294951 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.389507 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.389599 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.389629 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.389653 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 07:56:51 crc kubenswrapper[4789]: E1122 07:56:51.389740 4789 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 22 07:56:51 crc kubenswrapper[4789]: E1122 07:56:51.389817 4789 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 22 07:56:51 crc kubenswrapper[4789]: E1122 07:56:51.389842 4789 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 22 07:56:51 crc kubenswrapper[4789]: E1122 07:56:51.389857 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-22 07:56:52.38983703 +0000 UTC m=+26.624237303 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 22 07:56:51 crc kubenswrapper[4789]: E1122 07:56:51.389862 4789 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 22 07:56:51 crc kubenswrapper[4789]: E1122 07:56:51.389862 4789 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 22 07:56:51 crc kubenswrapper[4789]: E1122 07:56:51.389900 4789 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 22 07:56:51 crc kubenswrapper[4789]: E1122 07:56:51.389918 4789 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 07:56:51 crc kubenswrapper[4789]: E1122 07:56:51.389923 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-22 07:56:52.389902682 +0000 UTC m=+26.624302955 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 22 07:56:51 crc kubenswrapper[4789]: E1122 07:56:51.389881 4789 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 07:56:51 crc kubenswrapper[4789]: E1122 07:56:51.389960 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-22 07:56:52.389947403 +0000 UTC m=+26.624347886 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 07:56:51 crc kubenswrapper[4789]: E1122 07:56:51.389979 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-22 07:56:52.389972784 +0000 UTC m=+26.624373057 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.574684 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.576283 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.576326 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.576334 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.576391 4789 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.583233 4789 kubelet_node_status.go:115] "Node was previously registered" node="crc" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.583323 4789 kubelet_node_status.go:79] "Successfully registered node" node="crc" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.584334 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.584365 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.584386 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.584402 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.584412 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:56:51Z","lastTransitionTime":"2025-11-22T07:56:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:56:51 crc kubenswrapper[4789]: E1122 07:56:51.598938 4789 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"06b9c1b2-d063-4af7-880e-0654164bb896\\\",\\\"systemUUID\\\":\\\"fd8856dd-ccc1-49f3-a39b-e4a6f9389fc3\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.602459 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.602516 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.602531 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.602554 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.602570 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:56:51Z","lastTransitionTime":"2025-11-22T07:56:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:56:51 crc kubenswrapper[4789]: E1122 07:56:51.611254 4789 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"06b9c1b2-d063-4af7-880e-0654164bb896\\\",\\\"systemUUID\\\":\\\"fd8856dd-ccc1-49f3-a39b-e4a6f9389fc3\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.614409 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.614456 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.614466 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.614482 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.614493 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:56:51Z","lastTransitionTime":"2025-11-22T07:56:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:56:51 crc kubenswrapper[4789]: E1122 07:56:51.622634 4789 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"06b9c1b2-d063-4af7-880e-0654164bb896\\\",\\\"systemUUID\\\":\\\"fd8856dd-ccc1-49f3-a39b-e4a6f9389fc3\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.626322 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.626362 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.626371 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.626404 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.626414 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:56:51Z","lastTransitionTime":"2025-11-22T07:56:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:56:51 crc kubenswrapper[4789]: E1122 07:56:51.635739 4789 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"06b9c1b2-d063-4af7-880e-0654164bb896\\\",\\\"systemUUID\\\":\\\"fd8856dd-ccc1-49f3-a39b-e4a6f9389fc3\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.639789 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.639858 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.639870 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.639890 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.639903 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:56:51Z","lastTransitionTime":"2025-11-22T07:56:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:56:51 crc kubenswrapper[4789]: E1122 07:56:51.648893 4789 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"06b9c1b2-d063-4af7-880e-0654164bb896\\\",\\\"systemUUID\\\":\\\"fd8856dd-ccc1-49f3-a39b-e4a6f9389fc3\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:51 crc kubenswrapper[4789]: E1122 07:56:51.649023 4789 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.650793 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.650844 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.650857 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.650876 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.650888 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:56:51Z","lastTransitionTime":"2025-11-22T07:56:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.752822 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.752871 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.752882 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.752898 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.752909 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:56:51Z","lastTransitionTime":"2025-11-22T07:56:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.856102 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.856186 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.856199 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.856218 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.856230 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:56:51Z","lastTransitionTime":"2025-11-22T07:56:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.890736 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-9f74f"] Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.891009 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-pxvrm"] Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.891125 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-9f74f" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.891536 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.892970 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.893080 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.893081 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.893202 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.893167 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.893324 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.893886 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.899039 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.903192 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.912883 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.919876 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9f74f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"29ad5e00-4918-4718-9d49-013b12621f7d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mnz6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9f74f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.929511 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93be48ce-7d87-448b-b03e-ab6736cfc34c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba91466edcf49f73751622f98632b3040a046ccd8a42ba8ecee7ab08688a8c24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b6666e0ce42a46f669505659298a19931b1bf7c253ae6a97c6acfd80a9e717f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://251e72e1cd0f3e0ee195bacf33ab38765d1dd943adf4395f9d6dd282bda024cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://098c31b822e7c2351f1142af36b942d67ff8262896e28022c96a187d40bd0ddb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b0233c87ba1f1a4ee0ab1214fc45db43ad6122e128c5890cf29134b0e0217ce6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T07:56:44Z\\\",\\\"message\\\":\\\"W1122 07:56:33.475056 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1122 07:56:33.475654 1 crypto.go:601] Generating new CA for check-endpoints-signer@1763798193 cert, and key in /tmp/serving-cert-1893313301/serving-signer.crt, /tmp/serving-cert-1893313301/serving-signer.key\\\\nI1122 07:56:33.744277 1 observer_polling.go:159] Starting file observer\\\\nW1122 07:56:33.747344 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1122 07:56:33.747899 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 07:56:33.749500 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1893313301/tls.crt::/tmp/serving-cert-1893313301/tls.key\\\\\\\"\\\\nF1122 07:56:44.205890 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T07:56:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://098c31b822e7c2351f1142af36b942d67ff8262896e28022c96a187d40bd0ddb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"message\\\":\\\"le observer\\\\nW1122 07:56:50.641735 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1122 07:56:50.642792 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 07:56:50.643945 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3588433834/tls.crt::/tmp/serving-cert-3588433834/tls.key\\\\\\\"\\\\nI1122 07:56:50.949040 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 07:56:50.952068 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 07:56:50.952092 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 07:56:50.952116 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 07:56:50.952121 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 07:56:50.958549 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1122 07:56:50.958565 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1122 07:56:50.958579 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 07:56:50.958585 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 07:56:50.958591 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 07:56:50.958596 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1122 07:56:50.958600 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 07:56:50.958605 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1122 07:56:50.962189 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T07:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://263dce4c3f4f3fe388b21d4bb55a9fc7d956c3cca7da34ee00f470d88ef501f1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac2b567fcebc517d3fb4366aebce03351183ea03ce62f86c2abdf5b95e20b34a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac2b567fcebc517d3fb4366aebce03351183ea03ce62f86c2abdf5b95e20b34a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T07:56:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T07:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.937427 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.946019 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.954502 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.958021 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.958053 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.958061 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.958074 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.958083 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:56:51Z","lastTransitionTime":"2025-11-22T07:56:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.962639 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.964239 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 07:56:51 crc kubenswrapper[4789]: E1122 07:56:51.964348 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.968921 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.969895 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.971645 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.971977 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.972366 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.973342 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.973917 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.974681 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.975801 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.976427 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.977421 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.978024 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.979294 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.979887 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9f74f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"29ad5e00-4918-4718-9d49-013b12621f7d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mnz6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9f74f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.979969 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.980626 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.981609 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.982126 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.983072 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.983440 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.984106 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.985096 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.985520 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.986451 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.986978 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.987944 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.988324 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.988903 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.989919 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.990352 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.991284 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.991706 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.992599 4789 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.992697 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.992650 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93be48ce-7d87-448b-b03e-ab6736cfc34c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba91466edcf49f73751622f98632b3040a046ccd8a42ba8ecee7ab08688a8c24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b6666e0ce42a46f669505659298a19931b1bf7c253ae6a97c6acfd80a9e717f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://251e72e1cd0f3e0ee195bacf33ab38765d1dd943adf4395f9d6dd282bda024cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://098c31b822e7c2351f1142af36b942d67ff8262896e28022c96a187d40bd0ddb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b0233c87ba1f1a4ee0ab1214fc45db43ad6122e128c5890cf29134b0e0217ce6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T07:56:44Z\\\",\\\"message\\\":\\\"W1122 07:56:33.475056 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1122 07:56:33.475654 1 crypto.go:601] Generating new CA for check-endpoints-signer@1763798193 cert, and key in /tmp/serving-cert-1893313301/serving-signer.crt, /tmp/serving-cert-1893313301/serving-signer.key\\\\nI1122 07:56:33.744277 1 observer_polling.go:159] Starting file observer\\\\nW1122 07:56:33.747344 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1122 07:56:33.747899 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 07:56:33.749500 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1893313301/tls.crt::/tmp/serving-cert-1893313301/tls.key\\\\\\\"\\\\nF1122 07:56:44.205890 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T07:56:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://098c31b822e7c2351f1142af36b942d67ff8262896e28022c96a187d40bd0ddb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"message\\\":\\\"le observer\\\\nW1122 07:56:50.641735 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1122 07:56:50.642792 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 07:56:50.643945 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3588433834/tls.crt::/tmp/serving-cert-3588433834/tls.key\\\\\\\"\\\\nI1122 07:56:50.949040 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 07:56:50.952068 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 07:56:50.952092 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 07:56:50.952116 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 07:56:50.952121 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 07:56:50.958549 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1122 07:56:50.958565 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1122 07:56:50.958579 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 07:56:50.958585 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 07:56:50.958591 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 07:56:50.958596 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1122 07:56:50.958600 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 07:56:50.958605 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1122 07:56:50.962189 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T07:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://263dce4c3f4f3fe388b21d4bb55a9fc7d956c3cca7da34ee00f470d88ef501f1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac2b567fcebc517d3fb4366aebce03351183ea03ce62f86c2abdf5b95e20b34a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac2b567fcebc517d3fb4366aebce03351183ea03ce62f86c2abdf5b95e20b34a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T07:56:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T07:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.993870 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/58aa8071-7ecc-4692-8789-2db04ec70647-proxy-tls\") pod \"machine-config-daemon-pxvrm\" (UID: \"58aa8071-7ecc-4692-8789-2db04ec70647\") " pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.993906 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/58aa8071-7ecc-4692-8789-2db04ec70647-rootfs\") pod \"machine-config-daemon-pxvrm\" (UID: \"58aa8071-7ecc-4692-8789-2db04ec70647\") " pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.993938 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/58aa8071-7ecc-4692-8789-2db04ec70647-mcd-auth-proxy-config\") pod \"machine-config-daemon-pxvrm\" (UID: \"58aa8071-7ecc-4692-8789-2db04ec70647\") " pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.993975 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bfjgw\" (UniqueName: \"kubernetes.io/projected/58aa8071-7ecc-4692-8789-2db04ec70647-kube-api-access-bfjgw\") pod \"machine-config-daemon-pxvrm\" (UID: \"58aa8071-7ecc-4692-8789-2db04ec70647\") " pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.994002 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/29ad5e00-4918-4718-9d49-013b12621f7d-hosts-file\") pod \"node-resolver-9f74f\" (UID: \"29ad5e00-4918-4718-9d49-013b12621f7d\") " pod="openshift-dns/node-resolver-9f74f" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.994034 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mnz6k\" (UniqueName: \"kubernetes.io/projected/29ad5e00-4918-4718-9d49-013b12621f7d-kube-api-access-mnz6k\") pod \"node-resolver-9f74f\" (UID: \"29ad5e00-4918-4718-9d49-013b12621f7d\") " pod="openshift-dns/node-resolver-9f74f" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.994310 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.995187 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.995573 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.996966 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.997583 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.998443 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Nov 22 07:56:51 crc kubenswrapper[4789]: I1122 07:56:51.999052 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.000019 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.000440 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.001402 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.002042 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.002157 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.003380 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.003899 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.004815 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.005295 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.006512 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.007132 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.007963 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.008402 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.009355 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.009914 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.010349 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.011801 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.020851 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.030461 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.040020 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.049026 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58aa8071-7ecc-4692-8789-2db04ec70647\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfjgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfjgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pxvrm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.061337 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.061385 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.061397 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.061415 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.061427 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:56:52Z","lastTransitionTime":"2025-11-22T07:56:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.094937 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/29ad5e00-4918-4718-9d49-013b12621f7d-hosts-file\") pod \"node-resolver-9f74f\" (UID: \"29ad5e00-4918-4718-9d49-013b12621f7d\") " pod="openshift-dns/node-resolver-9f74f" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.095012 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mnz6k\" (UniqueName: \"kubernetes.io/projected/29ad5e00-4918-4718-9d49-013b12621f7d-kube-api-access-mnz6k\") pod \"node-resolver-9f74f\" (UID: \"29ad5e00-4918-4718-9d49-013b12621f7d\") " pod="openshift-dns/node-resolver-9f74f" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.095036 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/58aa8071-7ecc-4692-8789-2db04ec70647-proxy-tls\") pod \"machine-config-daemon-pxvrm\" (UID: \"58aa8071-7ecc-4692-8789-2db04ec70647\") " pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.095061 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/58aa8071-7ecc-4692-8789-2db04ec70647-rootfs\") pod \"machine-config-daemon-pxvrm\" (UID: \"58aa8071-7ecc-4692-8789-2db04ec70647\") " pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.095084 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/58aa8071-7ecc-4692-8789-2db04ec70647-mcd-auth-proxy-config\") pod \"machine-config-daemon-pxvrm\" (UID: \"58aa8071-7ecc-4692-8789-2db04ec70647\") " pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.095092 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/29ad5e00-4918-4718-9d49-013b12621f7d-hosts-file\") pod \"node-resolver-9f74f\" (UID: \"29ad5e00-4918-4718-9d49-013b12621f7d\") " pod="openshift-dns/node-resolver-9f74f" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.095125 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bfjgw\" (UniqueName: \"kubernetes.io/projected/58aa8071-7ecc-4692-8789-2db04ec70647-kube-api-access-bfjgw\") pod \"machine-config-daemon-pxvrm\" (UID: \"58aa8071-7ecc-4692-8789-2db04ec70647\") " pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.095226 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/58aa8071-7ecc-4692-8789-2db04ec70647-rootfs\") pod \"machine-config-daemon-pxvrm\" (UID: \"58aa8071-7ecc-4692-8789-2db04ec70647\") " pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.095965 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/58aa8071-7ecc-4692-8789-2db04ec70647-mcd-auth-proxy-config\") pod \"machine-config-daemon-pxvrm\" (UID: \"58aa8071-7ecc-4692-8789-2db04ec70647\") " pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.099729 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/58aa8071-7ecc-4692-8789-2db04ec70647-proxy-tls\") pod \"machine-config-daemon-pxvrm\" (UID: \"58aa8071-7ecc-4692-8789-2db04ec70647\") " pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.112244 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mnz6k\" (UniqueName: \"kubernetes.io/projected/29ad5e00-4918-4718-9d49-013b12621f7d-kube-api-access-mnz6k\") pod \"node-resolver-9f74f\" (UID: \"29ad5e00-4918-4718-9d49-013b12621f7d\") " pod="openshift-dns/node-resolver-9f74f" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.112281 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bfjgw\" (UniqueName: \"kubernetes.io/projected/58aa8071-7ecc-4692-8789-2db04ec70647-kube-api-access-bfjgw\") pod \"machine-config-daemon-pxvrm\" (UID: \"58aa8071-7ecc-4692-8789-2db04ec70647\") " pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.112846 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.116257 4789 scope.go:117] "RemoveContainer" containerID="098c31b822e7c2351f1142af36b942d67ff8262896e28022c96a187d40bd0ddb" Nov 22 07:56:52 crc kubenswrapper[4789]: E1122 07:56:52.116430 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.126243 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.138139 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.148513 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.155719 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9f74f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"29ad5e00-4918-4718-9d49-013b12621f7d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mnz6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9f74f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.164332 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.164377 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.164387 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.164402 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.164415 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:56:52Z","lastTransitionTime":"2025-11-22T07:56:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.167017 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93be48ce-7d87-448b-b03e-ab6736cfc34c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba91466edcf49f73751622f98632b3040a046ccd8a42ba8ecee7ab08688a8c24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b6666e0ce42a46f669505659298a19931b1bf7c253ae6a97c6acfd80a9e717f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://251e72e1cd0f3e0ee195bacf33ab38765d1dd943adf4395f9d6dd282bda024cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://098c31b822e7c2351f1142af36b942d67ff8262896e28022c96a187d40bd0ddb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://098c31b822e7c2351f1142af36b942d67ff8262896e28022c96a187d40bd0ddb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"message\\\":\\\"le observer\\\\nW1122 07:56:50.641735 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1122 07:56:50.642792 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 07:56:50.643945 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3588433834/tls.crt::/tmp/serving-cert-3588433834/tls.key\\\\\\\"\\\\nI1122 07:56:50.949040 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 07:56:50.952068 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 07:56:50.952092 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 07:56:50.952116 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 07:56:50.952121 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 07:56:50.958549 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1122 07:56:50.958565 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1122 07:56:50.958579 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 07:56:50.958585 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 07:56:50.958591 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 07:56:50.958596 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1122 07:56:50.958600 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 07:56:50.958605 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1122 07:56:50.962189 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T07:56:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://263dce4c3f4f3fe388b21d4bb55a9fc7d956c3cca7da34ee00f470d88ef501f1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac2b567fcebc517d3fb4366aebce03351183ea03ce62f86c2abdf5b95e20b34a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac2b567fcebc517d3fb4366aebce03351183ea03ce62f86c2abdf5b95e20b34a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T07:56:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T07:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.196916 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.204954 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-9f74f" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.212974 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.215388 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:52 crc kubenswrapper[4789]: W1122 07:56:52.228720 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod29ad5e00_4918_4718_9d49_013b12621f7d.slice/crio-e6fb47fa3f3cc35dc53b3f5cc52b1b60e44be4b2de4afac3241e9a52dc8d4c60 WatchSource:0}: Error finding container e6fb47fa3f3cc35dc53b3f5cc52b1b60e44be4b2de4afac3241e9a52dc8d4c60: Status 404 returned error can't find the container with id e6fb47fa3f3cc35dc53b3f5cc52b1b60e44be4b2de4afac3241e9a52dc8d4c60 Nov 22 07:56:52 crc kubenswrapper[4789]: W1122 07:56:52.228999 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod58aa8071_7ecc_4692_8789_2db04ec70647.slice/crio-809d26d975296963367b04cb00e88486f5364de5d7effac9e2eca6cb08b0a5b6 WatchSource:0}: Error finding container 809d26d975296963367b04cb00e88486f5364de5d7effac9e2eca6cb08b0a5b6: Status 404 returned error can't find the container with id 809d26d975296963367b04cb00e88486f5364de5d7effac9e2eca6cb08b0a5b6 Nov 22 07:56:52 crc kubenswrapper[4789]: E1122 07:56:52.230491 4789 kuberuntime_manager.go:1274] "Unhandled Error" err=< Nov 22 07:56:52 crc kubenswrapper[4789]: container &Container{Name:dns-node-resolver,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2,Command:[/bin/bash -c #!/bin/bash Nov 22 07:56:52 crc kubenswrapper[4789]: set -uo pipefail Nov 22 07:56:52 crc kubenswrapper[4789]: Nov 22 07:56:52 crc kubenswrapper[4789]: trap 'jobs -p | xargs kill || true; wait; exit 0' TERM Nov 22 07:56:52 crc kubenswrapper[4789]: Nov 22 07:56:52 crc kubenswrapper[4789]: OPENSHIFT_MARKER="openshift-generated-node-resolver" Nov 22 07:56:52 crc kubenswrapper[4789]: HOSTS_FILE="/etc/hosts" Nov 22 07:56:52 crc kubenswrapper[4789]: TEMP_FILE="/etc/hosts.tmp" Nov 22 07:56:52 crc kubenswrapper[4789]: Nov 22 07:56:52 crc kubenswrapper[4789]: IFS=', ' read -r -a services <<< "${SERVICES}" Nov 22 07:56:52 crc kubenswrapper[4789]: Nov 22 07:56:52 crc kubenswrapper[4789]: # Make a temporary file with the old hosts file's attributes. Nov 22 07:56:52 crc kubenswrapper[4789]: if ! cp -f --attributes-only "${HOSTS_FILE}" "${TEMP_FILE}"; then Nov 22 07:56:52 crc kubenswrapper[4789]: echo "Failed to preserve hosts file. Exiting." Nov 22 07:56:52 crc kubenswrapper[4789]: exit 1 Nov 22 07:56:52 crc kubenswrapper[4789]: fi Nov 22 07:56:52 crc kubenswrapper[4789]: Nov 22 07:56:52 crc kubenswrapper[4789]: while true; do Nov 22 07:56:52 crc kubenswrapper[4789]: declare -A svc_ips Nov 22 07:56:52 crc kubenswrapper[4789]: for svc in "${services[@]}"; do Nov 22 07:56:52 crc kubenswrapper[4789]: # Fetch service IP from cluster dns if present. We make several tries Nov 22 07:56:52 crc kubenswrapper[4789]: # to do it: IPv4, IPv6, IPv4 over TCP and IPv6 over TCP. The two last ones Nov 22 07:56:52 crc kubenswrapper[4789]: # are for deployments with Kuryr on older OpenStack (OSP13) - those do not Nov 22 07:56:52 crc kubenswrapper[4789]: # support UDP loadbalancers and require reaching DNS through TCP. Nov 22 07:56:52 crc kubenswrapper[4789]: cmds=('dig -t A @"${NAMESERVER}" +short "${svc}.${CLUSTER_DOMAIN}"|grep -v "^;"' Nov 22 07:56:52 crc kubenswrapper[4789]: 'dig -t AAAA @"${NAMESERVER}" +short "${svc}.${CLUSTER_DOMAIN}"|grep -v "^;"' Nov 22 07:56:52 crc kubenswrapper[4789]: 'dig -t A +tcp +retry=0 @"${NAMESERVER}" +short "${svc}.${CLUSTER_DOMAIN}"|grep -v "^;"' Nov 22 07:56:52 crc kubenswrapper[4789]: 'dig -t AAAA +tcp +retry=0 @"${NAMESERVER}" +short "${svc}.${CLUSTER_DOMAIN}"|grep -v "^;"') Nov 22 07:56:52 crc kubenswrapper[4789]: for i in ${!cmds[*]} Nov 22 07:56:52 crc kubenswrapper[4789]: do Nov 22 07:56:52 crc kubenswrapper[4789]: ips=($(eval "${cmds[i]}")) Nov 22 07:56:52 crc kubenswrapper[4789]: if [[ "$?" -eq 0 && "${#ips[@]}" -ne 0 ]]; then Nov 22 07:56:52 crc kubenswrapper[4789]: svc_ips["${svc}"]="${ips[@]}" Nov 22 07:56:52 crc kubenswrapper[4789]: break Nov 22 07:56:52 crc kubenswrapper[4789]: fi Nov 22 07:56:52 crc kubenswrapper[4789]: done Nov 22 07:56:52 crc kubenswrapper[4789]: done Nov 22 07:56:52 crc kubenswrapper[4789]: Nov 22 07:56:52 crc kubenswrapper[4789]: # Update /etc/hosts only if we get valid service IPs Nov 22 07:56:52 crc kubenswrapper[4789]: # We will not update /etc/hosts when there is coredns service outage or api unavailability Nov 22 07:56:52 crc kubenswrapper[4789]: # Stale entries could exist in /etc/hosts if the service is deleted Nov 22 07:56:52 crc kubenswrapper[4789]: if [[ -n "${svc_ips[*]-}" ]]; then Nov 22 07:56:52 crc kubenswrapper[4789]: # Build a new hosts file from /etc/hosts with our custom entries filtered out Nov 22 07:56:52 crc kubenswrapper[4789]: if ! sed --silent "/# ${OPENSHIFT_MARKER}/d; w ${TEMP_FILE}" "${HOSTS_FILE}"; then Nov 22 07:56:52 crc kubenswrapper[4789]: # Only continue rebuilding the hosts entries if its original content is preserved Nov 22 07:56:52 crc kubenswrapper[4789]: sleep 60 & wait Nov 22 07:56:52 crc kubenswrapper[4789]: continue Nov 22 07:56:52 crc kubenswrapper[4789]: fi Nov 22 07:56:52 crc kubenswrapper[4789]: Nov 22 07:56:52 crc kubenswrapper[4789]: # Append resolver entries for services Nov 22 07:56:52 crc kubenswrapper[4789]: rc=0 Nov 22 07:56:52 crc kubenswrapper[4789]: for svc in "${!svc_ips[@]}"; do Nov 22 07:56:52 crc kubenswrapper[4789]: for ip in ${svc_ips[${svc}]}; do Nov 22 07:56:52 crc kubenswrapper[4789]: echo "${ip} ${svc} ${svc}.${CLUSTER_DOMAIN} # ${OPENSHIFT_MARKER}" >> "${TEMP_FILE}" || rc=$? Nov 22 07:56:52 crc kubenswrapper[4789]: done Nov 22 07:56:52 crc kubenswrapper[4789]: done Nov 22 07:56:52 crc kubenswrapper[4789]: if [[ $rc -ne 0 ]]; then Nov 22 07:56:52 crc kubenswrapper[4789]: sleep 60 & wait Nov 22 07:56:52 crc kubenswrapper[4789]: continue Nov 22 07:56:52 crc kubenswrapper[4789]: fi Nov 22 07:56:52 crc kubenswrapper[4789]: Nov 22 07:56:52 crc kubenswrapper[4789]: Nov 22 07:56:52 crc kubenswrapper[4789]: # TODO: Update /etc/hosts atomically to avoid any inconsistent behavior Nov 22 07:56:52 crc kubenswrapper[4789]: # Replace /etc/hosts with our modified version if needed Nov 22 07:56:52 crc kubenswrapper[4789]: cmp "${TEMP_FILE}" "${HOSTS_FILE}" || cp -f "${TEMP_FILE}" "${HOSTS_FILE}" Nov 22 07:56:52 crc kubenswrapper[4789]: # TEMP_FILE is not removed to avoid file create/delete and attributes copy churn Nov 22 07:56:52 crc kubenswrapper[4789]: fi Nov 22 07:56:52 crc kubenswrapper[4789]: sleep 60 & wait Nov 22 07:56:52 crc kubenswrapper[4789]: unset svc_ips Nov 22 07:56:52 crc kubenswrapper[4789]: done Nov 22 07:56:52 crc kubenswrapper[4789]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:SERVICES,Value:image-registry.openshift-image-registry.svc,ValueFrom:nil,},EnvVar{Name:NAMESERVER,Value:10.217.4.10,ValueFrom:nil,},EnvVar{Name:CLUSTER_DOMAIN,Value:cluster.local,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{22020096 0} {} 21Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:hosts-file,ReadOnly:false,MountPath:/etc/hosts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mnz6k,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod node-resolver-9f74f_openshift-dns(29ad5e00-4918-4718-9d49-013b12621f7d): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Nov 22 07:56:52 crc kubenswrapper[4789]: > logger="UnhandledError" Nov 22 07:56:52 crc kubenswrapper[4789]: E1122 07:56:52.231159 4789 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:machine-config-daemon,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a,Command:[/usr/bin/machine-config-daemon],Args:[start --payload-version=4.18.1],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:health,HostPort:8798,ContainerPort:8798,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:rootfs,ReadOnly:false,MountPath:/rootfs,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bfjgw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/health,Port:{0 8798 },Host:127.0.0.1,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:120,TimeoutSeconds:1,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars" logger="UnhandledError" Nov 22 07:56:52 crc kubenswrapper[4789]: E1122 07:56:52.231850 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"dns-node-resolver\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-dns/node-resolver-9f74f" podUID="29ad5e00-4918-4718-9d49-013b12621f7d" Nov 22 07:56:52 crc kubenswrapper[4789]: E1122 07:56:52.233907 4789 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09,Command:[],Args:[--secure-listen-address=0.0.0.0:9001 --config-file=/etc/kube-rbac-proxy/config-file.yaml --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12 --upstream=http://127.0.0.1:8797 --logtostderr=true --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:9001,ContainerPort:9001,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:proxy-tls,ReadOnly:false,MountPath:/etc/tls/private,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:mcd-auth-proxy-config,ReadOnly:false,MountPath:/etc/kube-rbac-proxy,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bfjgw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars" logger="UnhandledError" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.235049 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:52 crc kubenswrapper[4789]: E1122 07:56:52.235218 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"machine-config-daemon\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"]" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.246831 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58aa8071-7ecc-4692-8789-2db04ec70647\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfjgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfjgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pxvrm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.256186 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-rks5j"] Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.256513 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-rks5j" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.256915 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-2wdp7"] Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.257466 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-2wdp7" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.258207 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.258650 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.258864 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.259079 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.260849 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.261003 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.262611 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.266270 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.266312 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.266320 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.266337 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.266347 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:56:52Z","lastTransitionTime":"2025-11-22T07:56:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.274161 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.284681 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.291869 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9f74f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"29ad5e00-4918-4718-9d49-013b12621f7d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mnz6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9f74f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.295890 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.295986 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/5613c209-59a0-4b8a-ac8b-5c633bc1bdc4-system-cni-dir\") pod \"multus-rks5j\" (UID: \"5613c209-59a0-4b8a-ac8b-5c633bc1bdc4\") " pod="openshift-multus/multus-rks5j" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.296016 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/5613c209-59a0-4b8a-ac8b-5c633bc1bdc4-multus-conf-dir\") pod \"multus-rks5j\" (UID: \"5613c209-59a0-4b8a-ac8b-5c633bc1bdc4\") " pod="openshift-multus/multus-rks5j" Nov 22 07:56:52 crc kubenswrapper[4789]: E1122 07:56:52.296036 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 07:56:54.296013901 +0000 UTC m=+28.530414174 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.296064 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/2fdee89f-559b-4f4e-8d20-6ecc3aafd22e-cni-binary-copy\") pod \"multus-additional-cni-plugins-2wdp7\" (UID: \"2fdee89f-559b-4f4e-8d20-6ecc3aafd22e\") " pod="openshift-multus/multus-additional-cni-plugins-2wdp7" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.296099 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/5613c209-59a0-4b8a-ac8b-5c633bc1bdc4-cnibin\") pod \"multus-rks5j\" (UID: \"5613c209-59a0-4b8a-ac8b-5c633bc1bdc4\") " pod="openshift-multus/multus-rks5j" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.296120 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/5613c209-59a0-4b8a-ac8b-5c633bc1bdc4-cni-binary-copy\") pod \"multus-rks5j\" (UID: \"5613c209-59a0-4b8a-ac8b-5c633bc1bdc4\") " pod="openshift-multus/multus-rks5j" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.296163 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/5613c209-59a0-4b8a-ac8b-5c633bc1bdc4-multus-cni-dir\") pod \"multus-rks5j\" (UID: \"5613c209-59a0-4b8a-ac8b-5c633bc1bdc4\") " pod="openshift-multus/multus-rks5j" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.296182 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/5613c209-59a0-4b8a-ac8b-5c633bc1bdc4-host-run-k8s-cni-cncf-io\") pod \"multus-rks5j\" (UID: \"5613c209-59a0-4b8a-ac8b-5c633bc1bdc4\") " pod="openshift-multus/multus-rks5j" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.296206 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/2fdee89f-559b-4f4e-8d20-6ecc3aafd22e-cnibin\") pod \"multus-additional-cni-plugins-2wdp7\" (UID: \"2fdee89f-559b-4f4e-8d20-6ecc3aafd22e\") " pod="openshift-multus/multus-additional-cni-plugins-2wdp7" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.296222 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/2fdee89f-559b-4f4e-8d20-6ecc3aafd22e-tuning-conf-dir\") pod \"multus-additional-cni-plugins-2wdp7\" (UID: \"2fdee89f-559b-4f4e-8d20-6ecc3aafd22e\") " pod="openshift-multus/multus-additional-cni-plugins-2wdp7" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.296238 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/2fdee89f-559b-4f4e-8d20-6ecc3aafd22e-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-2wdp7\" (UID: \"2fdee89f-559b-4f4e-8d20-6ecc3aafd22e\") " pod="openshift-multus/multus-additional-cni-plugins-2wdp7" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.296446 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/5613c209-59a0-4b8a-ac8b-5c633bc1bdc4-hostroot\") pod \"multus-rks5j\" (UID: \"5613c209-59a0-4b8a-ac8b-5c633bc1bdc4\") " pod="openshift-multus/multus-rks5j" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.296483 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/5613c209-59a0-4b8a-ac8b-5c633bc1bdc4-multus-daemon-config\") pod \"multus-rks5j\" (UID: \"5613c209-59a0-4b8a-ac8b-5c633bc1bdc4\") " pod="openshift-multus/multus-rks5j" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.296535 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/5613c209-59a0-4b8a-ac8b-5c633bc1bdc4-host-run-netns\") pod \"multus-rks5j\" (UID: \"5613c209-59a0-4b8a-ac8b-5c633bc1bdc4\") " pod="openshift-multus/multus-rks5j" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.296561 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/5613c209-59a0-4b8a-ac8b-5c633bc1bdc4-host-var-lib-kubelet\") pod \"multus-rks5j\" (UID: \"5613c209-59a0-4b8a-ac8b-5c633bc1bdc4\") " pod="openshift-multus/multus-rks5j" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.296581 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5613c209-59a0-4b8a-ac8b-5c633bc1bdc4-etc-kubernetes\") pod \"multus-rks5j\" (UID: \"5613c209-59a0-4b8a-ac8b-5c633bc1bdc4\") " pod="openshift-multus/multus-rks5j" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.296609 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/5613c209-59a0-4b8a-ac8b-5c633bc1bdc4-host-var-lib-cni-multus\") pod \"multus-rks5j\" (UID: \"5613c209-59a0-4b8a-ac8b-5c633bc1bdc4\") " pod="openshift-multus/multus-rks5j" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.296638 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/5613c209-59a0-4b8a-ac8b-5c633bc1bdc4-host-run-multus-certs\") pod \"multus-rks5j\" (UID: \"5613c209-59a0-4b8a-ac8b-5c633bc1bdc4\") " pod="openshift-multus/multus-rks5j" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.296653 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pxvhw\" (UniqueName: \"kubernetes.io/projected/2fdee89f-559b-4f4e-8d20-6ecc3aafd22e-kube-api-access-pxvhw\") pod \"multus-additional-cni-plugins-2wdp7\" (UID: \"2fdee89f-559b-4f4e-8d20-6ecc3aafd22e\") " pod="openshift-multus/multus-additional-cni-plugins-2wdp7" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.296676 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/5613c209-59a0-4b8a-ac8b-5c633bc1bdc4-host-var-lib-cni-bin\") pod \"multus-rks5j\" (UID: \"5613c209-59a0-4b8a-ac8b-5c633bc1bdc4\") " pod="openshift-multus/multus-rks5j" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.296696 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/2fdee89f-559b-4f4e-8d20-6ecc3aafd22e-system-cni-dir\") pod \"multus-additional-cni-plugins-2wdp7\" (UID: \"2fdee89f-559b-4f4e-8d20-6ecc3aafd22e\") " pod="openshift-multus/multus-additional-cni-plugins-2wdp7" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.296711 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/5613c209-59a0-4b8a-ac8b-5c633bc1bdc4-os-release\") pod \"multus-rks5j\" (UID: \"5613c209-59a0-4b8a-ac8b-5c633bc1bdc4\") " pod="openshift-multus/multus-rks5j" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.296725 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wj8sc\" (UniqueName: \"kubernetes.io/projected/5613c209-59a0-4b8a-ac8b-5c633bc1bdc4-kube-api-access-wj8sc\") pod \"multus-rks5j\" (UID: \"5613c209-59a0-4b8a-ac8b-5c633bc1bdc4\") " pod="openshift-multus/multus-rks5j" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.296740 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/2fdee89f-559b-4f4e-8d20-6ecc3aafd22e-os-release\") pod \"multus-additional-cni-plugins-2wdp7\" (UID: \"2fdee89f-559b-4f4e-8d20-6ecc3aafd22e\") " pod="openshift-multus/multus-additional-cni-plugins-2wdp7" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.296786 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/5613c209-59a0-4b8a-ac8b-5c633bc1bdc4-multus-socket-dir-parent\") pod \"multus-rks5j\" (UID: \"5613c209-59a0-4b8a-ac8b-5c633bc1bdc4\") " pod="openshift-multus/multus-rks5j" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.301711 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.311644 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.321510 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.329026 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58aa8071-7ecc-4692-8789-2db04ec70647\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfjgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfjgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pxvrm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.337773 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rks5j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5613c209-59a0-4b8a-ac8b-5c633bc1bdc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wj8sc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rks5j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.347204 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93be48ce-7d87-448b-b03e-ab6736cfc34c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba91466edcf49f73751622f98632b3040a046ccd8a42ba8ecee7ab08688a8c24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b6666e0ce42a46f669505659298a19931b1bf7c253ae6a97c6acfd80a9e717f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://251e72e1cd0f3e0ee195bacf33ab38765d1dd943adf4395f9d6dd282bda024cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://098c31b822e7c2351f1142af36b942d67ff8262896e28022c96a187d40bd0ddb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://098c31b822e7c2351f1142af36b942d67ff8262896e28022c96a187d40bd0ddb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"message\\\":\\\"le observer\\\\nW1122 07:56:50.641735 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1122 07:56:50.642792 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 07:56:50.643945 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3588433834/tls.crt::/tmp/serving-cert-3588433834/tls.key\\\\\\\"\\\\nI1122 07:56:50.949040 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 07:56:50.952068 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 07:56:50.952092 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 07:56:50.952116 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 07:56:50.952121 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 07:56:50.958549 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1122 07:56:50.958565 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1122 07:56:50.958579 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 07:56:50.958585 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 07:56:50.958591 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 07:56:50.958596 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1122 07:56:50.958600 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 07:56:50.958605 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1122 07:56:50.962189 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T07:56:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://263dce4c3f4f3fe388b21d4bb55a9fc7d956c3cca7da34ee00f470d88ef501f1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac2b567fcebc517d3fb4366aebce03351183ea03ce62f86c2abdf5b95e20b34a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac2b567fcebc517d3fb4366aebce03351183ea03ce62f86c2abdf5b95e20b34a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T07:56:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T07:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.355679 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.364442 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.367934 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.367978 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.367992 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.368010 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.368024 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:56:52Z","lastTransitionTime":"2025-11-22T07:56:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.368652 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.372417 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.373790 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58aa8071-7ecc-4692-8789-2db04ec70647\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfjgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfjgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pxvrm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.375665 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.383896 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93be48ce-7d87-448b-b03e-ab6736cfc34c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba91466edcf49f73751622f98632b3040a046ccd8a42ba8ecee7ab08688a8c24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b6666e0ce42a46f669505659298a19931b1bf7c253ae6a97c6acfd80a9e717f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://251e72e1cd0f3e0ee195bacf33ab38765d1dd943adf4395f9d6dd282bda024cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://098c31b822e7c2351f1142af36b942d67ff8262896e28022c96a187d40bd0ddb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://098c31b822e7c2351f1142af36b942d67ff8262896e28022c96a187d40bd0ddb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"message\\\":\\\"le observer\\\\nW1122 07:56:50.641735 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1122 07:56:50.642792 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 07:56:50.643945 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3588433834/tls.crt::/tmp/serving-cert-3588433834/tls.key\\\\\\\"\\\\nI1122 07:56:50.949040 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 07:56:50.952068 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 07:56:50.952092 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 07:56:50.952116 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 07:56:50.952121 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 07:56:50.958549 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1122 07:56:50.958565 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1122 07:56:50.958579 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 07:56:50.958585 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 07:56:50.958591 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 07:56:50.958596 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1122 07:56:50.958600 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 07:56:50.958605 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1122 07:56:50.962189 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T07:56:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://263dce4c3f4f3fe388b21d4bb55a9fc7d956c3cca7da34ee00f470d88ef501f1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac2b567fcebc517d3fb4366aebce03351183ea03ce62f86c2abdf5b95e20b34a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac2b567fcebc517d3fb4366aebce03351183ea03ce62f86c2abdf5b95e20b34a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T07:56:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T07:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.392149 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.398050 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/5613c209-59a0-4b8a-ac8b-5c633bc1bdc4-multus-daemon-config\") pod \"multus-rks5j\" (UID: \"5613c209-59a0-4b8a-ac8b-5c633bc1bdc4\") " pod="openshift-multus/multus-rks5j" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.398107 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.398134 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/5613c209-59a0-4b8a-ac8b-5c633bc1bdc4-host-run-netns\") pod \"multus-rks5j\" (UID: \"5613c209-59a0-4b8a-ac8b-5c633bc1bdc4\") " pod="openshift-multus/multus-rks5j" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.398153 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/5613c209-59a0-4b8a-ac8b-5c633bc1bdc4-host-var-lib-kubelet\") pod \"multus-rks5j\" (UID: \"5613c209-59a0-4b8a-ac8b-5c633bc1bdc4\") " pod="openshift-multus/multus-rks5j" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.398181 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5613c209-59a0-4b8a-ac8b-5c633bc1bdc4-etc-kubernetes\") pod \"multus-rks5j\" (UID: \"5613c209-59a0-4b8a-ac8b-5c633bc1bdc4\") " pod="openshift-multus/multus-rks5j" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.398197 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/5613c209-59a0-4b8a-ac8b-5c633bc1bdc4-host-var-lib-cni-multus\") pod \"multus-rks5j\" (UID: \"5613c209-59a0-4b8a-ac8b-5c633bc1bdc4\") " pod="openshift-multus/multus-rks5j" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.398213 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/5613c209-59a0-4b8a-ac8b-5c633bc1bdc4-host-run-multus-certs\") pod \"multus-rks5j\" (UID: \"5613c209-59a0-4b8a-ac8b-5c633bc1bdc4\") " pod="openshift-multus/multus-rks5j" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.398424 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pxvhw\" (UniqueName: \"kubernetes.io/projected/2fdee89f-559b-4f4e-8d20-6ecc3aafd22e-kube-api-access-pxvhw\") pod \"multus-additional-cni-plugins-2wdp7\" (UID: \"2fdee89f-559b-4f4e-8d20-6ecc3aafd22e\") " pod="openshift-multus/multus-additional-cni-plugins-2wdp7" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.398444 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/5613c209-59a0-4b8a-ac8b-5c633bc1bdc4-host-var-lib-cni-bin\") pod \"multus-rks5j\" (UID: \"5613c209-59a0-4b8a-ac8b-5c633bc1bdc4\") " pod="openshift-multus/multus-rks5j" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.398462 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/2fdee89f-559b-4f4e-8d20-6ecc3aafd22e-system-cni-dir\") pod \"multus-additional-cni-plugins-2wdp7\" (UID: \"2fdee89f-559b-4f4e-8d20-6ecc3aafd22e\") " pod="openshift-multus/multus-additional-cni-plugins-2wdp7" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.398468 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/5613c209-59a0-4b8a-ac8b-5c633bc1bdc4-host-var-lib-kubelet\") pod \"multus-rks5j\" (UID: \"5613c209-59a0-4b8a-ac8b-5c633bc1bdc4\") " pod="openshift-multus/multus-rks5j" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.398513 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5613c209-59a0-4b8a-ac8b-5c633bc1bdc4-etc-kubernetes\") pod \"multus-rks5j\" (UID: \"5613c209-59a0-4b8a-ac8b-5c633bc1bdc4\") " pod="openshift-multus/multus-rks5j" Nov 22 07:56:52 crc kubenswrapper[4789]: E1122 07:56:52.398514 4789 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.398548 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/5613c209-59a0-4b8a-ac8b-5c633bc1bdc4-os-release\") pod \"multus-rks5j\" (UID: \"5613c209-59a0-4b8a-ac8b-5c633bc1bdc4\") " pod="openshift-multus/multus-rks5j" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.398562 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/5613c209-59a0-4b8a-ac8b-5c633bc1bdc4-host-var-lib-cni-bin\") pod \"multus-rks5j\" (UID: \"5613c209-59a0-4b8a-ac8b-5c633bc1bdc4\") " pod="openshift-multus/multus-rks5j" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.398479 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/5613c209-59a0-4b8a-ac8b-5c633bc1bdc4-host-run-netns\") pod \"multus-rks5j\" (UID: \"5613c209-59a0-4b8a-ac8b-5c633bc1bdc4\") " pod="openshift-multus/multus-rks5j" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.398578 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/2fdee89f-559b-4f4e-8d20-6ecc3aafd22e-system-cni-dir\") pod \"multus-additional-cni-plugins-2wdp7\" (UID: \"2fdee89f-559b-4f4e-8d20-6ecc3aafd22e\") " pod="openshift-multus/multus-additional-cni-plugins-2wdp7" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.398541 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/5613c209-59a0-4b8a-ac8b-5c633bc1bdc4-host-var-lib-cni-multus\") pod \"multus-rks5j\" (UID: \"5613c209-59a0-4b8a-ac8b-5c633bc1bdc4\") " pod="openshift-multus/multus-rks5j" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.398527 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/5613c209-59a0-4b8a-ac8b-5c633bc1bdc4-host-run-multus-certs\") pod \"multus-rks5j\" (UID: \"5613c209-59a0-4b8a-ac8b-5c633bc1bdc4\") " pod="openshift-multus/multus-rks5j" Nov 22 07:56:52 crc kubenswrapper[4789]: E1122 07:56:52.398610 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-22 07:56:54.398590558 +0000 UTC m=+28.632990831 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.398478 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/5613c209-59a0-4b8a-ac8b-5c633bc1bdc4-os-release\") pod \"multus-rks5j\" (UID: \"5613c209-59a0-4b8a-ac8b-5c633bc1bdc4\") " pod="openshift-multus/multus-rks5j" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.398708 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wj8sc\" (UniqueName: \"kubernetes.io/projected/5613c209-59a0-4b8a-ac8b-5c633bc1bdc4-kube-api-access-wj8sc\") pod \"multus-rks5j\" (UID: \"5613c209-59a0-4b8a-ac8b-5c633bc1bdc4\") " pod="openshift-multus/multus-rks5j" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.398762 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/2fdee89f-559b-4f4e-8d20-6ecc3aafd22e-os-release\") pod \"multus-additional-cni-plugins-2wdp7\" (UID: \"2fdee89f-559b-4f4e-8d20-6ecc3aafd22e\") " pod="openshift-multus/multus-additional-cni-plugins-2wdp7" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.398798 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.398824 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/5613c209-59a0-4b8a-ac8b-5c633bc1bdc4-multus-socket-dir-parent\") pod \"multus-rks5j\" (UID: \"5613c209-59a0-4b8a-ac8b-5c633bc1bdc4\") " pod="openshift-multus/multus-rks5j" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.398845 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/5613c209-59a0-4b8a-ac8b-5c633bc1bdc4-system-cni-dir\") pod \"multus-rks5j\" (UID: \"5613c209-59a0-4b8a-ac8b-5c633bc1bdc4\") " pod="openshift-multus/multus-rks5j" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.398867 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/5613c209-59a0-4b8a-ac8b-5c633bc1bdc4-multus-conf-dir\") pod \"multus-rks5j\" (UID: \"5613c209-59a0-4b8a-ac8b-5c633bc1bdc4\") " pod="openshift-multus/multus-rks5j" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.398891 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/2fdee89f-559b-4f4e-8d20-6ecc3aafd22e-cni-binary-copy\") pod \"multus-additional-cni-plugins-2wdp7\" (UID: \"2fdee89f-559b-4f4e-8d20-6ecc3aafd22e\") " pod="openshift-multus/multus-additional-cni-plugins-2wdp7" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.398911 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/2fdee89f-559b-4f4e-8d20-6ecc3aafd22e-os-release\") pod \"multus-additional-cni-plugins-2wdp7\" (UID: \"2fdee89f-559b-4f4e-8d20-6ecc3aafd22e\") " pod="openshift-multus/multus-additional-cni-plugins-2wdp7" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.398920 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/5613c209-59a0-4b8a-ac8b-5c633bc1bdc4-cnibin\") pod \"multus-rks5j\" (UID: \"5613c209-59a0-4b8a-ac8b-5c633bc1bdc4\") " pod="openshift-multus/multus-rks5j" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.398964 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/5613c209-59a0-4b8a-ac8b-5c633bc1bdc4-cni-binary-copy\") pod \"multus-rks5j\" (UID: \"5613c209-59a0-4b8a-ac8b-5c633bc1bdc4\") " pod="openshift-multus/multus-rks5j" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.398971 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/5613c209-59a0-4b8a-ac8b-5c633bc1bdc4-cnibin\") pod \"multus-rks5j\" (UID: \"5613c209-59a0-4b8a-ac8b-5c633bc1bdc4\") " pod="openshift-multus/multus-rks5j" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.399000 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.399025 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/5613c209-59a0-4b8a-ac8b-5c633bc1bdc4-multus-cni-dir\") pod \"multus-rks5j\" (UID: \"5613c209-59a0-4b8a-ac8b-5c633bc1bdc4\") " pod="openshift-multus/multus-rks5j" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.399057 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/5613c209-59a0-4b8a-ac8b-5c633bc1bdc4-host-run-k8s-cni-cncf-io\") pod \"multus-rks5j\" (UID: \"5613c209-59a0-4b8a-ac8b-5c633bc1bdc4\") " pod="openshift-multus/multus-rks5j" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.399088 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 07:56:52 crc kubenswrapper[4789]: E1122 07:56:52.399100 4789 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.399111 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/2fdee89f-559b-4f4e-8d20-6ecc3aafd22e-cnibin\") pod \"multus-additional-cni-plugins-2wdp7\" (UID: \"2fdee89f-559b-4f4e-8d20-6ecc3aafd22e\") " pod="openshift-multus/multus-additional-cni-plugins-2wdp7" Nov 22 07:56:52 crc kubenswrapper[4789]: E1122 07:56:52.399125 4789 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.399134 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/2fdee89f-559b-4f4e-8d20-6ecc3aafd22e-tuning-conf-dir\") pod \"multus-additional-cni-plugins-2wdp7\" (UID: \"2fdee89f-559b-4f4e-8d20-6ecc3aafd22e\") " pod="openshift-multus/multus-additional-cni-plugins-2wdp7" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.399156 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/2fdee89f-559b-4f4e-8d20-6ecc3aafd22e-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-2wdp7\" (UID: \"2fdee89f-559b-4f4e-8d20-6ecc3aafd22e\") " pod="openshift-multus/multus-additional-cni-plugins-2wdp7" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.399177 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/5613c209-59a0-4b8a-ac8b-5c633bc1bdc4-hostroot\") pod \"multus-rks5j\" (UID: \"5613c209-59a0-4b8a-ac8b-5c633bc1bdc4\") " pod="openshift-multus/multus-rks5j" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.399181 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/5613c209-59a0-4b8a-ac8b-5c633bc1bdc4-multus-socket-dir-parent\") pod \"multus-rks5j\" (UID: \"5613c209-59a0-4b8a-ac8b-5c633bc1bdc4\") " pod="openshift-multus/multus-rks5j" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.399215 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/5613c209-59a0-4b8a-ac8b-5c633bc1bdc4-system-cni-dir\") pod \"multus-rks5j\" (UID: \"5613c209-59a0-4b8a-ac8b-5c633bc1bdc4\") " pod="openshift-multus/multus-rks5j" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.399234 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/5613c209-59a0-4b8a-ac8b-5c633bc1bdc4-multus-daemon-config\") pod \"multus-rks5j\" (UID: \"5613c209-59a0-4b8a-ac8b-5c633bc1bdc4\") " pod="openshift-multus/multus-rks5j" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.399253 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/5613c209-59a0-4b8a-ac8b-5c633bc1bdc4-multus-conf-dir\") pod \"multus-rks5j\" (UID: \"5613c209-59a0-4b8a-ac8b-5c633bc1bdc4\") " pod="openshift-multus/multus-rks5j" Nov 22 07:56:52 crc kubenswrapper[4789]: E1122 07:56:52.399137 4789 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.399441 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/5613c209-59a0-4b8a-ac8b-5c633bc1bdc4-hostroot\") pod \"multus-rks5j\" (UID: \"5613c209-59a0-4b8a-ac8b-5c633bc1bdc4\") " pod="openshift-multus/multus-rks5j" Nov 22 07:56:52 crc kubenswrapper[4789]: E1122 07:56:52.399454 4789 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.399411 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/5613c209-59a0-4b8a-ac8b-5c633bc1bdc4-multus-cni-dir\") pod \"multus-rks5j\" (UID: \"5613c209-59a0-4b8a-ac8b-5c633bc1bdc4\") " pod="openshift-multus/multus-rks5j" Nov 22 07:56:52 crc kubenswrapper[4789]: E1122 07:56:52.399105 4789 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.399531 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/2fdee89f-559b-4f4e-8d20-6ecc3aafd22e-cnibin\") pod \"multus-additional-cni-plugins-2wdp7\" (UID: \"2fdee89f-559b-4f4e-8d20-6ecc3aafd22e\") " pod="openshift-multus/multus-additional-cni-plugins-2wdp7" Nov 22 07:56:52 crc kubenswrapper[4789]: E1122 07:56:52.399542 4789 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 22 07:56:52 crc kubenswrapper[4789]: E1122 07:56:52.399554 4789 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.399546 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/5613c209-59a0-4b8a-ac8b-5c633bc1bdc4-host-run-k8s-cni-cncf-io\") pod \"multus-rks5j\" (UID: \"5613c209-59a0-4b8a-ac8b-5c633bc1bdc4\") " pod="openshift-multus/multus-rks5j" Nov 22 07:56:52 crc kubenswrapper[4789]: E1122 07:56:52.399519 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-22 07:56:54.399495202 +0000 UTC m=+28.633895555 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 07:56:52 crc kubenswrapper[4789]: E1122 07:56:52.399655 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-22 07:56:54.399626246 +0000 UTC m=+28.634026589 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.399663 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/5613c209-59a0-4b8a-ac8b-5c633bc1bdc4-cni-binary-copy\") pod \"multus-rks5j\" (UID: \"5613c209-59a0-4b8a-ac8b-5c633bc1bdc4\") " pod="openshift-multus/multus-rks5j" Nov 22 07:56:52 crc kubenswrapper[4789]: E1122 07:56:52.399697 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-22 07:56:54.399689667 +0000 UTC m=+28.634090070 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.400151 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/2fdee89f-559b-4f4e-8d20-6ecc3aafd22e-cni-binary-copy\") pod \"multus-additional-cni-plugins-2wdp7\" (UID: \"2fdee89f-559b-4f4e-8d20-6ecc3aafd22e\") " pod="openshift-multus/multus-additional-cni-plugins-2wdp7" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.400280 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/2fdee89f-559b-4f4e-8d20-6ecc3aafd22e-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-2wdp7\" (UID: \"2fdee89f-559b-4f4e-8d20-6ecc3aafd22e\") " pod="openshift-multus/multus-additional-cni-plugins-2wdp7" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.400388 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/2fdee89f-559b-4f4e-8d20-6ecc3aafd22e-tuning-conf-dir\") pod \"multus-additional-cni-plugins-2wdp7\" (UID: \"2fdee89f-559b-4f4e-8d20-6ecc3aafd22e\") " pod="openshift-multus/multus-additional-cni-plugins-2wdp7" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.402086 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rks5j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5613c209-59a0-4b8a-ac8b-5c633bc1bdc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wj8sc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rks5j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.412449 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.414695 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wj8sc\" (UniqueName: \"kubernetes.io/projected/5613c209-59a0-4b8a-ac8b-5c633bc1bdc4-kube-api-access-wj8sc\") pod \"multus-rks5j\" (UID: \"5613c209-59a0-4b8a-ac8b-5c633bc1bdc4\") " pod="openshift-multus/multus-rks5j" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.415699 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pxvhw\" (UniqueName: \"kubernetes.io/projected/2fdee89f-559b-4f4e-8d20-6ecc3aafd22e-kube-api-access-pxvhw\") pod \"multus-additional-cni-plugins-2wdp7\" (UID: \"2fdee89f-559b-4f4e-8d20-6ecc3aafd22e\") " pod="openshift-multus/multus-additional-cni-plugins-2wdp7" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.420718 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.429160 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.438393 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.446373 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9f74f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"29ad5e00-4918-4718-9d49-013b12621f7d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mnz6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9f74f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.457433 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2wdp7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2fdee89f-559b-4f4e-8d20-6ecc3aafd22e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxvhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxvhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxvhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxvhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxvhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxvhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxvhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2wdp7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.467731 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93be48ce-7d87-448b-b03e-ab6736cfc34c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba91466edcf49f73751622f98632b3040a046ccd8a42ba8ecee7ab08688a8c24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b6666e0ce42a46f669505659298a19931b1bf7c253ae6a97c6acfd80a9e717f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://251e72e1cd0f3e0ee195bacf33ab38765d1dd943adf4395f9d6dd282bda024cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://098c31b822e7c2351f1142af36b942d67ff8262896e28022c96a187d40bd0ddb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://098c31b822e7c2351f1142af36b942d67ff8262896e28022c96a187d40bd0ddb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"message\\\":\\\"le observer\\\\nW1122 07:56:50.641735 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1122 07:56:50.642792 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 07:56:50.643945 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3588433834/tls.crt::/tmp/serving-cert-3588433834/tls.key\\\\\\\"\\\\nI1122 07:56:50.949040 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 07:56:50.952068 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 07:56:50.952092 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 07:56:50.952116 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 07:56:50.952121 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 07:56:50.958549 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1122 07:56:50.958565 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1122 07:56:50.958579 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 07:56:50.958585 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 07:56:50.958591 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 07:56:50.958596 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1122 07:56:50.958600 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 07:56:50.958605 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1122 07:56:50.962189 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T07:56:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://263dce4c3f4f3fe388b21d4bb55a9fc7d956c3cca7da34ee00f470d88ef501f1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac2b567fcebc517d3fb4366aebce03351183ea03ce62f86c2abdf5b95e20b34a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac2b567fcebc517d3fb4366aebce03351183ea03ce62f86c2abdf5b95e20b34a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T07:56:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T07:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.470433 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.470474 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.470485 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.470500 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.470512 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:56:52Z","lastTransitionTime":"2025-11-22T07:56:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.476461 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.485112 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rks5j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5613c209-59a0-4b8a-ac8b-5c633bc1bdc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wj8sc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rks5j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.493707 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.501916 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.539397 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.572829 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.572867 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.572876 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.572908 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.572920 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:56:52Z","lastTransitionTime":"2025-11-22T07:56:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.574178 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-rks5j" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.579667 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.580744 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-2wdp7" Nov 22 07:56:52 crc kubenswrapper[4789]: W1122 07:56:52.584723 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5613c209_59a0_4b8a_ac8b_5c633bc1bdc4.slice/crio-d37ce6be707cf546a6075783d075e7c1ba7e639aee014a897bcc661b50a48877 WatchSource:0}: Error finding container d37ce6be707cf546a6075783d075e7c1ba7e639aee014a897bcc661b50a48877: Status 404 returned error can't find the container with id d37ce6be707cf546a6075783d075e7c1ba7e639aee014a897bcc661b50a48877 Nov 22 07:56:52 crc kubenswrapper[4789]: E1122 07:56:52.587263 4789 kuberuntime_manager.go:1274] "Unhandled Error" err=< Nov 22 07:56:52 crc kubenswrapper[4789]: container &Container{Name:kube-multus,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26,Command:[/bin/bash -ec --],Args:[MULTUS_DAEMON_OPT="" Nov 22 07:56:52 crc kubenswrapper[4789]: /entrypoint/cnibincopy.sh; exec /usr/src/multus-cni/bin/multus-daemon $MULTUS_DAEMON_OPT Nov 22 07:56:52 crc kubenswrapper[4789]: ],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:RHEL8_SOURCE_DIRECTORY,Value:/usr/src/multus-cni/rhel8/bin/,ValueFrom:nil,},EnvVar{Name:RHEL9_SOURCE_DIRECTORY,Value:/usr/src/multus-cni/rhel9/bin/,ValueFrom:nil,},EnvVar{Name:DEFAULT_SOURCE_DIRECTORY,Value:/usr/src/multus-cni/bin/,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:6443,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:api-int.crc.testing,ValueFrom:nil,},EnvVar{Name:MULTUS_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:K8S_NODE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{68157440 0} {} 65Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cni-binary-copy,ReadOnly:false,MountPath:/entrypoint,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:os-release,ReadOnly:false,MountPath:/host/etc/os-release,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:system-cni-dir,ReadOnly:false,MountPath:/host/etc/cni/net.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:multus-cni-dir,ReadOnly:false,MountPath:/host/run/multus/cni/net.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:cnibin,ReadOnly:false,MountPath:/host/opt/cni/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:multus-socket-dir-parent,ReadOnly:false,MountPath:/host/run/multus,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-run-k8s-cni-cncf-io,ReadOnly:false,MountPath:/run/k8s.cni.cncf.io,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-run-netns,ReadOnly:false,MountPath:/run/netns,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-var-lib-cni-bin,ReadOnly:false,MountPath:/var/lib/cni/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-var-lib-cni-multus,ReadOnly:false,MountPath:/var/lib/cni/multus,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-var-lib-kubelet,ReadOnly:false,MountPath:/var/lib/kubelet,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:hostroot,ReadOnly:false,MountPath:/hostroot,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:multus-conf-dir,ReadOnly:false,MountPath:/etc/cni/multus/net.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:multus-daemon-config,ReadOnly:true,MountPath:/etc/cni/net.d/multus.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-run-multus-certs,ReadOnly:false,MountPath:/etc/cni/multus/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:etc-kubernetes,ReadOnly:false,MountPath:/etc/kubernetes,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wj8sc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod multus-rks5j_openshift-multus(5613c209-59a0-4b8a-ac8b-5c633bc1bdc4): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Nov 22 07:56:52 crc kubenswrapper[4789]: > logger="UnhandledError" Nov 22 07:56:52 crc kubenswrapper[4789]: E1122 07:56:52.588981 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-multus/multus-rks5j" podUID="5613c209-59a0-4b8a-ac8b-5c633bc1bdc4" Nov 22 07:56:52 crc kubenswrapper[4789]: W1122 07:56:52.593790 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2fdee89f_559b_4f4e_8d20_6ecc3aafd22e.slice/crio-36008df6c4e477404c0c72914ccd243422a118d331cb54900d3b3b8a734e4e70 WatchSource:0}: Error finding container 36008df6c4e477404c0c72914ccd243422a118d331cb54900d3b3b8a734e4e70: Status 404 returned error can't find the container with id 36008df6c4e477404c0c72914ccd243422a118d331cb54900d3b3b8a734e4e70 Nov 22 07:56:52 crc kubenswrapper[4789]: E1122 07:56:52.597128 4789 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:egress-router-binary-copy,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c,Command:[/entrypoint/cnibincopy.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:RHEL8_SOURCE_DIRECTORY,Value:/usr/src/egress-router-cni/rhel8/bin/,ValueFrom:nil,},EnvVar{Name:RHEL9_SOURCE_DIRECTORY,Value:/usr/src/egress-router-cni/rhel9/bin/,ValueFrom:nil,},EnvVar{Name:DEFAULT_SOURCE_DIRECTORY,Value:/usr/src/egress-router-cni/bin/,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cni-binary-copy,ReadOnly:false,MountPath:/entrypoint,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:cnibin,ReadOnly:false,MountPath:/host/opt/cni/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:os-release,ReadOnly:true,MountPath:/host/etc/os-release,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-pxvhw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod multus-additional-cni-plugins-2wdp7_openshift-multus(2fdee89f-559b-4f4e-8d20-6ecc3aafd22e): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars" logger="UnhandledError" Nov 22 07:56:52 crc kubenswrapper[4789]: E1122 07:56:52.598299 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"egress-router-binary-copy\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-multus/multus-additional-cni-plugins-2wdp7" podUID="2fdee89f-559b-4f4e-8d20-6ecc3aafd22e" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.617568 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9f74f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"29ad5e00-4918-4718-9d49-013b12621f7d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mnz6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9f74f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.622636 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-qmxss"] Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.626114 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.653947 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.673948 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.675588 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.675622 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.675635 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.675651 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.675662 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:56:52Z","lastTransitionTime":"2025-11-22T07:56:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.694033 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.701613 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0cb4cf51-62d6-4f32-a8b4-4116555f3870-host-run-ovn-kubernetes\") pod \"ovnkube-node-qmxss\" (UID: \"0cb4cf51-62d6-4f32-a8b4-4116555f3870\") " pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.701674 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/0cb4cf51-62d6-4f32-a8b4-4116555f3870-host-slash\") pod \"ovnkube-node-qmxss\" (UID: \"0cb4cf51-62d6-4f32-a8b4-4116555f3870\") " pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.701699 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/0cb4cf51-62d6-4f32-a8b4-4116555f3870-log-socket\") pod \"ovnkube-node-qmxss\" (UID: \"0cb4cf51-62d6-4f32-a8b4-4116555f3870\") " pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.701721 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/0cb4cf51-62d6-4f32-a8b4-4116555f3870-run-systemd\") pod \"ovnkube-node-qmxss\" (UID: \"0cb4cf51-62d6-4f32-a8b4-4116555f3870\") " pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.701741 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/0cb4cf51-62d6-4f32-a8b4-4116555f3870-host-cni-netd\") pod \"ovnkube-node-qmxss\" (UID: \"0cb4cf51-62d6-4f32-a8b4-4116555f3870\") " pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.701787 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/0cb4cf51-62d6-4f32-a8b4-4116555f3870-ovn-node-metrics-cert\") pod \"ovnkube-node-qmxss\" (UID: \"0cb4cf51-62d6-4f32-a8b4-4116555f3870\") " pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.701808 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/0cb4cf51-62d6-4f32-a8b4-4116555f3870-ovnkube-script-lib\") pod \"ovnkube-node-qmxss\" (UID: \"0cb4cf51-62d6-4f32-a8b4-4116555f3870\") " pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.701827 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t7d59\" (UniqueName: \"kubernetes.io/projected/0cb4cf51-62d6-4f32-a8b4-4116555f3870-kube-api-access-t7d59\") pod \"ovnkube-node-qmxss\" (UID: \"0cb4cf51-62d6-4f32-a8b4-4116555f3870\") " pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.701858 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0cb4cf51-62d6-4f32-a8b4-4116555f3870-run-openvswitch\") pod \"ovnkube-node-qmxss\" (UID: \"0cb4cf51-62d6-4f32-a8b4-4116555f3870\") " pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.702292 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/0cb4cf51-62d6-4f32-a8b4-4116555f3870-ovnkube-config\") pod \"ovnkube-node-qmxss\" (UID: \"0cb4cf51-62d6-4f32-a8b4-4116555f3870\") " pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.702326 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0cb4cf51-62d6-4f32-a8b4-4116555f3870-var-lib-openvswitch\") pod \"ovnkube-node-qmxss\" (UID: \"0cb4cf51-62d6-4f32-a8b4-4116555f3870\") " pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.702345 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0cb4cf51-62d6-4f32-a8b4-4116555f3870-etc-openvswitch\") pod \"ovnkube-node-qmxss\" (UID: \"0cb4cf51-62d6-4f32-a8b4-4116555f3870\") " pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.702361 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/0cb4cf51-62d6-4f32-a8b4-4116555f3870-node-log\") pod \"ovnkube-node-qmxss\" (UID: \"0cb4cf51-62d6-4f32-a8b4-4116555f3870\") " pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.702377 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/0cb4cf51-62d6-4f32-a8b4-4116555f3870-host-cni-bin\") pod \"ovnkube-node-qmxss\" (UID: \"0cb4cf51-62d6-4f32-a8b4-4116555f3870\") " pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.702397 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0cb4cf51-62d6-4f32-a8b4-4116555f3870-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-qmxss\" (UID: \"0cb4cf51-62d6-4f32-a8b4-4116555f3870\") " pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.702415 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/0cb4cf51-62d6-4f32-a8b4-4116555f3870-systemd-units\") pod \"ovnkube-node-qmxss\" (UID: \"0cb4cf51-62d6-4f32-a8b4-4116555f3870\") " pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.702431 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/0cb4cf51-62d6-4f32-a8b4-4116555f3870-run-ovn\") pod \"ovnkube-node-qmxss\" (UID: \"0cb4cf51-62d6-4f32-a8b4-4116555f3870\") " pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.702458 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/0cb4cf51-62d6-4f32-a8b4-4116555f3870-host-run-netns\") pod \"ovnkube-node-qmxss\" (UID: \"0cb4cf51-62d6-4f32-a8b4-4116555f3870\") " pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.702475 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/0cb4cf51-62d6-4f32-a8b4-4116555f3870-env-overrides\") pod \"ovnkube-node-qmxss\" (UID: \"0cb4cf51-62d6-4f32-a8b4-4116555f3870\") " pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.702491 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/0cb4cf51-62d6-4f32-a8b4-4116555f3870-host-kubelet\") pod \"ovnkube-node-qmxss\" (UID: \"0cb4cf51-62d6-4f32-a8b4-4116555f3870\") " pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.713897 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.733859 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.753266 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.773925 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.778022 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.778056 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.778068 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.778083 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.778093 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:56:52Z","lastTransitionTime":"2025-11-22T07:56:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.800478 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2wdp7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2fdee89f-559b-4f4e-8d20-6ecc3aafd22e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxvhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxvhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxvhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxvhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxvhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxvhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxvhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2wdp7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.803788 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/0cb4cf51-62d6-4f32-a8b4-4116555f3870-run-ovn\") pod \"ovnkube-node-qmxss\" (UID: \"0cb4cf51-62d6-4f32-a8b4-4116555f3870\") " pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.803851 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/0cb4cf51-62d6-4f32-a8b4-4116555f3870-host-run-netns\") pod \"ovnkube-node-qmxss\" (UID: \"0cb4cf51-62d6-4f32-a8b4-4116555f3870\") " pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.803876 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/0cb4cf51-62d6-4f32-a8b4-4116555f3870-env-overrides\") pod \"ovnkube-node-qmxss\" (UID: \"0cb4cf51-62d6-4f32-a8b4-4116555f3870\") " pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.803900 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/0cb4cf51-62d6-4f32-a8b4-4116555f3870-host-kubelet\") pod \"ovnkube-node-qmxss\" (UID: \"0cb4cf51-62d6-4f32-a8b4-4116555f3870\") " pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.803931 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0cb4cf51-62d6-4f32-a8b4-4116555f3870-host-run-ovn-kubernetes\") pod \"ovnkube-node-qmxss\" (UID: \"0cb4cf51-62d6-4f32-a8b4-4116555f3870\") " pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.803939 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/0cb4cf51-62d6-4f32-a8b4-4116555f3870-run-ovn\") pod \"ovnkube-node-qmxss\" (UID: \"0cb4cf51-62d6-4f32-a8b4-4116555f3870\") " pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.803964 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/0cb4cf51-62d6-4f32-a8b4-4116555f3870-host-slash\") pod \"ovnkube-node-qmxss\" (UID: \"0cb4cf51-62d6-4f32-a8b4-4116555f3870\") " pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.803984 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/0cb4cf51-62d6-4f32-a8b4-4116555f3870-log-socket\") pod \"ovnkube-node-qmxss\" (UID: \"0cb4cf51-62d6-4f32-a8b4-4116555f3870\") " pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.804005 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/0cb4cf51-62d6-4f32-a8b4-4116555f3870-run-systemd\") pod \"ovnkube-node-qmxss\" (UID: \"0cb4cf51-62d6-4f32-a8b4-4116555f3870\") " pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.804010 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/0cb4cf51-62d6-4f32-a8b4-4116555f3870-host-kubelet\") pod \"ovnkube-node-qmxss\" (UID: \"0cb4cf51-62d6-4f32-a8b4-4116555f3870\") " pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.804024 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t7d59\" (UniqueName: \"kubernetes.io/projected/0cb4cf51-62d6-4f32-a8b4-4116555f3870-kube-api-access-t7d59\") pod \"ovnkube-node-qmxss\" (UID: \"0cb4cf51-62d6-4f32-a8b4-4116555f3870\") " pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.804039 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/0cb4cf51-62d6-4f32-a8b4-4116555f3870-host-run-netns\") pod \"ovnkube-node-qmxss\" (UID: \"0cb4cf51-62d6-4f32-a8b4-4116555f3870\") " pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.804046 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/0cb4cf51-62d6-4f32-a8b4-4116555f3870-host-cni-netd\") pod \"ovnkube-node-qmxss\" (UID: \"0cb4cf51-62d6-4f32-a8b4-4116555f3870\") " pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.804066 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/0cb4cf51-62d6-4f32-a8b4-4116555f3870-ovn-node-metrics-cert\") pod \"ovnkube-node-qmxss\" (UID: \"0cb4cf51-62d6-4f32-a8b4-4116555f3870\") " pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.804085 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/0cb4cf51-62d6-4f32-a8b4-4116555f3870-ovnkube-script-lib\") pod \"ovnkube-node-qmxss\" (UID: \"0cb4cf51-62d6-4f32-a8b4-4116555f3870\") " pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.804113 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0cb4cf51-62d6-4f32-a8b4-4116555f3870-run-openvswitch\") pod \"ovnkube-node-qmxss\" (UID: \"0cb4cf51-62d6-4f32-a8b4-4116555f3870\") " pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.804135 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/0cb4cf51-62d6-4f32-a8b4-4116555f3870-ovnkube-config\") pod \"ovnkube-node-qmxss\" (UID: \"0cb4cf51-62d6-4f32-a8b4-4116555f3870\") " pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.804159 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/0cb4cf51-62d6-4f32-a8b4-4116555f3870-host-cni-bin\") pod \"ovnkube-node-qmxss\" (UID: \"0cb4cf51-62d6-4f32-a8b4-4116555f3870\") " pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.804162 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/0cb4cf51-62d6-4f32-a8b4-4116555f3870-run-systemd\") pod \"ovnkube-node-qmxss\" (UID: \"0cb4cf51-62d6-4f32-a8b4-4116555f3870\") " pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.804181 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0cb4cf51-62d6-4f32-a8b4-4116555f3870-var-lib-openvswitch\") pod \"ovnkube-node-qmxss\" (UID: \"0cb4cf51-62d6-4f32-a8b4-4116555f3870\") " pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.804239 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0cb4cf51-62d6-4f32-a8b4-4116555f3870-etc-openvswitch\") pod \"ovnkube-node-qmxss\" (UID: \"0cb4cf51-62d6-4f32-a8b4-4116555f3870\") " pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.804250 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/0cb4cf51-62d6-4f32-a8b4-4116555f3870-host-slash\") pod \"ovnkube-node-qmxss\" (UID: \"0cb4cf51-62d6-4f32-a8b4-4116555f3870\") " pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.804288 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/0cb4cf51-62d6-4f32-a8b4-4116555f3870-node-log\") pod \"ovnkube-node-qmxss\" (UID: \"0cb4cf51-62d6-4f32-a8b4-4116555f3870\") " pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.804263 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/0cb4cf51-62d6-4f32-a8b4-4116555f3870-node-log\") pod \"ovnkube-node-qmxss\" (UID: \"0cb4cf51-62d6-4f32-a8b4-4116555f3870\") " pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.804316 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/0cb4cf51-62d6-4f32-a8b4-4116555f3870-log-socket\") pod \"ovnkube-node-qmxss\" (UID: \"0cb4cf51-62d6-4f32-a8b4-4116555f3870\") " pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.804335 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0cb4cf51-62d6-4f32-a8b4-4116555f3870-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-qmxss\" (UID: \"0cb4cf51-62d6-4f32-a8b4-4116555f3870\") " pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.804284 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/0cb4cf51-62d6-4f32-a8b4-4116555f3870-host-cni-netd\") pod \"ovnkube-node-qmxss\" (UID: \"0cb4cf51-62d6-4f32-a8b4-4116555f3870\") " pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.804366 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/0cb4cf51-62d6-4f32-a8b4-4116555f3870-systemd-units\") pod \"ovnkube-node-qmxss\" (UID: \"0cb4cf51-62d6-4f32-a8b4-4116555f3870\") " pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.804377 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0cb4cf51-62d6-4f32-a8b4-4116555f3870-run-openvswitch\") pod \"ovnkube-node-qmxss\" (UID: \"0cb4cf51-62d6-4f32-a8b4-4116555f3870\") " pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.804250 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0cb4cf51-62d6-4f32-a8b4-4116555f3870-host-run-ovn-kubernetes\") pod \"ovnkube-node-qmxss\" (UID: \"0cb4cf51-62d6-4f32-a8b4-4116555f3870\") " pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.804283 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/0cb4cf51-62d6-4f32-a8b4-4116555f3870-host-cni-bin\") pod \"ovnkube-node-qmxss\" (UID: \"0cb4cf51-62d6-4f32-a8b4-4116555f3870\") " pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.804426 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/0cb4cf51-62d6-4f32-a8b4-4116555f3870-systemd-units\") pod \"ovnkube-node-qmxss\" (UID: \"0cb4cf51-62d6-4f32-a8b4-4116555f3870\") " pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.804462 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0cb4cf51-62d6-4f32-a8b4-4116555f3870-var-lib-openvswitch\") pod \"ovnkube-node-qmxss\" (UID: \"0cb4cf51-62d6-4f32-a8b4-4116555f3870\") " pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.804543 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0cb4cf51-62d6-4f32-a8b4-4116555f3870-etc-openvswitch\") pod \"ovnkube-node-qmxss\" (UID: \"0cb4cf51-62d6-4f32-a8b4-4116555f3870\") " pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.804566 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0cb4cf51-62d6-4f32-a8b4-4116555f3870-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-qmxss\" (UID: \"0cb4cf51-62d6-4f32-a8b4-4116555f3870\") " pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.804768 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/0cb4cf51-62d6-4f32-a8b4-4116555f3870-env-overrides\") pod \"ovnkube-node-qmxss\" (UID: \"0cb4cf51-62d6-4f32-a8b4-4116555f3870\") " pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.804900 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/0cb4cf51-62d6-4f32-a8b4-4116555f3870-ovnkube-config\") pod \"ovnkube-node-qmxss\" (UID: \"0cb4cf51-62d6-4f32-a8b4-4116555f3870\") " pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.804954 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/0cb4cf51-62d6-4f32-a8b4-4116555f3870-ovnkube-script-lib\") pod \"ovnkube-node-qmxss\" (UID: \"0cb4cf51-62d6-4f32-a8b4-4116555f3870\") " pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.807543 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/0cb4cf51-62d6-4f32-a8b4-4116555f3870-ovn-node-metrics-cert\") pod \"ovnkube-node-qmxss\" (UID: \"0cb4cf51-62d6-4f32-a8b4-4116555f3870\") " pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.850981 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t7d59\" (UniqueName: \"kubernetes.io/projected/0cb4cf51-62d6-4f32-a8b4-4116555f3870-kube-api-access-t7d59\") pod \"ovnkube-node-qmxss\" (UID: \"0cb4cf51-62d6-4f32-a8b4-4116555f3870\") " pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.860412 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d081b29e-edd7-4fb0-a2e5-431ee8491bd0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43d9e10254883b6491bef617cf9f0b319171bc1fbc49c7770ed9630a8a1ae2cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://223c2300bb6afb8812a72eb7605015b0473f59046565cdf39fb9b9f0cf6634bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca5af01b2e23b2bcac23ed767bc7dcd2ac0be1010ee5e675b3b44333ee337180\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8e79bfb6172c894868a8e1b47ac701ba4fdaca1fc6b1582f7e70807aadfdf1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:29Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.880660 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.880700 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.880710 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.880724 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.880734 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:56:52Z","lastTransitionTime":"2025-11-22T07:56:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.902904 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.937402 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.938876 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58aa8071-7ecc-4692-8789-2db04ec70647\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfjgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfjgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pxvrm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:52 crc kubenswrapper[4789]: E1122 07:56:52.954234 4789 kuberuntime_manager.go:1274] "Unhandled Error" err=< Nov 22 07:56:52 crc kubenswrapper[4789]: init container &Container{Name:kubecfg-setup,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,Command:[/bin/bash -c cat << EOF > /etc/ovn/kubeconfig Nov 22 07:56:52 crc kubenswrapper[4789]: apiVersion: v1 Nov 22 07:56:52 crc kubenswrapper[4789]: clusters: Nov 22 07:56:52 crc kubenswrapper[4789]: - cluster: Nov 22 07:56:52 crc kubenswrapper[4789]: certificate-authority: /var/run/secrets/kubernetes.io/serviceaccount/ca.crt Nov 22 07:56:52 crc kubenswrapper[4789]: server: https://api-int.crc.testing:6443 Nov 22 07:56:52 crc kubenswrapper[4789]: name: default-cluster Nov 22 07:56:52 crc kubenswrapper[4789]: contexts: Nov 22 07:56:52 crc kubenswrapper[4789]: - context: Nov 22 07:56:52 crc kubenswrapper[4789]: cluster: default-cluster Nov 22 07:56:52 crc kubenswrapper[4789]: namespace: default Nov 22 07:56:52 crc kubenswrapper[4789]: user: default-auth Nov 22 07:56:52 crc kubenswrapper[4789]: name: default-context Nov 22 07:56:52 crc kubenswrapper[4789]: current-context: default-context Nov 22 07:56:52 crc kubenswrapper[4789]: kind: Config Nov 22 07:56:52 crc kubenswrapper[4789]: preferences: {} Nov 22 07:56:52 crc kubenswrapper[4789]: users: Nov 22 07:56:52 crc kubenswrapper[4789]: - name: default-auth Nov 22 07:56:52 crc kubenswrapper[4789]: user: Nov 22 07:56:52 crc kubenswrapper[4789]: client-certificate: /etc/ovn/ovnkube-node-certs/ovnkube-client-current.pem Nov 22 07:56:52 crc kubenswrapper[4789]: client-key: /etc/ovn/ovnkube-node-certs/ovnkube-client-current.pem Nov 22 07:56:52 crc kubenswrapper[4789]: EOF Nov 22 07:56:52 crc kubenswrapper[4789]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-openvswitch,ReadOnly:false,MountPath:/etc/ovn/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-t7d59,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovnkube-node-qmxss_openshift-ovn-kubernetes(0cb4cf51-62d6-4f32-a8b4-4116555f3870): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Nov 22 07:56:52 crc kubenswrapper[4789]: > logger="UnhandledError" Nov 22 07:56:52 crc kubenswrapper[4789]: E1122 07:56:52.955627 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kubecfg-setup\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" podUID="0cb4cf51-62d6-4f32-a8b4-4116555f3870" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.964712 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.964736 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 07:56:52 crc kubenswrapper[4789]: E1122 07:56:52.964880 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 07:56:52 crc kubenswrapper[4789]: E1122 07:56:52.964943 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.978073 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.982518 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.982555 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.982566 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.982583 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.982595 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:56:52Z","lastTransitionTime":"2025-11-22T07:56:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:56:52 crc kubenswrapper[4789]: I1122 07:56:52.990052 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Nov 22 07:56:53 crc kubenswrapper[4789]: I1122 07:56:53.001661 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Nov 22 07:56:53 crc kubenswrapper[4789]: I1122 07:56:53.024267 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Nov 22 07:56:53 crc kubenswrapper[4789]: I1122 07:56:53.043121 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rks5j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5613c209-59a0-4b8a-ac8b-5c633bc1bdc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wj8sc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rks5j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:53 crc kubenswrapper[4789]: I1122 07:56:53.080775 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93be48ce-7d87-448b-b03e-ab6736cfc34c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba91466edcf49f73751622f98632b3040a046ccd8a42ba8ecee7ab08688a8c24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b6666e0ce42a46f669505659298a19931b1bf7c253ae6a97c6acfd80a9e717f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://251e72e1cd0f3e0ee195bacf33ab38765d1dd943adf4395f9d6dd282bda024cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://098c31b822e7c2351f1142af36b942d67ff8262896e28022c96a187d40bd0ddb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://098c31b822e7c2351f1142af36b942d67ff8262896e28022c96a187d40bd0ddb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"message\\\":\\\"le observer\\\\nW1122 07:56:50.641735 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1122 07:56:50.642792 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 07:56:50.643945 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3588433834/tls.crt::/tmp/serving-cert-3588433834/tls.key\\\\\\\"\\\\nI1122 07:56:50.949040 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 07:56:50.952068 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 07:56:50.952092 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 07:56:50.952116 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 07:56:50.952121 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 07:56:50.958549 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1122 07:56:50.958565 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1122 07:56:50.958579 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 07:56:50.958585 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 07:56:50.958591 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 07:56:50.958596 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1122 07:56:50.958600 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 07:56:50.958605 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1122 07:56:50.962189 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T07:56:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://263dce4c3f4f3fe388b21d4bb55a9fc7d956c3cca7da34ee00f470d88ef501f1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac2b567fcebc517d3fb4366aebce03351183ea03ce62f86c2abdf5b95e20b34a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac2b567fcebc517d3fb4366aebce03351183ea03ce62f86c2abdf5b95e20b34a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T07:56:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T07:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:53 crc kubenswrapper[4789]: I1122 07:56:53.085221 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:53 crc kubenswrapper[4789]: I1122 07:56:53.085257 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:53 crc kubenswrapper[4789]: I1122 07:56:53.085268 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:53 crc kubenswrapper[4789]: I1122 07:56:53.085284 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:56:53 crc kubenswrapper[4789]: I1122 07:56:53.085295 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:56:53Z","lastTransitionTime":"2025-11-22T07:56:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:56:53 crc kubenswrapper[4789]: I1122 07:56:53.119574 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" event={"ID":"58aa8071-7ecc-4692-8789-2db04ec70647","Type":"ContainerStarted","Data":"809d26d975296963367b04cb00e88486f5364de5d7effac9e2eca6cb08b0a5b6"} Nov 22 07:56:53 crc kubenswrapper[4789]: I1122 07:56:53.120138 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:53 crc kubenswrapper[4789]: I1122 07:56:53.120523 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" event={"ID":"0cb4cf51-62d6-4f32-a8b4-4116555f3870","Type":"ContainerStarted","Data":"0d23c5a22905ab46e9052894cc4a7cbfb9d56d8ee528d3286a086123b9218c21"} Nov 22 07:56:53 crc kubenswrapper[4789]: E1122 07:56:53.121331 4789 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:machine-config-daemon,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a,Command:[/usr/bin/machine-config-daemon],Args:[start --payload-version=4.18.1],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:health,HostPort:8798,ContainerPort:8798,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:rootfs,ReadOnly:false,MountPath:/rootfs,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bfjgw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/health,Port:{0 8798 },Host:127.0.0.1,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:120,TimeoutSeconds:1,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars" logger="UnhandledError" Nov 22 07:56:53 crc kubenswrapper[4789]: E1122 07:56:53.121543 4789 kuberuntime_manager.go:1274] "Unhandled Error" err=< Nov 22 07:56:53 crc kubenswrapper[4789]: init container &Container{Name:kubecfg-setup,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,Command:[/bin/bash -c cat << EOF > /etc/ovn/kubeconfig Nov 22 07:56:53 crc kubenswrapper[4789]: apiVersion: v1 Nov 22 07:56:53 crc kubenswrapper[4789]: clusters: Nov 22 07:56:53 crc kubenswrapper[4789]: - cluster: Nov 22 07:56:53 crc kubenswrapper[4789]: certificate-authority: /var/run/secrets/kubernetes.io/serviceaccount/ca.crt Nov 22 07:56:53 crc kubenswrapper[4789]: server: https://api-int.crc.testing:6443 Nov 22 07:56:53 crc kubenswrapper[4789]: name: default-cluster Nov 22 07:56:53 crc kubenswrapper[4789]: contexts: Nov 22 07:56:53 crc kubenswrapper[4789]: - context: Nov 22 07:56:53 crc kubenswrapper[4789]: cluster: default-cluster Nov 22 07:56:53 crc kubenswrapper[4789]: namespace: default Nov 22 07:56:53 crc kubenswrapper[4789]: user: default-auth Nov 22 07:56:53 crc kubenswrapper[4789]: name: default-context Nov 22 07:56:53 crc kubenswrapper[4789]: current-context: default-context Nov 22 07:56:53 crc kubenswrapper[4789]: kind: Config Nov 22 07:56:53 crc kubenswrapper[4789]: preferences: {} Nov 22 07:56:53 crc kubenswrapper[4789]: users: Nov 22 07:56:53 crc kubenswrapper[4789]: - name: default-auth Nov 22 07:56:53 crc kubenswrapper[4789]: user: Nov 22 07:56:53 crc kubenswrapper[4789]: client-certificate: /etc/ovn/ovnkube-node-certs/ovnkube-client-current.pem Nov 22 07:56:53 crc kubenswrapper[4789]: client-key: /etc/ovn/ovnkube-node-certs/ovnkube-client-current.pem Nov 22 07:56:53 crc kubenswrapper[4789]: EOF Nov 22 07:56:53 crc kubenswrapper[4789]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-openvswitch,ReadOnly:false,MountPath:/etc/ovn/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-t7d59,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovnkube-node-qmxss_openshift-ovn-kubernetes(0cb4cf51-62d6-4f32-a8b4-4116555f3870): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Nov 22 07:56:53 crc kubenswrapper[4789]: > logger="UnhandledError" Nov 22 07:56:53 crc kubenswrapper[4789]: I1122 07:56:53.121644 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2wdp7" event={"ID":"2fdee89f-559b-4f4e-8d20-6ecc3aafd22e","Type":"ContainerStarted","Data":"36008df6c4e477404c0c72914ccd243422a118d331cb54900d3b3b8a734e4e70"} Nov 22 07:56:53 crc kubenswrapper[4789]: I1122 07:56:53.122471 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-9f74f" event={"ID":"29ad5e00-4918-4718-9d49-013b12621f7d","Type":"ContainerStarted","Data":"e6fb47fa3f3cc35dc53b3f5cc52b1b60e44be4b2de4afac3241e9a52dc8d4c60"} Nov 22 07:56:53 crc kubenswrapper[4789]: E1122 07:56:53.122614 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kubecfg-setup\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" podUID="0cb4cf51-62d6-4f32-a8b4-4116555f3870" Nov 22 07:56:53 crc kubenswrapper[4789]: E1122 07:56:53.122789 4789 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:egress-router-binary-copy,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c,Command:[/entrypoint/cnibincopy.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:RHEL8_SOURCE_DIRECTORY,Value:/usr/src/egress-router-cni/rhel8/bin/,ValueFrom:nil,},EnvVar{Name:RHEL9_SOURCE_DIRECTORY,Value:/usr/src/egress-router-cni/rhel9/bin/,ValueFrom:nil,},EnvVar{Name:DEFAULT_SOURCE_DIRECTORY,Value:/usr/src/egress-router-cni/bin/,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cni-binary-copy,ReadOnly:false,MountPath:/entrypoint,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:cnibin,ReadOnly:false,MountPath:/host/opt/cni/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:os-release,ReadOnly:true,MountPath:/host/etc/os-release,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-pxvhw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod multus-additional-cni-plugins-2wdp7_openshift-multus(2fdee89f-559b-4f4e-8d20-6ecc3aafd22e): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars" logger="UnhandledError" Nov 22 07:56:53 crc kubenswrapper[4789]: E1122 07:56:53.122934 4789 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09,Command:[],Args:[--secure-listen-address=0.0.0.0:9001 --config-file=/etc/kube-rbac-proxy/config-file.yaml --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12 --upstream=http://127.0.0.1:8797 --logtostderr=true --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:9001,ContainerPort:9001,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:proxy-tls,ReadOnly:false,MountPath:/etc/tls/private,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:mcd-auth-proxy-config,ReadOnly:false,MountPath:/etc/kube-rbac-proxy,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bfjgw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars" logger="UnhandledError" Nov 22 07:56:53 crc kubenswrapper[4789]: E1122 07:56:53.123295 4789 kuberuntime_manager.go:1274] "Unhandled Error" err=< Nov 22 07:56:53 crc kubenswrapper[4789]: container &Container{Name:dns-node-resolver,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2,Command:[/bin/bash -c #!/bin/bash Nov 22 07:56:53 crc kubenswrapper[4789]: set -uo pipefail Nov 22 07:56:53 crc kubenswrapper[4789]: Nov 22 07:56:53 crc kubenswrapper[4789]: trap 'jobs -p | xargs kill || true; wait; exit 0' TERM Nov 22 07:56:53 crc kubenswrapper[4789]: Nov 22 07:56:53 crc kubenswrapper[4789]: OPENSHIFT_MARKER="openshift-generated-node-resolver" Nov 22 07:56:53 crc kubenswrapper[4789]: HOSTS_FILE="/etc/hosts" Nov 22 07:56:53 crc kubenswrapper[4789]: TEMP_FILE="/etc/hosts.tmp" Nov 22 07:56:53 crc kubenswrapper[4789]: Nov 22 07:56:53 crc kubenswrapper[4789]: IFS=', ' read -r -a services <<< "${SERVICES}" Nov 22 07:56:53 crc kubenswrapper[4789]: Nov 22 07:56:53 crc kubenswrapper[4789]: # Make a temporary file with the old hosts file's attributes. Nov 22 07:56:53 crc kubenswrapper[4789]: if ! cp -f --attributes-only "${HOSTS_FILE}" "${TEMP_FILE}"; then Nov 22 07:56:53 crc kubenswrapper[4789]: echo "Failed to preserve hosts file. Exiting." Nov 22 07:56:53 crc kubenswrapper[4789]: exit 1 Nov 22 07:56:53 crc kubenswrapper[4789]: fi Nov 22 07:56:53 crc kubenswrapper[4789]: Nov 22 07:56:53 crc kubenswrapper[4789]: while true; do Nov 22 07:56:53 crc kubenswrapper[4789]: declare -A svc_ips Nov 22 07:56:53 crc kubenswrapper[4789]: for svc in "${services[@]}"; do Nov 22 07:56:53 crc kubenswrapper[4789]: # Fetch service IP from cluster dns if present. We make several tries Nov 22 07:56:53 crc kubenswrapper[4789]: # to do it: IPv4, IPv6, IPv4 over TCP and IPv6 over TCP. The two last ones Nov 22 07:56:53 crc kubenswrapper[4789]: # are for deployments with Kuryr on older OpenStack (OSP13) - those do not Nov 22 07:56:53 crc kubenswrapper[4789]: # support UDP loadbalancers and require reaching DNS through TCP. Nov 22 07:56:53 crc kubenswrapper[4789]: cmds=('dig -t A @"${NAMESERVER}" +short "${svc}.${CLUSTER_DOMAIN}"|grep -v "^;"' Nov 22 07:56:53 crc kubenswrapper[4789]: 'dig -t AAAA @"${NAMESERVER}" +short "${svc}.${CLUSTER_DOMAIN}"|grep -v "^;"' Nov 22 07:56:53 crc kubenswrapper[4789]: 'dig -t A +tcp +retry=0 @"${NAMESERVER}" +short "${svc}.${CLUSTER_DOMAIN}"|grep -v "^;"' Nov 22 07:56:53 crc kubenswrapper[4789]: 'dig -t AAAA +tcp +retry=0 @"${NAMESERVER}" +short "${svc}.${CLUSTER_DOMAIN}"|grep -v "^;"') Nov 22 07:56:53 crc kubenswrapper[4789]: for i in ${!cmds[*]} Nov 22 07:56:53 crc kubenswrapper[4789]: do Nov 22 07:56:53 crc kubenswrapper[4789]: ips=($(eval "${cmds[i]}")) Nov 22 07:56:53 crc kubenswrapper[4789]: if [[ "$?" -eq 0 && "${#ips[@]}" -ne 0 ]]; then Nov 22 07:56:53 crc kubenswrapper[4789]: svc_ips["${svc}"]="${ips[@]}" Nov 22 07:56:53 crc kubenswrapper[4789]: break Nov 22 07:56:53 crc kubenswrapper[4789]: fi Nov 22 07:56:53 crc kubenswrapper[4789]: done Nov 22 07:56:53 crc kubenswrapper[4789]: done Nov 22 07:56:53 crc kubenswrapper[4789]: Nov 22 07:56:53 crc kubenswrapper[4789]: # Update /etc/hosts only if we get valid service IPs Nov 22 07:56:53 crc kubenswrapper[4789]: # We will not update /etc/hosts when there is coredns service outage or api unavailability Nov 22 07:56:53 crc kubenswrapper[4789]: # Stale entries could exist in /etc/hosts if the service is deleted Nov 22 07:56:53 crc kubenswrapper[4789]: if [[ -n "${svc_ips[*]-}" ]]; then Nov 22 07:56:53 crc kubenswrapper[4789]: # Build a new hosts file from /etc/hosts with our custom entries filtered out Nov 22 07:56:53 crc kubenswrapper[4789]: if ! sed --silent "/# ${OPENSHIFT_MARKER}/d; w ${TEMP_FILE}" "${HOSTS_FILE}"; then Nov 22 07:56:53 crc kubenswrapper[4789]: # Only continue rebuilding the hosts entries if its original content is preserved Nov 22 07:56:53 crc kubenswrapper[4789]: sleep 60 & wait Nov 22 07:56:53 crc kubenswrapper[4789]: continue Nov 22 07:56:53 crc kubenswrapper[4789]: fi Nov 22 07:56:53 crc kubenswrapper[4789]: Nov 22 07:56:53 crc kubenswrapper[4789]: # Append resolver entries for services Nov 22 07:56:53 crc kubenswrapper[4789]: rc=0 Nov 22 07:56:53 crc kubenswrapper[4789]: for svc in "${!svc_ips[@]}"; do Nov 22 07:56:53 crc kubenswrapper[4789]: for ip in ${svc_ips[${svc}]}; do Nov 22 07:56:53 crc kubenswrapper[4789]: echo "${ip} ${svc} ${svc}.${CLUSTER_DOMAIN} # ${OPENSHIFT_MARKER}" >> "${TEMP_FILE}" || rc=$? Nov 22 07:56:53 crc kubenswrapper[4789]: done Nov 22 07:56:53 crc kubenswrapper[4789]: done Nov 22 07:56:53 crc kubenswrapper[4789]: if [[ $rc -ne 0 ]]; then Nov 22 07:56:53 crc kubenswrapper[4789]: sleep 60 & wait Nov 22 07:56:53 crc kubenswrapper[4789]: continue Nov 22 07:56:53 crc kubenswrapper[4789]: fi Nov 22 07:56:53 crc kubenswrapper[4789]: Nov 22 07:56:53 crc kubenswrapper[4789]: Nov 22 07:56:53 crc kubenswrapper[4789]: # TODO: Update /etc/hosts atomically to avoid any inconsistent behavior Nov 22 07:56:53 crc kubenswrapper[4789]: # Replace /etc/hosts with our modified version if needed Nov 22 07:56:53 crc kubenswrapper[4789]: cmp "${TEMP_FILE}" "${HOSTS_FILE}" || cp -f "${TEMP_FILE}" "${HOSTS_FILE}" Nov 22 07:56:53 crc kubenswrapper[4789]: # TEMP_FILE is not removed to avoid file create/delete and attributes copy churn Nov 22 07:56:53 crc kubenswrapper[4789]: fi Nov 22 07:56:53 crc kubenswrapper[4789]: sleep 60 & wait Nov 22 07:56:53 crc kubenswrapper[4789]: unset svc_ips Nov 22 07:56:53 crc kubenswrapper[4789]: done Nov 22 07:56:53 crc kubenswrapper[4789]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:SERVICES,Value:image-registry.openshift-image-registry.svc,ValueFrom:nil,},EnvVar{Name:NAMESERVER,Value:10.217.4.10,ValueFrom:nil,},EnvVar{Name:CLUSTER_DOMAIN,Value:cluster.local,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{22020096 0} {} 21Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:hosts-file,ReadOnly:false,MountPath:/etc/hosts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mnz6k,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod node-resolver-9f74f_openshift-dns(29ad5e00-4918-4718-9d49-013b12621f7d): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Nov 22 07:56:53 crc kubenswrapper[4789]: > logger="UnhandledError" Nov 22 07:56:53 crc kubenswrapper[4789]: I1122 07:56:53.123355 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-rks5j" event={"ID":"5613c209-59a0-4b8a-ac8b-5c633bc1bdc4","Type":"ContainerStarted","Data":"d37ce6be707cf546a6075783d075e7c1ba7e639aee014a897bcc661b50a48877"} Nov 22 07:56:53 crc kubenswrapper[4789]: E1122 07:56:53.123868 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"egress-router-binary-copy\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-multus/multus-additional-cni-plugins-2wdp7" podUID="2fdee89f-559b-4f4e-8d20-6ecc3aafd22e" Nov 22 07:56:53 crc kubenswrapper[4789]: E1122 07:56:53.124097 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"machine-config-daemon\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"]" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 07:56:53 crc kubenswrapper[4789]: E1122 07:56:53.124592 4789 kuberuntime_manager.go:1274] "Unhandled Error" err=< Nov 22 07:56:53 crc kubenswrapper[4789]: container &Container{Name:kube-multus,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26,Command:[/bin/bash -ec --],Args:[MULTUS_DAEMON_OPT="" Nov 22 07:56:53 crc kubenswrapper[4789]: /entrypoint/cnibincopy.sh; exec /usr/src/multus-cni/bin/multus-daemon $MULTUS_DAEMON_OPT Nov 22 07:56:53 crc kubenswrapper[4789]: ],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:RHEL8_SOURCE_DIRECTORY,Value:/usr/src/multus-cni/rhel8/bin/,ValueFrom:nil,},EnvVar{Name:RHEL9_SOURCE_DIRECTORY,Value:/usr/src/multus-cni/rhel9/bin/,ValueFrom:nil,},EnvVar{Name:DEFAULT_SOURCE_DIRECTORY,Value:/usr/src/multus-cni/bin/,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:6443,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:api-int.crc.testing,ValueFrom:nil,},EnvVar{Name:MULTUS_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:K8S_NODE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{68157440 0} {} 65Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cni-binary-copy,ReadOnly:false,MountPath:/entrypoint,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:os-release,ReadOnly:false,MountPath:/host/etc/os-release,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:system-cni-dir,ReadOnly:false,MountPath:/host/etc/cni/net.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:multus-cni-dir,ReadOnly:false,MountPath:/host/run/multus/cni/net.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:cnibin,ReadOnly:false,MountPath:/host/opt/cni/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:multus-socket-dir-parent,ReadOnly:false,MountPath:/host/run/multus,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-run-k8s-cni-cncf-io,ReadOnly:false,MountPath:/run/k8s.cni.cncf.io,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-run-netns,ReadOnly:false,MountPath:/run/netns,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-var-lib-cni-bin,ReadOnly:false,MountPath:/var/lib/cni/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-var-lib-cni-multus,ReadOnly:false,MountPath:/var/lib/cni/multus,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-var-lib-kubelet,ReadOnly:false,MountPath:/var/lib/kubelet,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:hostroot,ReadOnly:false,MountPath:/hostroot,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:multus-conf-dir,ReadOnly:false,MountPath:/etc/cni/multus/net.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:multus-daemon-config,ReadOnly:true,MountPath:/etc/cni/net.d/multus.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-run-multus-certs,ReadOnly:false,MountPath:/etc/cni/multus/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:etc-kubernetes,ReadOnly:false,MountPath:/etc/kubernetes,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wj8sc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod multus-rks5j_openshift-multus(5613c209-59a0-4b8a-ac8b-5c633bc1bdc4): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Nov 22 07:56:53 crc kubenswrapper[4789]: > logger="UnhandledError" Nov 22 07:56:53 crc kubenswrapper[4789]: E1122 07:56:53.124598 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"dns-node-resolver\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-dns/node-resolver-9f74f" podUID="29ad5e00-4918-4718-9d49-013b12621f7d" Nov 22 07:56:53 crc kubenswrapper[4789]: E1122 07:56:53.125713 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-multus/multus-rks5j" podUID="5613c209-59a0-4b8a-ac8b-5c633bc1bdc4" Nov 22 07:56:53 crc kubenswrapper[4789]: I1122 07:56:53.159405 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:53 crc kubenswrapper[4789]: I1122 07:56:53.187349 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:53 crc kubenswrapper[4789]: I1122 07:56:53.187392 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:53 crc kubenswrapper[4789]: I1122 07:56:53.187405 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:53 crc kubenswrapper[4789]: I1122 07:56:53.187421 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:56:53 crc kubenswrapper[4789]: I1122 07:56:53.187432 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:56:53Z","lastTransitionTime":"2025-11-22T07:56:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:56:53 crc kubenswrapper[4789]: I1122 07:56:53.203777 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0cb4cf51-62d6-4f32-a8b4-4116555f3870\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qmxss\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:53 crc kubenswrapper[4789]: I1122 07:56:53.242769 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:53 crc kubenswrapper[4789]: I1122 07:56:53.277412 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9f74f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"29ad5e00-4918-4718-9d49-013b12621f7d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mnz6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9f74f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:53 crc kubenswrapper[4789]: I1122 07:56:53.290156 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:53 crc kubenswrapper[4789]: I1122 07:56:53.290211 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:53 crc kubenswrapper[4789]: I1122 07:56:53.290223 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:53 crc kubenswrapper[4789]: I1122 07:56:53.290240 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:56:53 crc kubenswrapper[4789]: I1122 07:56:53.290251 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:56:53Z","lastTransitionTime":"2025-11-22T07:56:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:56:53 crc kubenswrapper[4789]: I1122 07:56:53.321275 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2wdp7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2fdee89f-559b-4f4e-8d20-6ecc3aafd22e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxvhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxvhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxvhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxvhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxvhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxvhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxvhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2wdp7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:53 crc kubenswrapper[4789]: I1122 07:56:53.360179 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:53 crc kubenswrapper[4789]: I1122 07:56:53.393559 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:53 crc kubenswrapper[4789]: I1122 07:56:53.393627 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:53 crc kubenswrapper[4789]: I1122 07:56:53.393642 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:53 crc kubenswrapper[4789]: I1122 07:56:53.393661 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:56:53 crc kubenswrapper[4789]: I1122 07:56:53.393676 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:56:53Z","lastTransitionTime":"2025-11-22T07:56:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:56:53 crc kubenswrapper[4789]: I1122 07:56:53.403704 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:53 crc kubenswrapper[4789]: I1122 07:56:53.439087 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58aa8071-7ecc-4692-8789-2db04ec70647\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfjgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfjgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pxvrm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:53 crc kubenswrapper[4789]: I1122 07:56:53.481511 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d081b29e-edd7-4fb0-a2e5-431ee8491bd0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43d9e10254883b6491bef617cf9f0b319171bc1fbc49c7770ed9630a8a1ae2cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://223c2300bb6afb8812a72eb7605015b0473f59046565cdf39fb9b9f0cf6634bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca5af01b2e23b2bcac23ed767bc7dcd2ac0be1010ee5e675b3b44333ee337180\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8e79bfb6172c894868a8e1b47ac701ba4fdaca1fc6b1582f7e70807aadfdf1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:29Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:53 crc kubenswrapper[4789]: I1122 07:56:53.496360 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:53 crc kubenswrapper[4789]: I1122 07:56:53.496405 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:53 crc kubenswrapper[4789]: I1122 07:56:53.496417 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:53 crc kubenswrapper[4789]: I1122 07:56:53.496433 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:56:53 crc kubenswrapper[4789]: I1122 07:56:53.496444 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:56:53Z","lastTransitionTime":"2025-11-22T07:56:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:56:53 crc kubenswrapper[4789]: I1122 07:56:53.505903 4789 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 22 07:56:53 crc kubenswrapper[4789]: I1122 07:56:53.506860 4789 scope.go:117] "RemoveContainer" containerID="098c31b822e7c2351f1142af36b942d67ff8262896e28022c96a187d40bd0ddb" Nov 22 07:56:53 crc kubenswrapper[4789]: E1122 07:56:53.507074 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Nov 22 07:56:53 crc kubenswrapper[4789]: I1122 07:56:53.529784 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66f74155-c96e-4faa-876f-62b8b5e6b3bc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53953e73ae4a11a47b7a5332786b97d9f129d9576e66aa2664836ec5f3786683\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65813cc305e2105a598eb0b701dd079f1f26db8cc1b2e67679e7086c2ba3630f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1d47e3e0a1939863a1ae9e1f52262d8fad029cd90552c21e73a98cdf0a803ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9d09f0deafe8d22df512d2acd120ce10eec182173daaf5029ab72f848ceecd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c410501c102d72819d29aa0036b6270b1ce9804bba84092161b2c2d8ebaa885a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0755dd1f1e7b1fda23c8973b56521ab15b62b9eb76f9b5604110e04e2f25b422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0755dd1f1e7b1fda23c8973b56521ab15b62b9eb76f9b5604110e04e2f25b422\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T07:56:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T07:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://680b49efb88db68b75efc6ca47cdfa449c1cf688d277b3e1eaa75cb466e90480\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://680b49efb88db68b75efc6ca47cdfa449c1cf688d277b3e1eaa75cb466e90480\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T07:56:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T07:56:32Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c5b1cf28577e60931184499a22559f30b0bdf5dab0b6cb542c948fec5d9c54cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5b1cf28577e60931184499a22559f30b0bdf5dab0b6cb542c948fec5d9c54cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T07:56:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T07:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:29Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:53 crc kubenswrapper[4789]: I1122 07:56:53.561949 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:53 crc kubenswrapper[4789]: I1122 07:56:53.598834 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:53 crc kubenswrapper[4789]: I1122 07:56:53.598887 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:53 crc kubenswrapper[4789]: I1122 07:56:53.598898 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:53 crc kubenswrapper[4789]: I1122 07:56:53.598916 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:56:53 crc kubenswrapper[4789]: I1122 07:56:53.598931 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:56:53Z","lastTransitionTime":"2025-11-22T07:56:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:56:53 crc kubenswrapper[4789]: I1122 07:56:53.600665 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:53 crc kubenswrapper[4789]: I1122 07:56:53.647532 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0cb4cf51-62d6-4f32-a8b4-4116555f3870\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qmxss\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:53 crc kubenswrapper[4789]: I1122 07:56:53.679015 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:53 crc kubenswrapper[4789]: I1122 07:56:53.701647 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:53 crc kubenswrapper[4789]: I1122 07:56:53.701684 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:53 crc kubenswrapper[4789]: I1122 07:56:53.701694 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:53 crc kubenswrapper[4789]: I1122 07:56:53.701710 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:56:53 crc kubenswrapper[4789]: I1122 07:56:53.701721 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:56:53Z","lastTransitionTime":"2025-11-22T07:56:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:56:53 crc kubenswrapper[4789]: I1122 07:56:53.720906 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:53 crc kubenswrapper[4789]: I1122 07:56:53.758466 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9f74f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"29ad5e00-4918-4718-9d49-013b12621f7d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mnz6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9f74f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:53 crc kubenswrapper[4789]: I1122 07:56:53.801904 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2wdp7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2fdee89f-559b-4f4e-8d20-6ecc3aafd22e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxvhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxvhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxvhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxvhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxvhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxvhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxvhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2wdp7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:53 crc kubenswrapper[4789]: I1122 07:56:53.803278 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:53 crc kubenswrapper[4789]: I1122 07:56:53.803304 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:53 crc kubenswrapper[4789]: I1122 07:56:53.803314 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:53 crc kubenswrapper[4789]: I1122 07:56:53.803326 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:56:53 crc kubenswrapper[4789]: I1122 07:56:53.803335 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:56:53Z","lastTransitionTime":"2025-11-22T07:56:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:56:53 crc kubenswrapper[4789]: I1122 07:56:53.839128 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d081b29e-edd7-4fb0-a2e5-431ee8491bd0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43d9e10254883b6491bef617cf9f0b319171bc1fbc49c7770ed9630a8a1ae2cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://223c2300bb6afb8812a72eb7605015b0473f59046565cdf39fb9b9f0cf6634bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca5af01b2e23b2bcac23ed767bc7dcd2ac0be1010ee5e675b3b44333ee337180\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8e79bfb6172c894868a8e1b47ac701ba4fdaca1fc6b1582f7e70807aadfdf1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:29Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:53 crc kubenswrapper[4789]: I1122 07:56:53.880241 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:53 crc kubenswrapper[4789]: I1122 07:56:53.905521 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:53 crc kubenswrapper[4789]: I1122 07:56:53.905560 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:53 crc kubenswrapper[4789]: I1122 07:56:53.905569 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:53 crc kubenswrapper[4789]: I1122 07:56:53.905586 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:56:53 crc kubenswrapper[4789]: I1122 07:56:53.905595 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:56:53Z","lastTransitionTime":"2025-11-22T07:56:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:56:53 crc kubenswrapper[4789]: I1122 07:56:53.921061 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58aa8071-7ecc-4692-8789-2db04ec70647\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfjgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfjgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pxvrm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:53 crc kubenswrapper[4789]: I1122 07:56:53.962039 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93be48ce-7d87-448b-b03e-ab6736cfc34c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba91466edcf49f73751622f98632b3040a046ccd8a42ba8ecee7ab08688a8c24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b6666e0ce42a46f669505659298a19931b1bf7c253ae6a97c6acfd80a9e717f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://251e72e1cd0f3e0ee195bacf33ab38765d1dd943adf4395f9d6dd282bda024cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://098c31b822e7c2351f1142af36b942d67ff8262896e28022c96a187d40bd0ddb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://098c31b822e7c2351f1142af36b942d67ff8262896e28022c96a187d40bd0ddb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"message\\\":\\\"le observer\\\\nW1122 07:56:50.641735 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1122 07:56:50.642792 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 07:56:50.643945 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3588433834/tls.crt::/tmp/serving-cert-3588433834/tls.key\\\\\\\"\\\\nI1122 07:56:50.949040 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 07:56:50.952068 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 07:56:50.952092 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 07:56:50.952116 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 07:56:50.952121 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 07:56:50.958549 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1122 07:56:50.958565 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1122 07:56:50.958579 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 07:56:50.958585 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 07:56:50.958591 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 07:56:50.958596 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1122 07:56:50.958600 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 07:56:50.958605 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1122 07:56:50.962189 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T07:56:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://263dce4c3f4f3fe388b21d4bb55a9fc7d956c3cca7da34ee00f470d88ef501f1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac2b567fcebc517d3fb4366aebce03351183ea03ce62f86c2abdf5b95e20b34a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac2b567fcebc517d3fb4366aebce03351183ea03ce62f86c2abdf5b95e20b34a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T07:56:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T07:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:53 crc kubenswrapper[4789]: I1122 07:56:53.964831 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 07:56:53 crc kubenswrapper[4789]: E1122 07:56:53.964966 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 07:56:54 crc kubenswrapper[4789]: I1122 07:56:54.001215 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:54 crc kubenswrapper[4789]: I1122 07:56:54.008273 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:54 crc kubenswrapper[4789]: I1122 07:56:54.008319 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:54 crc kubenswrapper[4789]: I1122 07:56:54.008334 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:54 crc kubenswrapper[4789]: I1122 07:56:54.008351 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:56:54 crc kubenswrapper[4789]: I1122 07:56:54.008365 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:56:54Z","lastTransitionTime":"2025-11-22T07:56:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:56:54 crc kubenswrapper[4789]: I1122 07:56:54.022432 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-vrkrh"] Nov 22 07:56:54 crc kubenswrapper[4789]: I1122 07:56:54.022845 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-vrkrh" Nov 22 07:56:54 crc kubenswrapper[4789]: I1122 07:56:54.044728 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rks5j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5613c209-59a0-4b8a-ac8b-5c633bc1bdc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wj8sc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rks5j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:54 crc kubenswrapper[4789]: I1122 07:56:54.054974 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Nov 22 07:56:54 crc kubenswrapper[4789]: I1122 07:56:54.075018 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Nov 22 07:56:54 crc kubenswrapper[4789]: I1122 07:56:54.093877 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Nov 22 07:56:54 crc kubenswrapper[4789]: I1122 07:56:54.110032 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:54 crc kubenswrapper[4789]: I1122 07:56:54.110094 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:54 crc kubenswrapper[4789]: I1122 07:56:54.110111 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:54 crc kubenswrapper[4789]: I1122 07:56:54.110134 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:56:54 crc kubenswrapper[4789]: I1122 07:56:54.110150 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:56:54Z","lastTransitionTime":"2025-11-22T07:56:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:56:54 crc kubenswrapper[4789]: I1122 07:56:54.114018 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Nov 22 07:56:54 crc kubenswrapper[4789]: I1122 07:56:54.117349 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/371275b1-1d13-4f3c-8190-8c59c080fef5-serviceca\") pod \"node-ca-vrkrh\" (UID: \"371275b1-1d13-4f3c-8190-8c59c080fef5\") " pod="openshift-image-registry/node-ca-vrkrh" Nov 22 07:56:54 crc kubenswrapper[4789]: I1122 07:56:54.117500 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/371275b1-1d13-4f3c-8190-8c59c080fef5-host\") pod \"node-ca-vrkrh\" (UID: \"371275b1-1d13-4f3c-8190-8c59c080fef5\") " pod="openshift-image-registry/node-ca-vrkrh" Nov 22 07:56:54 crc kubenswrapper[4789]: I1122 07:56:54.117606 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8w4z2\" (UniqueName: \"kubernetes.io/projected/371275b1-1d13-4f3c-8190-8c59c080fef5-kube-api-access-8w4z2\") pod \"node-ca-vrkrh\" (UID: \"371275b1-1d13-4f3c-8190-8c59c080fef5\") " pod="openshift-image-registry/node-ca-vrkrh" Nov 22 07:56:54 crc kubenswrapper[4789]: I1122 07:56:54.164599 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:54 crc kubenswrapper[4789]: I1122 07:56:54.200926 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:54 crc kubenswrapper[4789]: I1122 07:56:54.212697 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:54 crc kubenswrapper[4789]: I1122 07:56:54.212782 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:54 crc kubenswrapper[4789]: I1122 07:56:54.212797 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:54 crc kubenswrapper[4789]: I1122 07:56:54.212812 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:56:54 crc kubenswrapper[4789]: I1122 07:56:54.212838 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:56:54Z","lastTransitionTime":"2025-11-22T07:56:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:56:54 crc kubenswrapper[4789]: I1122 07:56:54.218625 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/371275b1-1d13-4f3c-8190-8c59c080fef5-host\") pod \"node-ca-vrkrh\" (UID: \"371275b1-1d13-4f3c-8190-8c59c080fef5\") " pod="openshift-image-registry/node-ca-vrkrh" Nov 22 07:56:54 crc kubenswrapper[4789]: I1122 07:56:54.218687 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8w4z2\" (UniqueName: \"kubernetes.io/projected/371275b1-1d13-4f3c-8190-8c59c080fef5-kube-api-access-8w4z2\") pod \"node-ca-vrkrh\" (UID: \"371275b1-1d13-4f3c-8190-8c59c080fef5\") " pod="openshift-image-registry/node-ca-vrkrh" Nov 22 07:56:54 crc kubenswrapper[4789]: I1122 07:56:54.218694 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/371275b1-1d13-4f3c-8190-8c59c080fef5-host\") pod \"node-ca-vrkrh\" (UID: \"371275b1-1d13-4f3c-8190-8c59c080fef5\") " pod="openshift-image-registry/node-ca-vrkrh" Nov 22 07:56:54 crc kubenswrapper[4789]: I1122 07:56:54.218727 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/371275b1-1d13-4f3c-8190-8c59c080fef5-serviceca\") pod \"node-ca-vrkrh\" (UID: \"371275b1-1d13-4f3c-8190-8c59c080fef5\") " pod="openshift-image-registry/node-ca-vrkrh" Nov 22 07:56:54 crc kubenswrapper[4789]: I1122 07:56:54.219586 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/371275b1-1d13-4f3c-8190-8c59c080fef5-serviceca\") pod \"node-ca-vrkrh\" (UID: \"371275b1-1d13-4f3c-8190-8c59c080fef5\") " pod="openshift-image-registry/node-ca-vrkrh" Nov 22 07:56:54 crc kubenswrapper[4789]: I1122 07:56:54.249880 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0cb4cf51-62d6-4f32-a8b4-4116555f3870\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qmxss\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:54 crc kubenswrapper[4789]: I1122 07:56:54.276656 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8w4z2\" (UniqueName: \"kubernetes.io/projected/371275b1-1d13-4f3c-8190-8c59c080fef5-kube-api-access-8w4z2\") pod \"node-ca-vrkrh\" (UID: \"371275b1-1d13-4f3c-8190-8c59c080fef5\") " pod="openshift-image-registry/node-ca-vrkrh" Nov 22 07:56:54 crc kubenswrapper[4789]: I1122 07:56:54.315969 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:54 crc kubenswrapper[4789]: I1122 07:56:54.316025 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:54 crc kubenswrapper[4789]: I1122 07:56:54.316048 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:54 crc kubenswrapper[4789]: I1122 07:56:54.316080 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:56:54 crc kubenswrapper[4789]: I1122 07:56:54.316104 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:56:54Z","lastTransitionTime":"2025-11-22T07:56:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:56:54 crc kubenswrapper[4789]: I1122 07:56:54.317460 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66f74155-c96e-4faa-876f-62b8b5e6b3bc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53953e73ae4a11a47b7a5332786b97d9f129d9576e66aa2664836ec5f3786683\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65813cc305e2105a598eb0b701dd079f1f26db8cc1b2e67679e7086c2ba3630f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1d47e3e0a1939863a1ae9e1f52262d8fad029cd90552c21e73a98cdf0a803ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9d09f0deafe8d22df512d2acd120ce10eec182173daaf5029ab72f848ceecd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c410501c102d72819d29aa0036b6270b1ce9804bba84092161b2c2d8ebaa885a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0755dd1f1e7b1fda23c8973b56521ab15b62b9eb76f9b5604110e04e2f25b422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0755dd1f1e7b1fda23c8973b56521ab15b62b9eb76f9b5604110e04e2f25b422\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T07:56:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T07:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://680b49efb88db68b75efc6ca47cdfa449c1cf688d277b3e1eaa75cb466e90480\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://680b49efb88db68b75efc6ca47cdfa449c1cf688d277b3e1eaa75cb466e90480\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T07:56:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T07:56:32Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c5b1cf28577e60931184499a22559f30b0bdf5dab0b6cb542c948fec5d9c54cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5b1cf28577e60931184499a22559f30b0bdf5dab0b6cb542c948fec5d9c54cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T07:56:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T07:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:29Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:54 crc kubenswrapper[4789]: I1122 07:56:54.319953 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 07:56:54 crc kubenswrapper[4789]: E1122 07:56:54.320123 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 07:56:58.320096814 +0000 UTC m=+32.554497127 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:56:54 crc kubenswrapper[4789]: I1122 07:56:54.333155 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-vrkrh" Nov 22 07:56:54 crc kubenswrapper[4789]: I1122 07:56:54.340815 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9f74f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"29ad5e00-4918-4718-9d49-013b12621f7d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mnz6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9f74f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:54 crc kubenswrapper[4789]: E1122 07:56:54.347919 4789 kuberuntime_manager.go:1274] "Unhandled Error" err=< Nov 22 07:56:54 crc kubenswrapper[4789]: container &Container{Name:node-ca,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f,Command:[/bin/sh -c trap 'jobs -p | xargs -r kill; echo shutting down node-ca; exit 0' TERM Nov 22 07:56:54 crc kubenswrapper[4789]: while [ true ]; Nov 22 07:56:54 crc kubenswrapper[4789]: do Nov 22 07:56:54 crc kubenswrapper[4789]: for f in $(ls /tmp/serviceca); do Nov 22 07:56:54 crc kubenswrapper[4789]: echo $f Nov 22 07:56:54 crc kubenswrapper[4789]: ca_file_path="/tmp/serviceca/${f}" Nov 22 07:56:54 crc kubenswrapper[4789]: f=$(echo $f | sed -r 's/(.*)\.\./\1:/') Nov 22 07:56:54 crc kubenswrapper[4789]: reg_dir_path="/etc/docker/certs.d/${f}" Nov 22 07:56:54 crc kubenswrapper[4789]: if [ -e "${reg_dir_path}" ]; then Nov 22 07:56:54 crc kubenswrapper[4789]: cp -u $ca_file_path $reg_dir_path/ca.crt Nov 22 07:56:54 crc kubenswrapper[4789]: else Nov 22 07:56:54 crc kubenswrapper[4789]: mkdir $reg_dir_path Nov 22 07:56:54 crc kubenswrapper[4789]: cp $ca_file_path $reg_dir_path/ca.crt Nov 22 07:56:54 crc kubenswrapper[4789]: fi Nov 22 07:56:54 crc kubenswrapper[4789]: done Nov 22 07:56:54 crc kubenswrapper[4789]: for d in $(ls /etc/docker/certs.d); do Nov 22 07:56:54 crc kubenswrapper[4789]: echo $d Nov 22 07:56:54 crc kubenswrapper[4789]: dp=$(echo $d | sed -r 's/(.*):/\1\.\./') Nov 22 07:56:54 crc kubenswrapper[4789]: reg_conf_path="/tmp/serviceca/${dp}" Nov 22 07:56:54 crc kubenswrapper[4789]: if [ ! -e "${reg_conf_path}" ]; then Nov 22 07:56:54 crc kubenswrapper[4789]: rm -rf /etc/docker/certs.d/$d Nov 22 07:56:54 crc kubenswrapper[4789]: fi Nov 22 07:56:54 crc kubenswrapper[4789]: done Nov 22 07:56:54 crc kubenswrapper[4789]: sleep 60 & wait ${!} Nov 22 07:56:54 crc kubenswrapper[4789]: done Nov 22 07:56:54 crc kubenswrapper[4789]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{10485760 0} {} 10Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:serviceca,ReadOnly:false,MountPath:/tmp/serviceca,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host,ReadOnly:false,MountPath:/etc/docker/certs.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8w4z2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:*1001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod node-ca-vrkrh_openshift-image-registry(371275b1-1d13-4f3c-8190-8c59c080fef5): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Nov 22 07:56:54 crc kubenswrapper[4789]: > logger="UnhandledError" Nov 22 07:56:54 crc kubenswrapper[4789]: E1122 07:56:54.349439 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"node-ca\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-image-registry/node-ca-vrkrh" podUID="371275b1-1d13-4f3c-8190-8c59c080fef5" Nov 22 07:56:54 crc kubenswrapper[4789]: I1122 07:56:54.385183 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2wdp7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2fdee89f-559b-4f4e-8d20-6ecc3aafd22e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxvhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxvhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxvhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxvhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxvhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxvhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxvhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2wdp7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:54 crc kubenswrapper[4789]: I1122 07:56:54.419215 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:54 crc kubenswrapper[4789]: I1122 07:56:54.419256 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:54 crc kubenswrapper[4789]: I1122 07:56:54.419267 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:54 crc kubenswrapper[4789]: I1122 07:56:54.419279 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:56:54 crc kubenswrapper[4789]: I1122 07:56:54.419288 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:56:54Z","lastTransitionTime":"2025-11-22T07:56:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:56:54 crc kubenswrapper[4789]: I1122 07:56:54.420802 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 07:56:54 crc kubenswrapper[4789]: I1122 07:56:54.420852 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 07:56:54 crc kubenswrapper[4789]: I1122 07:56:54.420882 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 07:56:54 crc kubenswrapper[4789]: I1122 07:56:54.420907 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 07:56:54 crc kubenswrapper[4789]: E1122 07:56:54.421003 4789 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 22 07:56:54 crc kubenswrapper[4789]: E1122 07:56:54.421052 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-22 07:56:58.421035657 +0000 UTC m=+32.655435930 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 22 07:56:54 crc kubenswrapper[4789]: E1122 07:56:54.421236 4789 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 22 07:56:54 crc kubenswrapper[4789]: E1122 07:56:54.421267 4789 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 22 07:56:54 crc kubenswrapper[4789]: E1122 07:56:54.421277 4789 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 22 07:56:54 crc kubenswrapper[4789]: E1122 07:56:54.421287 4789 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 07:56:54 crc kubenswrapper[4789]: E1122 07:56:54.421327 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-22 07:56:58.421311934 +0000 UTC m=+32.655712207 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 22 07:56:54 crc kubenswrapper[4789]: E1122 07:56:54.421345 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-22 07:56:58.421336515 +0000 UTC m=+32.655736788 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 07:56:54 crc kubenswrapper[4789]: E1122 07:56:54.421343 4789 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 22 07:56:54 crc kubenswrapper[4789]: E1122 07:56:54.421388 4789 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 22 07:56:54 crc kubenswrapper[4789]: E1122 07:56:54.421409 4789 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 07:56:54 crc kubenswrapper[4789]: E1122 07:56:54.421502 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-22 07:56:58.421476498 +0000 UTC m=+32.655876811 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 07:56:54 crc kubenswrapper[4789]: I1122 07:56:54.421923 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:54 crc kubenswrapper[4789]: I1122 07:56:54.463303 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:54 crc kubenswrapper[4789]: I1122 07:56:54.499901 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:54 crc kubenswrapper[4789]: I1122 07:56:54.521172 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:54 crc kubenswrapper[4789]: I1122 07:56:54.521228 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:54 crc kubenswrapper[4789]: I1122 07:56:54.521244 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:54 crc kubenswrapper[4789]: I1122 07:56:54.521262 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:56:54 crc kubenswrapper[4789]: I1122 07:56:54.521319 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:56:54Z","lastTransitionTime":"2025-11-22T07:56:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:56:54 crc kubenswrapper[4789]: I1122 07:56:54.540664 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58aa8071-7ecc-4692-8789-2db04ec70647\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfjgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfjgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pxvrm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:54 crc kubenswrapper[4789]: I1122 07:56:54.581362 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vrkrh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"371275b1-1d13-4f3c-8190-8c59c080fef5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:54Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:54Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8w4z2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vrkrh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:54 crc kubenswrapper[4789]: I1122 07:56:54.621921 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d081b29e-edd7-4fb0-a2e5-431ee8491bd0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43d9e10254883b6491bef617cf9f0b319171bc1fbc49c7770ed9630a8a1ae2cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://223c2300bb6afb8812a72eb7605015b0473f59046565cdf39fb9b9f0cf6634bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca5af01b2e23b2bcac23ed767bc7dcd2ac0be1010ee5e675b3b44333ee337180\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8e79bfb6172c894868a8e1b47ac701ba4fdaca1fc6b1582f7e70807aadfdf1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:29Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:54 crc kubenswrapper[4789]: I1122 07:56:54.623394 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:54 crc kubenswrapper[4789]: I1122 07:56:54.623441 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:54 crc kubenswrapper[4789]: I1122 07:56:54.623453 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:54 crc kubenswrapper[4789]: I1122 07:56:54.623471 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:56:54 crc kubenswrapper[4789]: I1122 07:56:54.623485 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:56:54Z","lastTransitionTime":"2025-11-22T07:56:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:56:54 crc kubenswrapper[4789]: I1122 07:56:54.661848 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rks5j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5613c209-59a0-4b8a-ac8b-5c633bc1bdc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wj8sc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rks5j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:54 crc kubenswrapper[4789]: I1122 07:56:54.701501 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93be48ce-7d87-448b-b03e-ab6736cfc34c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba91466edcf49f73751622f98632b3040a046ccd8a42ba8ecee7ab08688a8c24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b6666e0ce42a46f669505659298a19931b1bf7c253ae6a97c6acfd80a9e717f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://251e72e1cd0f3e0ee195bacf33ab38765d1dd943adf4395f9d6dd282bda024cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://098c31b822e7c2351f1142af36b942d67ff8262896e28022c96a187d40bd0ddb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://098c31b822e7c2351f1142af36b942d67ff8262896e28022c96a187d40bd0ddb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"message\\\":\\\"le observer\\\\nW1122 07:56:50.641735 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1122 07:56:50.642792 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 07:56:50.643945 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3588433834/tls.crt::/tmp/serving-cert-3588433834/tls.key\\\\\\\"\\\\nI1122 07:56:50.949040 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 07:56:50.952068 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 07:56:50.952092 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 07:56:50.952116 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 07:56:50.952121 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 07:56:50.958549 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1122 07:56:50.958565 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1122 07:56:50.958579 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 07:56:50.958585 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 07:56:50.958591 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 07:56:50.958596 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1122 07:56:50.958600 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 07:56:50.958605 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1122 07:56:50.962189 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T07:56:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://263dce4c3f4f3fe388b21d4bb55a9fc7d956c3cca7da34ee00f470d88ef501f1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac2b567fcebc517d3fb4366aebce03351183ea03ce62f86c2abdf5b95e20b34a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac2b567fcebc517d3fb4366aebce03351183ea03ce62f86c2abdf5b95e20b34a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T07:56:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T07:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:54 crc kubenswrapper[4789]: I1122 07:56:54.726269 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:54 crc kubenswrapper[4789]: I1122 07:56:54.726322 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:54 crc kubenswrapper[4789]: I1122 07:56:54.726340 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:54 crc kubenswrapper[4789]: I1122 07:56:54.726363 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:56:54 crc kubenswrapper[4789]: I1122 07:56:54.726383 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:56:54Z","lastTransitionTime":"2025-11-22T07:56:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:56:54 crc kubenswrapper[4789]: I1122 07:56:54.744068 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:54 crc kubenswrapper[4789]: I1122 07:56:54.829356 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:54 crc kubenswrapper[4789]: I1122 07:56:54.829504 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:54 crc kubenswrapper[4789]: I1122 07:56:54.829531 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:54 crc kubenswrapper[4789]: I1122 07:56:54.829565 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:56:54 crc kubenswrapper[4789]: I1122 07:56:54.829589 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:56:54Z","lastTransitionTime":"2025-11-22T07:56:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:56:54 crc kubenswrapper[4789]: I1122 07:56:54.933515 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:54 crc kubenswrapper[4789]: I1122 07:56:54.933568 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:54 crc kubenswrapper[4789]: I1122 07:56:54.933580 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:54 crc kubenswrapper[4789]: I1122 07:56:54.933599 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:56:54 crc kubenswrapper[4789]: I1122 07:56:54.933611 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:56:54Z","lastTransitionTime":"2025-11-22T07:56:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:56:54 crc kubenswrapper[4789]: I1122 07:56:54.956005 4789 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Nov 22 07:56:54 crc kubenswrapper[4789]: I1122 07:56:54.964470 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 07:56:54 crc kubenswrapper[4789]: I1122 07:56:54.964513 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 07:56:54 crc kubenswrapper[4789]: E1122 07:56:54.964664 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 07:56:54 crc kubenswrapper[4789]: E1122 07:56:54.964790 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 07:56:55 crc kubenswrapper[4789]: I1122 07:56:55.036793 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:55 crc kubenswrapper[4789]: I1122 07:56:55.036851 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:55 crc kubenswrapper[4789]: I1122 07:56:55.036872 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:55 crc kubenswrapper[4789]: I1122 07:56:55.036891 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:56:55 crc kubenswrapper[4789]: I1122 07:56:55.036904 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:56:55Z","lastTransitionTime":"2025-11-22T07:56:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:56:55 crc kubenswrapper[4789]: I1122 07:56:55.130276 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-vrkrh" event={"ID":"371275b1-1d13-4f3c-8190-8c59c080fef5","Type":"ContainerStarted","Data":"736aedd65d4f2c7a4a1a81fe0ef3d672bd8690f36600de63340c37a88b3bc705"} Nov 22 07:56:55 crc kubenswrapper[4789]: E1122 07:56:55.132319 4789 kuberuntime_manager.go:1274] "Unhandled Error" err=< Nov 22 07:56:55 crc kubenswrapper[4789]: container &Container{Name:node-ca,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f,Command:[/bin/sh -c trap 'jobs -p | xargs -r kill; echo shutting down node-ca; exit 0' TERM Nov 22 07:56:55 crc kubenswrapper[4789]: while [ true ]; Nov 22 07:56:55 crc kubenswrapper[4789]: do Nov 22 07:56:55 crc kubenswrapper[4789]: for f in $(ls /tmp/serviceca); do Nov 22 07:56:55 crc kubenswrapper[4789]: echo $f Nov 22 07:56:55 crc kubenswrapper[4789]: ca_file_path="/tmp/serviceca/${f}" Nov 22 07:56:55 crc kubenswrapper[4789]: f=$(echo $f | sed -r 's/(.*)\.\./\1:/') Nov 22 07:56:55 crc kubenswrapper[4789]: reg_dir_path="/etc/docker/certs.d/${f}" Nov 22 07:56:55 crc kubenswrapper[4789]: if [ -e "${reg_dir_path}" ]; then Nov 22 07:56:55 crc kubenswrapper[4789]: cp -u $ca_file_path $reg_dir_path/ca.crt Nov 22 07:56:55 crc kubenswrapper[4789]: else Nov 22 07:56:55 crc kubenswrapper[4789]: mkdir $reg_dir_path Nov 22 07:56:55 crc kubenswrapper[4789]: cp $ca_file_path $reg_dir_path/ca.crt Nov 22 07:56:55 crc kubenswrapper[4789]: fi Nov 22 07:56:55 crc kubenswrapper[4789]: done Nov 22 07:56:55 crc kubenswrapper[4789]: for d in $(ls /etc/docker/certs.d); do Nov 22 07:56:55 crc kubenswrapper[4789]: echo $d Nov 22 07:56:55 crc kubenswrapper[4789]: dp=$(echo $d | sed -r 's/(.*):/\1\.\./') Nov 22 07:56:55 crc kubenswrapper[4789]: reg_conf_path="/tmp/serviceca/${dp}" Nov 22 07:56:55 crc kubenswrapper[4789]: if [ ! -e "${reg_conf_path}" ]; then Nov 22 07:56:55 crc kubenswrapper[4789]: rm -rf /etc/docker/certs.d/$d Nov 22 07:56:55 crc kubenswrapper[4789]: fi Nov 22 07:56:55 crc kubenswrapper[4789]: done Nov 22 07:56:55 crc kubenswrapper[4789]: sleep 60 & wait ${!} Nov 22 07:56:55 crc kubenswrapper[4789]: done Nov 22 07:56:55 crc kubenswrapper[4789]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{10485760 0} {} 10Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:serviceca,ReadOnly:false,MountPath:/tmp/serviceca,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host,ReadOnly:false,MountPath:/etc/docker/certs.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8w4z2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:*1001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod node-ca-vrkrh_openshift-image-registry(371275b1-1d13-4f3c-8190-8c59c080fef5): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Nov 22 07:56:55 crc kubenswrapper[4789]: > logger="UnhandledError" Nov 22 07:56:55 crc kubenswrapper[4789]: E1122 07:56:55.133550 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"node-ca\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-image-registry/node-ca-vrkrh" podUID="371275b1-1d13-4f3c-8190-8c59c080fef5" Nov 22 07:56:55 crc kubenswrapper[4789]: I1122 07:56:55.139529 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:55 crc kubenswrapper[4789]: I1122 07:56:55.139586 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:55 crc kubenswrapper[4789]: I1122 07:56:55.139598 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:55 crc kubenswrapper[4789]: I1122 07:56:55.139617 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:56:55 crc kubenswrapper[4789]: I1122 07:56:55.139630 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:56:55Z","lastTransitionTime":"2025-11-22T07:56:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:56:55 crc kubenswrapper[4789]: I1122 07:56:55.145113 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:55 crc kubenswrapper[4789]: I1122 07:56:55.158607 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:55 crc kubenswrapper[4789]: I1122 07:56:55.166303 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9f74f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"29ad5e00-4918-4718-9d49-013b12621f7d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mnz6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9f74f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:55 crc kubenswrapper[4789]: I1122 07:56:55.180354 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2wdp7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2fdee89f-559b-4f4e-8d20-6ecc3aafd22e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxvhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxvhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxvhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxvhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxvhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxvhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxvhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2wdp7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:55 crc kubenswrapper[4789]: I1122 07:56:55.195364 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d081b29e-edd7-4fb0-a2e5-431ee8491bd0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43d9e10254883b6491bef617cf9f0b319171bc1fbc49c7770ed9630a8a1ae2cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://223c2300bb6afb8812a72eb7605015b0473f59046565cdf39fb9b9f0cf6634bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca5af01b2e23b2bcac23ed767bc7dcd2ac0be1010ee5e675b3b44333ee337180\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8e79bfb6172c894868a8e1b47ac701ba4fdaca1fc6b1582f7e70807aadfdf1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:29Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:55 crc kubenswrapper[4789]: I1122 07:56:55.206170 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:55 crc kubenswrapper[4789]: I1122 07:56:55.214655 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58aa8071-7ecc-4692-8789-2db04ec70647\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfjgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfjgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pxvrm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:55 crc kubenswrapper[4789]: I1122 07:56:55.221901 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vrkrh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"371275b1-1d13-4f3c-8190-8c59c080fef5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:54Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:54Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8w4z2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vrkrh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:55 crc kubenswrapper[4789]: I1122 07:56:55.232290 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93be48ce-7d87-448b-b03e-ab6736cfc34c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba91466edcf49f73751622f98632b3040a046ccd8a42ba8ecee7ab08688a8c24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b6666e0ce42a46f669505659298a19931b1bf7c253ae6a97c6acfd80a9e717f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://251e72e1cd0f3e0ee195bacf33ab38765d1dd943adf4395f9d6dd282bda024cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://098c31b822e7c2351f1142af36b942d67ff8262896e28022c96a187d40bd0ddb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://098c31b822e7c2351f1142af36b942d67ff8262896e28022c96a187d40bd0ddb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"message\\\":\\\"le observer\\\\nW1122 07:56:50.641735 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1122 07:56:50.642792 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 07:56:50.643945 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3588433834/tls.crt::/tmp/serving-cert-3588433834/tls.key\\\\\\\"\\\\nI1122 07:56:50.949040 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 07:56:50.952068 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 07:56:50.952092 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 07:56:50.952116 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 07:56:50.952121 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 07:56:50.958549 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1122 07:56:50.958565 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1122 07:56:50.958579 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 07:56:50.958585 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 07:56:50.958591 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 07:56:50.958596 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1122 07:56:50.958600 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 07:56:50.958605 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1122 07:56:50.962189 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T07:56:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://263dce4c3f4f3fe388b21d4bb55a9fc7d956c3cca7da34ee00f470d88ef501f1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac2b567fcebc517d3fb4366aebce03351183ea03ce62f86c2abdf5b95e20b34a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac2b567fcebc517d3fb4366aebce03351183ea03ce62f86c2abdf5b95e20b34a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T07:56:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T07:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:55 crc kubenswrapper[4789]: I1122 07:56:55.242166 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:55 crc kubenswrapper[4789]: I1122 07:56:55.242207 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:55 crc kubenswrapper[4789]: I1122 07:56:55.242215 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:55 crc kubenswrapper[4789]: I1122 07:56:55.242230 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:56:55 crc kubenswrapper[4789]: I1122 07:56:55.242241 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:56:55Z","lastTransitionTime":"2025-11-22T07:56:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:56:55 crc kubenswrapper[4789]: I1122 07:56:55.242429 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:55 crc kubenswrapper[4789]: I1122 07:56:55.252420 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rks5j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5613c209-59a0-4b8a-ac8b-5c633bc1bdc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wj8sc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rks5j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:55 crc kubenswrapper[4789]: I1122 07:56:55.267717 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66f74155-c96e-4faa-876f-62b8b5e6b3bc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53953e73ae4a11a47b7a5332786b97d9f129d9576e66aa2664836ec5f3786683\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65813cc305e2105a598eb0b701dd079f1f26db8cc1b2e67679e7086c2ba3630f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1d47e3e0a1939863a1ae9e1f52262d8fad029cd90552c21e73a98cdf0a803ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9d09f0deafe8d22df512d2acd120ce10eec182173daaf5029ab72f848ceecd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c410501c102d72819d29aa0036b6270b1ce9804bba84092161b2c2d8ebaa885a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0755dd1f1e7b1fda23c8973b56521ab15b62b9eb76f9b5604110e04e2f25b422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0755dd1f1e7b1fda23c8973b56521ab15b62b9eb76f9b5604110e04e2f25b422\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T07:56:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T07:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://680b49efb88db68b75efc6ca47cdfa449c1cf688d277b3e1eaa75cb466e90480\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://680b49efb88db68b75efc6ca47cdfa449c1cf688d277b3e1eaa75cb466e90480\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T07:56:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T07:56:32Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c5b1cf28577e60931184499a22559f30b0bdf5dab0b6cb542c948fec5d9c54cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5b1cf28577e60931184499a22559f30b0bdf5dab0b6cb542c948fec5d9c54cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T07:56:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T07:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:29Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:55 crc kubenswrapper[4789]: I1122 07:56:55.279609 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:55 crc kubenswrapper[4789]: I1122 07:56:55.318653 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:55 crc kubenswrapper[4789]: I1122 07:56:55.344318 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:55 crc kubenswrapper[4789]: I1122 07:56:55.344374 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:55 crc kubenswrapper[4789]: I1122 07:56:55.344384 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:55 crc kubenswrapper[4789]: I1122 07:56:55.344403 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:56:55 crc kubenswrapper[4789]: I1122 07:56:55.344416 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:56:55Z","lastTransitionTime":"2025-11-22T07:56:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:56:55 crc kubenswrapper[4789]: I1122 07:56:55.368069 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0cb4cf51-62d6-4f32-a8b4-4116555f3870\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qmxss\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:55 crc kubenswrapper[4789]: I1122 07:56:55.447049 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:55 crc kubenswrapper[4789]: I1122 07:56:55.447095 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:55 crc kubenswrapper[4789]: I1122 07:56:55.447108 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:55 crc kubenswrapper[4789]: I1122 07:56:55.447122 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:56:55 crc kubenswrapper[4789]: I1122 07:56:55.447132 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:56:55Z","lastTransitionTime":"2025-11-22T07:56:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:56:55 crc kubenswrapper[4789]: I1122 07:56:55.549074 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:55 crc kubenswrapper[4789]: I1122 07:56:55.549116 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:55 crc kubenswrapper[4789]: I1122 07:56:55.549127 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:55 crc kubenswrapper[4789]: I1122 07:56:55.549141 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:56:55 crc kubenswrapper[4789]: I1122 07:56:55.549149 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:56:55Z","lastTransitionTime":"2025-11-22T07:56:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:56:55 crc kubenswrapper[4789]: I1122 07:56:55.652223 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:55 crc kubenswrapper[4789]: I1122 07:56:55.652333 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:55 crc kubenswrapper[4789]: I1122 07:56:55.652360 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:55 crc kubenswrapper[4789]: I1122 07:56:55.652391 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:56:55 crc kubenswrapper[4789]: I1122 07:56:55.652413 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:56:55Z","lastTransitionTime":"2025-11-22T07:56:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:56:55 crc kubenswrapper[4789]: I1122 07:56:55.755411 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:55 crc kubenswrapper[4789]: I1122 07:56:55.755452 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:55 crc kubenswrapper[4789]: I1122 07:56:55.755463 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:55 crc kubenswrapper[4789]: I1122 07:56:55.755480 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:56:55 crc kubenswrapper[4789]: I1122 07:56:55.755493 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:56:55Z","lastTransitionTime":"2025-11-22T07:56:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:56:55 crc kubenswrapper[4789]: I1122 07:56:55.857869 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:55 crc kubenswrapper[4789]: I1122 07:56:55.857945 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:55 crc kubenswrapper[4789]: I1122 07:56:55.857963 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:55 crc kubenswrapper[4789]: I1122 07:56:55.857991 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:56:55 crc kubenswrapper[4789]: I1122 07:56:55.858009 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:56:55Z","lastTransitionTime":"2025-11-22T07:56:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:56:55 crc kubenswrapper[4789]: I1122 07:56:55.961551 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:55 crc kubenswrapper[4789]: I1122 07:56:55.961600 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:55 crc kubenswrapper[4789]: I1122 07:56:55.961618 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:55 crc kubenswrapper[4789]: I1122 07:56:55.961636 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:56:55 crc kubenswrapper[4789]: I1122 07:56:55.961647 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:56:55Z","lastTransitionTime":"2025-11-22T07:56:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:56:55 crc kubenswrapper[4789]: I1122 07:56:55.964907 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 07:56:55 crc kubenswrapper[4789]: E1122 07:56:55.965064 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 07:56:56 crc kubenswrapper[4789]: I1122 07:56:56.064457 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:56 crc kubenswrapper[4789]: I1122 07:56:56.064534 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:56 crc kubenswrapper[4789]: I1122 07:56:56.064542 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:56 crc kubenswrapper[4789]: I1122 07:56:56.064555 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:56:56 crc kubenswrapper[4789]: I1122 07:56:56.064564 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:56:56Z","lastTransitionTime":"2025-11-22T07:56:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:56:56 crc kubenswrapper[4789]: I1122 07:56:56.166552 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:56 crc kubenswrapper[4789]: I1122 07:56:56.166593 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:56 crc kubenswrapper[4789]: I1122 07:56:56.166602 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:56 crc kubenswrapper[4789]: I1122 07:56:56.166615 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:56:56 crc kubenswrapper[4789]: I1122 07:56:56.166624 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:56:56Z","lastTransitionTime":"2025-11-22T07:56:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:56:56 crc kubenswrapper[4789]: I1122 07:56:56.269271 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:56 crc kubenswrapper[4789]: I1122 07:56:56.269311 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:56 crc kubenswrapper[4789]: I1122 07:56:56.269319 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:56 crc kubenswrapper[4789]: I1122 07:56:56.269335 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:56:56 crc kubenswrapper[4789]: I1122 07:56:56.269347 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:56:56Z","lastTransitionTime":"2025-11-22T07:56:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:56:56 crc kubenswrapper[4789]: I1122 07:56:56.371889 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:56 crc kubenswrapper[4789]: I1122 07:56:56.371935 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:56 crc kubenswrapper[4789]: I1122 07:56:56.371945 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:56 crc kubenswrapper[4789]: I1122 07:56:56.371961 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:56:56 crc kubenswrapper[4789]: I1122 07:56:56.371973 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:56:56Z","lastTransitionTime":"2025-11-22T07:56:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:56:56 crc kubenswrapper[4789]: I1122 07:56:56.474671 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:56 crc kubenswrapper[4789]: I1122 07:56:56.474785 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:56 crc kubenswrapper[4789]: I1122 07:56:56.474804 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:56 crc kubenswrapper[4789]: I1122 07:56:56.474827 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:56:56 crc kubenswrapper[4789]: I1122 07:56:56.474841 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:56:56Z","lastTransitionTime":"2025-11-22T07:56:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:56:56 crc kubenswrapper[4789]: I1122 07:56:56.577248 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:56 crc kubenswrapper[4789]: I1122 07:56:56.577284 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:56 crc kubenswrapper[4789]: I1122 07:56:56.577293 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:56 crc kubenswrapper[4789]: I1122 07:56:56.577307 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:56:56 crc kubenswrapper[4789]: I1122 07:56:56.577317 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:56:56Z","lastTransitionTime":"2025-11-22T07:56:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:56:56 crc kubenswrapper[4789]: I1122 07:56:56.679723 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:56 crc kubenswrapper[4789]: I1122 07:56:56.679787 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:56 crc kubenswrapper[4789]: I1122 07:56:56.679800 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:56 crc kubenswrapper[4789]: I1122 07:56:56.679814 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:56:56 crc kubenswrapper[4789]: I1122 07:56:56.679824 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:56:56Z","lastTransitionTime":"2025-11-22T07:56:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:56:56 crc kubenswrapper[4789]: I1122 07:56:56.782232 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:56 crc kubenswrapper[4789]: I1122 07:56:56.782285 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:56 crc kubenswrapper[4789]: I1122 07:56:56.782298 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:56 crc kubenswrapper[4789]: I1122 07:56:56.782315 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:56:56 crc kubenswrapper[4789]: I1122 07:56:56.782328 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:56:56Z","lastTransitionTime":"2025-11-22T07:56:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:56:56 crc kubenswrapper[4789]: I1122 07:56:56.885347 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:56 crc kubenswrapper[4789]: I1122 07:56:56.885412 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:56 crc kubenswrapper[4789]: I1122 07:56:56.885434 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:56 crc kubenswrapper[4789]: I1122 07:56:56.885471 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:56:56 crc kubenswrapper[4789]: I1122 07:56:56.885505 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:56:56Z","lastTransitionTime":"2025-11-22T07:56:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:56:56 crc kubenswrapper[4789]: I1122 07:56:56.943380 4789 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Nov 22 07:56:56 crc kubenswrapper[4789]: I1122 07:56:56.964657 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 07:56:56 crc kubenswrapper[4789]: I1122 07:56:56.964732 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 07:56:56 crc kubenswrapper[4789]: E1122 07:56:56.964838 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 07:56:56 crc kubenswrapper[4789]: E1122 07:56:56.965037 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 07:56:56 crc kubenswrapper[4789]: I1122 07:56:56.988747 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:56 crc kubenswrapper[4789]: I1122 07:56:56.988840 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:56 crc kubenswrapper[4789]: I1122 07:56:56.988858 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:56 crc kubenswrapper[4789]: I1122 07:56:56.988881 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:56:56 crc kubenswrapper[4789]: I1122 07:56:56.988901 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:56:56Z","lastTransitionTime":"2025-11-22T07:56:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:56:57 crc kubenswrapper[4789]: I1122 07:56:57.091728 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:57 crc kubenswrapper[4789]: I1122 07:56:57.091804 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:57 crc kubenswrapper[4789]: I1122 07:56:57.091814 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:57 crc kubenswrapper[4789]: I1122 07:56:57.091826 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:56:57 crc kubenswrapper[4789]: I1122 07:56:57.091834 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:56:57Z","lastTransitionTime":"2025-11-22T07:56:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:56:57 crc kubenswrapper[4789]: I1122 07:56:57.194506 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:57 crc kubenswrapper[4789]: I1122 07:56:57.194541 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:57 crc kubenswrapper[4789]: I1122 07:56:57.194548 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:57 crc kubenswrapper[4789]: I1122 07:56:57.194563 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:56:57 crc kubenswrapper[4789]: I1122 07:56:57.194574 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:56:57Z","lastTransitionTime":"2025-11-22T07:56:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:56:57 crc kubenswrapper[4789]: I1122 07:56:57.297720 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:57 crc kubenswrapper[4789]: I1122 07:56:57.297785 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:57 crc kubenswrapper[4789]: I1122 07:56:57.297800 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:57 crc kubenswrapper[4789]: I1122 07:56:57.297817 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:56:57 crc kubenswrapper[4789]: I1122 07:56:57.297830 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:56:57Z","lastTransitionTime":"2025-11-22T07:56:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:56:57 crc kubenswrapper[4789]: I1122 07:56:57.401154 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:57 crc kubenswrapper[4789]: I1122 07:56:57.401212 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:57 crc kubenswrapper[4789]: I1122 07:56:57.401233 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:57 crc kubenswrapper[4789]: I1122 07:56:57.401255 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:56:57 crc kubenswrapper[4789]: I1122 07:56:57.401269 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:56:57Z","lastTransitionTime":"2025-11-22T07:56:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:56:57 crc kubenswrapper[4789]: I1122 07:56:57.503941 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:57 crc kubenswrapper[4789]: I1122 07:56:57.504001 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:57 crc kubenswrapper[4789]: I1122 07:56:57.504010 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:57 crc kubenswrapper[4789]: I1122 07:56:57.504029 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:56:57 crc kubenswrapper[4789]: I1122 07:56:57.504040 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:56:57Z","lastTransitionTime":"2025-11-22T07:56:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:56:57 crc kubenswrapper[4789]: I1122 07:56:57.606236 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:57 crc kubenswrapper[4789]: I1122 07:56:57.606288 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:57 crc kubenswrapper[4789]: I1122 07:56:57.606299 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:57 crc kubenswrapper[4789]: I1122 07:56:57.606318 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:56:57 crc kubenswrapper[4789]: I1122 07:56:57.606331 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:56:57Z","lastTransitionTime":"2025-11-22T07:56:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:56:57 crc kubenswrapper[4789]: I1122 07:56:57.709909 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:57 crc kubenswrapper[4789]: I1122 07:56:57.709968 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:57 crc kubenswrapper[4789]: I1122 07:56:57.709978 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:57 crc kubenswrapper[4789]: I1122 07:56:57.710005 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:56:57 crc kubenswrapper[4789]: I1122 07:56:57.710016 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:56:57Z","lastTransitionTime":"2025-11-22T07:56:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:56:57 crc kubenswrapper[4789]: I1122 07:56:57.812537 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:57 crc kubenswrapper[4789]: I1122 07:56:57.812599 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:57 crc kubenswrapper[4789]: I1122 07:56:57.812609 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:57 crc kubenswrapper[4789]: I1122 07:56:57.812632 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:56:57 crc kubenswrapper[4789]: I1122 07:56:57.812644 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:56:57Z","lastTransitionTime":"2025-11-22T07:56:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:56:57 crc kubenswrapper[4789]: I1122 07:56:57.914826 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:57 crc kubenswrapper[4789]: I1122 07:56:57.914890 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:57 crc kubenswrapper[4789]: I1122 07:56:57.914928 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:57 crc kubenswrapper[4789]: I1122 07:56:57.914946 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:56:57 crc kubenswrapper[4789]: I1122 07:56:57.914960 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:56:57Z","lastTransitionTime":"2025-11-22T07:56:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:56:57 crc kubenswrapper[4789]: I1122 07:56:57.964918 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 07:56:57 crc kubenswrapper[4789]: E1122 07:56:57.965107 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 07:56:57 crc kubenswrapper[4789]: I1122 07:56:57.985092 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66f74155-c96e-4faa-876f-62b8b5e6b3bc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53953e73ae4a11a47b7a5332786b97d9f129d9576e66aa2664836ec5f3786683\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65813cc305e2105a598eb0b701dd079f1f26db8cc1b2e67679e7086c2ba3630f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1d47e3e0a1939863a1ae9e1f52262d8fad029cd90552c21e73a98cdf0a803ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9d09f0deafe8d22df512d2acd120ce10eec182173daaf5029ab72f848ceecd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c410501c102d72819d29aa0036b6270b1ce9804bba84092161b2c2d8ebaa885a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0755dd1f1e7b1fda23c8973b56521ab15b62b9eb76f9b5604110e04e2f25b422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0755dd1f1e7b1fda23c8973b56521ab15b62b9eb76f9b5604110e04e2f25b422\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T07:56:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T07:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://680b49efb88db68b75efc6ca47cdfa449c1cf688d277b3e1eaa75cb466e90480\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://680b49efb88db68b75efc6ca47cdfa449c1cf688d277b3e1eaa75cb466e90480\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T07:56:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T07:56:32Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c5b1cf28577e60931184499a22559f30b0bdf5dab0b6cb542c948fec5d9c54cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5b1cf28577e60931184499a22559f30b0bdf5dab0b6cb542c948fec5d9c54cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T07:56:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T07:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:29Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:58 crc kubenswrapper[4789]: I1122 07:56:58.002880 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:58 crc kubenswrapper[4789]: I1122 07:56:58.017537 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:58 crc kubenswrapper[4789]: I1122 07:56:58.017605 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:58 crc kubenswrapper[4789]: I1122 07:56:58.017629 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:58 crc kubenswrapper[4789]: I1122 07:56:58.017659 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:56:58 crc kubenswrapper[4789]: I1122 07:56:58.017682 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:56:58Z","lastTransitionTime":"2025-11-22T07:56:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:56:58 crc kubenswrapper[4789]: I1122 07:56:58.021787 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:58 crc kubenswrapper[4789]: I1122 07:56:58.039916 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0cb4cf51-62d6-4f32-a8b4-4116555f3870\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qmxss\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:58 crc kubenswrapper[4789]: I1122 07:56:58.052830 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:58 crc kubenswrapper[4789]: I1122 07:56:58.062349 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9f74f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"29ad5e00-4918-4718-9d49-013b12621f7d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mnz6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9f74f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:58 crc kubenswrapper[4789]: I1122 07:56:58.076362 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2wdp7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2fdee89f-559b-4f4e-8d20-6ecc3aafd22e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxvhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxvhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxvhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxvhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxvhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxvhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxvhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2wdp7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:58 crc kubenswrapper[4789]: I1122 07:56:58.088341 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:58 crc kubenswrapper[4789]: I1122 07:56:58.100396 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:58 crc kubenswrapper[4789]: I1122 07:56:58.112201 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58aa8071-7ecc-4692-8789-2db04ec70647\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfjgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfjgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pxvrm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:58 crc kubenswrapper[4789]: I1122 07:56:58.120237 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:58 crc kubenswrapper[4789]: I1122 07:56:58.120282 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:58 crc kubenswrapper[4789]: I1122 07:56:58.120292 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:58 crc kubenswrapper[4789]: I1122 07:56:58.120307 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:56:58 crc kubenswrapper[4789]: I1122 07:56:58.120319 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:56:58Z","lastTransitionTime":"2025-11-22T07:56:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:56:58 crc kubenswrapper[4789]: I1122 07:56:58.124449 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vrkrh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"371275b1-1d13-4f3c-8190-8c59c080fef5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:54Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:54Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8w4z2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vrkrh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:58 crc kubenswrapper[4789]: I1122 07:56:58.136665 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d081b29e-edd7-4fb0-a2e5-431ee8491bd0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43d9e10254883b6491bef617cf9f0b319171bc1fbc49c7770ed9630a8a1ae2cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://223c2300bb6afb8812a72eb7605015b0473f59046565cdf39fb9b9f0cf6634bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca5af01b2e23b2bcac23ed767bc7dcd2ac0be1010ee5e675b3b44333ee337180\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8e79bfb6172c894868a8e1b47ac701ba4fdaca1fc6b1582f7e70807aadfdf1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:29Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:58 crc kubenswrapper[4789]: I1122 07:56:58.153719 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:58 crc kubenswrapper[4789]: I1122 07:56:58.164825 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rks5j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5613c209-59a0-4b8a-ac8b-5c633bc1bdc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wj8sc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rks5j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:58 crc kubenswrapper[4789]: I1122 07:56:58.176416 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93be48ce-7d87-448b-b03e-ab6736cfc34c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba91466edcf49f73751622f98632b3040a046ccd8a42ba8ecee7ab08688a8c24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b6666e0ce42a46f669505659298a19931b1bf7c253ae6a97c6acfd80a9e717f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://251e72e1cd0f3e0ee195bacf33ab38765d1dd943adf4395f9d6dd282bda024cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://098c31b822e7c2351f1142af36b942d67ff8262896e28022c96a187d40bd0ddb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://098c31b822e7c2351f1142af36b942d67ff8262896e28022c96a187d40bd0ddb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"message\\\":\\\"le observer\\\\nW1122 07:56:50.641735 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1122 07:56:50.642792 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 07:56:50.643945 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3588433834/tls.crt::/tmp/serving-cert-3588433834/tls.key\\\\\\\"\\\\nI1122 07:56:50.949040 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 07:56:50.952068 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 07:56:50.952092 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 07:56:50.952116 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 07:56:50.952121 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 07:56:50.958549 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1122 07:56:50.958565 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1122 07:56:50.958579 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 07:56:50.958585 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 07:56:50.958591 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 07:56:50.958596 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1122 07:56:50.958600 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 07:56:50.958605 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1122 07:56:50.962189 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T07:56:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://263dce4c3f4f3fe388b21d4bb55a9fc7d956c3cca7da34ee00f470d88ef501f1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac2b567fcebc517d3fb4366aebce03351183ea03ce62f86c2abdf5b95e20b34a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac2b567fcebc517d3fb4366aebce03351183ea03ce62f86c2abdf5b95e20b34a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T07:56:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T07:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:56:58 crc kubenswrapper[4789]: I1122 07:56:58.223092 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:58 crc kubenswrapper[4789]: I1122 07:56:58.223129 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:58 crc kubenswrapper[4789]: I1122 07:56:58.223140 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:58 crc kubenswrapper[4789]: I1122 07:56:58.223157 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:56:58 crc kubenswrapper[4789]: I1122 07:56:58.223170 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:56:58Z","lastTransitionTime":"2025-11-22T07:56:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:56:58 crc kubenswrapper[4789]: I1122 07:56:58.325857 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:58 crc kubenswrapper[4789]: I1122 07:56:58.325920 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:58 crc kubenswrapper[4789]: I1122 07:56:58.325935 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:58 crc kubenswrapper[4789]: I1122 07:56:58.325962 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:56:58 crc kubenswrapper[4789]: I1122 07:56:58.325977 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:56:58Z","lastTransitionTime":"2025-11-22T07:56:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:56:58 crc kubenswrapper[4789]: I1122 07:56:58.363710 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 07:56:58 crc kubenswrapper[4789]: E1122 07:56:58.364173 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 07:57:06.364127529 +0000 UTC m=+40.598527802 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:56:58 crc kubenswrapper[4789]: I1122 07:56:58.429463 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:58 crc kubenswrapper[4789]: I1122 07:56:58.429521 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:58 crc kubenswrapper[4789]: I1122 07:56:58.429531 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:58 crc kubenswrapper[4789]: I1122 07:56:58.429552 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:56:58 crc kubenswrapper[4789]: I1122 07:56:58.429565 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:56:58Z","lastTransitionTime":"2025-11-22T07:56:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:56:58 crc kubenswrapper[4789]: I1122 07:56:58.465166 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 07:56:58 crc kubenswrapper[4789]: I1122 07:56:58.465230 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 07:56:58 crc kubenswrapper[4789]: I1122 07:56:58.465252 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 07:56:58 crc kubenswrapper[4789]: I1122 07:56:58.465274 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 07:56:58 crc kubenswrapper[4789]: E1122 07:56:58.465434 4789 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 22 07:56:58 crc kubenswrapper[4789]: E1122 07:56:58.465522 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-22 07:57:06.465500143 +0000 UTC m=+40.699900416 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 22 07:56:58 crc kubenswrapper[4789]: E1122 07:56:58.465548 4789 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 22 07:56:58 crc kubenswrapper[4789]: E1122 07:56:58.465588 4789 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 22 07:56:58 crc kubenswrapper[4789]: E1122 07:56:58.465607 4789 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 07:56:58 crc kubenswrapper[4789]: E1122 07:56:58.465655 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-22 07:57:06.465643967 +0000 UTC m=+40.700044240 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 07:56:58 crc kubenswrapper[4789]: E1122 07:56:58.465706 4789 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 22 07:56:58 crc kubenswrapper[4789]: E1122 07:56:58.465888 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-22 07:57:06.465865943 +0000 UTC m=+40.700266216 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 22 07:56:58 crc kubenswrapper[4789]: E1122 07:56:58.465726 4789 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 22 07:56:58 crc kubenswrapper[4789]: E1122 07:56:58.465939 4789 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 22 07:56:58 crc kubenswrapper[4789]: E1122 07:56:58.465955 4789 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 07:56:58 crc kubenswrapper[4789]: E1122 07:56:58.466015 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-22 07:57:06.466003987 +0000 UTC m=+40.700404440 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 07:56:58 crc kubenswrapper[4789]: I1122 07:56:58.531180 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:58 crc kubenswrapper[4789]: I1122 07:56:58.531224 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:58 crc kubenswrapper[4789]: I1122 07:56:58.531246 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:58 crc kubenswrapper[4789]: I1122 07:56:58.531260 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:56:58 crc kubenswrapper[4789]: I1122 07:56:58.531277 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:56:58Z","lastTransitionTime":"2025-11-22T07:56:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:56:58 crc kubenswrapper[4789]: I1122 07:56:58.634405 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:58 crc kubenswrapper[4789]: I1122 07:56:58.634459 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:58 crc kubenswrapper[4789]: I1122 07:56:58.634472 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:58 crc kubenswrapper[4789]: I1122 07:56:58.634491 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:56:58 crc kubenswrapper[4789]: I1122 07:56:58.634503 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:56:58Z","lastTransitionTime":"2025-11-22T07:56:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:56:58 crc kubenswrapper[4789]: I1122 07:56:58.736932 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:58 crc kubenswrapper[4789]: I1122 07:56:58.736971 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:58 crc kubenswrapper[4789]: I1122 07:56:58.736984 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:58 crc kubenswrapper[4789]: I1122 07:56:58.736998 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:56:58 crc kubenswrapper[4789]: I1122 07:56:58.737007 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:56:58Z","lastTransitionTime":"2025-11-22T07:56:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:56:58 crc kubenswrapper[4789]: I1122 07:56:58.838942 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:58 crc kubenswrapper[4789]: I1122 07:56:58.838983 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:58 crc kubenswrapper[4789]: I1122 07:56:58.838993 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:58 crc kubenswrapper[4789]: I1122 07:56:58.839011 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:56:58 crc kubenswrapper[4789]: I1122 07:56:58.839021 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:56:58Z","lastTransitionTime":"2025-11-22T07:56:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:56:58 crc kubenswrapper[4789]: I1122 07:56:58.941713 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:58 crc kubenswrapper[4789]: I1122 07:56:58.941803 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:58 crc kubenswrapper[4789]: I1122 07:56:58.941813 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:58 crc kubenswrapper[4789]: I1122 07:56:58.941828 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:56:58 crc kubenswrapper[4789]: I1122 07:56:58.941839 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:56:58Z","lastTransitionTime":"2025-11-22T07:56:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:56:58 crc kubenswrapper[4789]: I1122 07:56:58.965031 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 07:56:58 crc kubenswrapper[4789]: I1122 07:56:58.965059 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 07:56:58 crc kubenswrapper[4789]: E1122 07:56:58.965159 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 07:56:58 crc kubenswrapper[4789]: E1122 07:56:58.965258 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 07:56:59 crc kubenswrapper[4789]: I1122 07:56:59.043847 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:59 crc kubenswrapper[4789]: I1122 07:56:59.043894 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:59 crc kubenswrapper[4789]: I1122 07:56:59.043912 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:59 crc kubenswrapper[4789]: I1122 07:56:59.043933 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:56:59 crc kubenswrapper[4789]: I1122 07:56:59.043949 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:56:59Z","lastTransitionTime":"2025-11-22T07:56:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:56:59 crc kubenswrapper[4789]: I1122 07:56:59.146363 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:59 crc kubenswrapper[4789]: I1122 07:56:59.146409 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:59 crc kubenswrapper[4789]: I1122 07:56:59.146421 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:59 crc kubenswrapper[4789]: I1122 07:56:59.146439 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:56:59 crc kubenswrapper[4789]: I1122 07:56:59.146452 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:56:59Z","lastTransitionTime":"2025-11-22T07:56:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:56:59 crc kubenswrapper[4789]: I1122 07:56:59.248766 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:59 crc kubenswrapper[4789]: I1122 07:56:59.248818 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:59 crc kubenswrapper[4789]: I1122 07:56:59.248831 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:59 crc kubenswrapper[4789]: I1122 07:56:59.248885 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:56:59 crc kubenswrapper[4789]: I1122 07:56:59.248899 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:56:59Z","lastTransitionTime":"2025-11-22T07:56:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:56:59 crc kubenswrapper[4789]: I1122 07:56:59.352287 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:59 crc kubenswrapper[4789]: I1122 07:56:59.352341 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:59 crc kubenswrapper[4789]: I1122 07:56:59.352356 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:59 crc kubenswrapper[4789]: I1122 07:56:59.352372 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:56:59 crc kubenswrapper[4789]: I1122 07:56:59.352381 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:56:59Z","lastTransitionTime":"2025-11-22T07:56:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:56:59 crc kubenswrapper[4789]: I1122 07:56:59.455354 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:59 crc kubenswrapper[4789]: I1122 07:56:59.455413 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:59 crc kubenswrapper[4789]: I1122 07:56:59.455426 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:59 crc kubenswrapper[4789]: I1122 07:56:59.455445 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:56:59 crc kubenswrapper[4789]: I1122 07:56:59.455456 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:56:59Z","lastTransitionTime":"2025-11-22T07:56:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:56:59 crc kubenswrapper[4789]: I1122 07:56:59.558398 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:59 crc kubenswrapper[4789]: I1122 07:56:59.558449 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:59 crc kubenswrapper[4789]: I1122 07:56:59.558458 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:59 crc kubenswrapper[4789]: I1122 07:56:59.558477 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:56:59 crc kubenswrapper[4789]: I1122 07:56:59.558489 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:56:59Z","lastTransitionTime":"2025-11-22T07:56:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:56:59 crc kubenswrapper[4789]: I1122 07:56:59.661363 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:59 crc kubenswrapper[4789]: I1122 07:56:59.661450 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:59 crc kubenswrapper[4789]: I1122 07:56:59.661460 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:59 crc kubenswrapper[4789]: I1122 07:56:59.661484 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:56:59 crc kubenswrapper[4789]: I1122 07:56:59.661499 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:56:59Z","lastTransitionTime":"2025-11-22T07:56:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:56:59 crc kubenswrapper[4789]: I1122 07:56:59.763460 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:59 crc kubenswrapper[4789]: I1122 07:56:59.764060 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:59 crc kubenswrapper[4789]: I1122 07:56:59.764163 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:59 crc kubenswrapper[4789]: I1122 07:56:59.764272 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:56:59 crc kubenswrapper[4789]: I1122 07:56:59.764364 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:56:59Z","lastTransitionTime":"2025-11-22T07:56:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:56:59 crc kubenswrapper[4789]: I1122 07:56:59.867681 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:59 crc kubenswrapper[4789]: I1122 07:56:59.867782 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:59 crc kubenswrapper[4789]: I1122 07:56:59.867798 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:59 crc kubenswrapper[4789]: I1122 07:56:59.867823 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:56:59 crc kubenswrapper[4789]: I1122 07:56:59.867844 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:56:59Z","lastTransitionTime":"2025-11-22T07:56:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:56:59 crc kubenswrapper[4789]: I1122 07:56:59.965099 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 07:56:59 crc kubenswrapper[4789]: E1122 07:56:59.965296 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 07:56:59 crc kubenswrapper[4789]: I1122 07:56:59.970931 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:56:59 crc kubenswrapper[4789]: I1122 07:56:59.970971 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:56:59 crc kubenswrapper[4789]: I1122 07:56:59.970984 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:56:59 crc kubenswrapper[4789]: I1122 07:56:59.971002 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:56:59 crc kubenswrapper[4789]: I1122 07:56:59.971019 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:56:59Z","lastTransitionTime":"2025-11-22T07:56:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:00 crc kubenswrapper[4789]: I1122 07:57:00.074906 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:00 crc kubenswrapper[4789]: I1122 07:57:00.074947 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:00 crc kubenswrapper[4789]: I1122 07:57:00.074958 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:00 crc kubenswrapper[4789]: I1122 07:57:00.074973 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:00 crc kubenswrapper[4789]: I1122 07:57:00.074984 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:00Z","lastTransitionTime":"2025-11-22T07:57:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:00 crc kubenswrapper[4789]: I1122 07:57:00.177385 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:00 crc kubenswrapper[4789]: I1122 07:57:00.177426 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:00 crc kubenswrapper[4789]: I1122 07:57:00.177438 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:00 crc kubenswrapper[4789]: I1122 07:57:00.177453 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:00 crc kubenswrapper[4789]: I1122 07:57:00.177463 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:00Z","lastTransitionTime":"2025-11-22T07:57:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:00 crc kubenswrapper[4789]: I1122 07:57:00.279965 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:00 crc kubenswrapper[4789]: I1122 07:57:00.280032 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:00 crc kubenswrapper[4789]: I1122 07:57:00.280043 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:00 crc kubenswrapper[4789]: I1122 07:57:00.280060 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:00 crc kubenswrapper[4789]: I1122 07:57:00.280071 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:00Z","lastTransitionTime":"2025-11-22T07:57:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:00 crc kubenswrapper[4789]: I1122 07:57:00.382670 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:00 crc kubenswrapper[4789]: I1122 07:57:00.382716 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:00 crc kubenswrapper[4789]: I1122 07:57:00.382726 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:00 crc kubenswrapper[4789]: I1122 07:57:00.382740 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:00 crc kubenswrapper[4789]: I1122 07:57:00.382766 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:00Z","lastTransitionTime":"2025-11-22T07:57:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:00 crc kubenswrapper[4789]: I1122 07:57:00.484733 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:00 crc kubenswrapper[4789]: I1122 07:57:00.484803 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:00 crc kubenswrapper[4789]: I1122 07:57:00.484813 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:00 crc kubenswrapper[4789]: I1122 07:57:00.484829 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:00 crc kubenswrapper[4789]: I1122 07:57:00.484839 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:00Z","lastTransitionTime":"2025-11-22T07:57:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:00 crc kubenswrapper[4789]: I1122 07:57:00.587831 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:00 crc kubenswrapper[4789]: I1122 07:57:00.587956 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:00 crc kubenswrapper[4789]: I1122 07:57:00.587988 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:00 crc kubenswrapper[4789]: I1122 07:57:00.588019 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:00 crc kubenswrapper[4789]: I1122 07:57:00.588040 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:00Z","lastTransitionTime":"2025-11-22T07:57:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:00 crc kubenswrapper[4789]: I1122 07:57:00.690573 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:00 crc kubenswrapper[4789]: I1122 07:57:00.690634 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:00 crc kubenswrapper[4789]: I1122 07:57:00.690654 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:00 crc kubenswrapper[4789]: I1122 07:57:00.690674 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:00 crc kubenswrapper[4789]: I1122 07:57:00.690688 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:00Z","lastTransitionTime":"2025-11-22T07:57:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:00 crc kubenswrapper[4789]: I1122 07:57:00.793632 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:00 crc kubenswrapper[4789]: I1122 07:57:00.793665 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:00 crc kubenswrapper[4789]: I1122 07:57:00.793674 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:00 crc kubenswrapper[4789]: I1122 07:57:00.793686 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:00 crc kubenswrapper[4789]: I1122 07:57:00.793694 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:00Z","lastTransitionTime":"2025-11-22T07:57:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:00 crc kubenswrapper[4789]: I1122 07:57:00.895482 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:00 crc kubenswrapper[4789]: I1122 07:57:00.895531 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:00 crc kubenswrapper[4789]: I1122 07:57:00.895542 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:00 crc kubenswrapper[4789]: I1122 07:57:00.895557 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:00 crc kubenswrapper[4789]: I1122 07:57:00.895567 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:00Z","lastTransitionTime":"2025-11-22T07:57:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:00 crc kubenswrapper[4789]: I1122 07:57:00.964149 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 07:57:00 crc kubenswrapper[4789]: E1122 07:57:00.964309 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 07:57:00 crc kubenswrapper[4789]: I1122 07:57:00.964170 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 07:57:00 crc kubenswrapper[4789]: E1122 07:57:00.964726 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 07:57:00 crc kubenswrapper[4789]: I1122 07:57:00.998224 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:00 crc kubenswrapper[4789]: I1122 07:57:00.998280 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:00 crc kubenswrapper[4789]: I1122 07:57:00.998292 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:00 crc kubenswrapper[4789]: I1122 07:57:00.998314 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:00 crc kubenswrapper[4789]: I1122 07:57:00.998329 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:00Z","lastTransitionTime":"2025-11-22T07:57:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:01 crc kubenswrapper[4789]: I1122 07:57:01.101083 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:01 crc kubenswrapper[4789]: I1122 07:57:01.101151 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:01 crc kubenswrapper[4789]: I1122 07:57:01.101174 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:01 crc kubenswrapper[4789]: I1122 07:57:01.101202 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:01 crc kubenswrapper[4789]: I1122 07:57:01.101225 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:01Z","lastTransitionTime":"2025-11-22T07:57:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:01 crc kubenswrapper[4789]: I1122 07:57:01.204584 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:01 crc kubenswrapper[4789]: I1122 07:57:01.204652 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:01 crc kubenswrapper[4789]: I1122 07:57:01.204670 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:01 crc kubenswrapper[4789]: I1122 07:57:01.204693 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:01 crc kubenswrapper[4789]: I1122 07:57:01.204712 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:01Z","lastTransitionTime":"2025-11-22T07:57:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:01 crc kubenswrapper[4789]: I1122 07:57:01.306980 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:01 crc kubenswrapper[4789]: I1122 07:57:01.307027 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:01 crc kubenswrapper[4789]: I1122 07:57:01.307036 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:01 crc kubenswrapper[4789]: I1122 07:57:01.307057 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:01 crc kubenswrapper[4789]: I1122 07:57:01.307067 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:01Z","lastTransitionTime":"2025-11-22T07:57:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:01 crc kubenswrapper[4789]: I1122 07:57:01.410368 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:01 crc kubenswrapper[4789]: I1122 07:57:01.410424 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:01 crc kubenswrapper[4789]: I1122 07:57:01.410442 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:01 crc kubenswrapper[4789]: I1122 07:57:01.410463 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:01 crc kubenswrapper[4789]: I1122 07:57:01.410478 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:01Z","lastTransitionTime":"2025-11-22T07:57:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:01 crc kubenswrapper[4789]: I1122 07:57:01.513708 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:01 crc kubenswrapper[4789]: I1122 07:57:01.513781 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:01 crc kubenswrapper[4789]: I1122 07:57:01.513797 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:01 crc kubenswrapper[4789]: I1122 07:57:01.513820 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:01 crc kubenswrapper[4789]: I1122 07:57:01.513836 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:01Z","lastTransitionTime":"2025-11-22T07:57:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:01 crc kubenswrapper[4789]: I1122 07:57:01.617373 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:01 crc kubenswrapper[4789]: I1122 07:57:01.617450 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:01 crc kubenswrapper[4789]: I1122 07:57:01.617497 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:01 crc kubenswrapper[4789]: I1122 07:57:01.617535 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:01 crc kubenswrapper[4789]: I1122 07:57:01.617547 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:01Z","lastTransitionTime":"2025-11-22T07:57:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:01 crc kubenswrapper[4789]: I1122 07:57:01.720158 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:01 crc kubenswrapper[4789]: I1122 07:57:01.720246 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:01 crc kubenswrapper[4789]: I1122 07:57:01.720280 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:01 crc kubenswrapper[4789]: I1122 07:57:01.720295 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:01 crc kubenswrapper[4789]: I1122 07:57:01.720322 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:01Z","lastTransitionTime":"2025-11-22T07:57:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:01 crc kubenswrapper[4789]: I1122 07:57:01.798152 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:01 crc kubenswrapper[4789]: I1122 07:57:01.798203 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:01 crc kubenswrapper[4789]: I1122 07:57:01.798214 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:01 crc kubenswrapper[4789]: I1122 07:57:01.798231 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:01 crc kubenswrapper[4789]: I1122 07:57:01.798240 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:01Z","lastTransitionTime":"2025-11-22T07:57:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:01 crc kubenswrapper[4789]: E1122 07:57:01.816486 4789 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T07:57:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T07:57:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T07:57:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T07:57:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T07:57:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T07:57:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T07:57:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T07:57:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"06b9c1b2-d063-4af7-880e-0654164bb896\\\",\\\"systemUUID\\\":\\\"fd8856dd-ccc1-49f3-a39b-e4a6f9389fc3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:01 crc kubenswrapper[4789]: I1122 07:57:01.820662 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:01 crc kubenswrapper[4789]: I1122 07:57:01.820974 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:01 crc kubenswrapper[4789]: I1122 07:57:01.821081 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:01 crc kubenswrapper[4789]: I1122 07:57:01.821118 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:01 crc kubenswrapper[4789]: I1122 07:57:01.821141 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:01Z","lastTransitionTime":"2025-11-22T07:57:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:01 crc kubenswrapper[4789]: E1122 07:57:01.833528 4789 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T07:57:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T07:57:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T07:57:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T07:57:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T07:57:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T07:57:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T07:57:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T07:57:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"06b9c1b2-d063-4af7-880e-0654164bb896\\\",\\\"systemUUID\\\":\\\"fd8856dd-ccc1-49f3-a39b-e4a6f9389fc3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:01 crc kubenswrapper[4789]: I1122 07:57:01.837194 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:01 crc kubenswrapper[4789]: I1122 07:57:01.837244 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:01 crc kubenswrapper[4789]: I1122 07:57:01.837257 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:01 crc kubenswrapper[4789]: I1122 07:57:01.837276 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:01 crc kubenswrapper[4789]: I1122 07:57:01.837287 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:01Z","lastTransitionTime":"2025-11-22T07:57:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:01 crc kubenswrapper[4789]: E1122 07:57:01.846657 4789 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T07:57:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T07:57:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T07:57:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T07:57:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T07:57:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T07:57:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T07:57:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T07:57:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"06b9c1b2-d063-4af7-880e-0654164bb896\\\",\\\"systemUUID\\\":\\\"fd8856dd-ccc1-49f3-a39b-e4a6f9389fc3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:01 crc kubenswrapper[4789]: I1122 07:57:01.851286 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:01 crc kubenswrapper[4789]: I1122 07:57:01.851335 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:01 crc kubenswrapper[4789]: I1122 07:57:01.851348 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:01 crc kubenswrapper[4789]: I1122 07:57:01.851370 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:01 crc kubenswrapper[4789]: I1122 07:57:01.851385 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:01Z","lastTransitionTime":"2025-11-22T07:57:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:01 crc kubenswrapper[4789]: E1122 07:57:01.865002 4789 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T07:57:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T07:57:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T07:57:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T07:57:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T07:57:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T07:57:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T07:57:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T07:57:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"06b9c1b2-d063-4af7-880e-0654164bb896\\\",\\\"systemUUID\\\":\\\"fd8856dd-ccc1-49f3-a39b-e4a6f9389fc3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:01 crc kubenswrapper[4789]: I1122 07:57:01.869139 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:01 crc kubenswrapper[4789]: I1122 07:57:01.869180 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:01 crc kubenswrapper[4789]: I1122 07:57:01.869190 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:01 crc kubenswrapper[4789]: I1122 07:57:01.869204 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:01 crc kubenswrapper[4789]: I1122 07:57:01.869216 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:01Z","lastTransitionTime":"2025-11-22T07:57:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:01 crc kubenswrapper[4789]: E1122 07:57:01.880005 4789 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T07:57:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T07:57:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T07:57:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T07:57:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T07:57:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T07:57:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T07:57:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T07:57:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"06b9c1b2-d063-4af7-880e-0654164bb896\\\",\\\"systemUUID\\\":\\\"fd8856dd-ccc1-49f3-a39b-e4a6f9389fc3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:01 crc kubenswrapper[4789]: E1122 07:57:01.880231 4789 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 22 07:57:01 crc kubenswrapper[4789]: I1122 07:57:01.881952 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:01 crc kubenswrapper[4789]: I1122 07:57:01.881996 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:01 crc kubenswrapper[4789]: I1122 07:57:01.882038 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:01 crc kubenswrapper[4789]: I1122 07:57:01.882056 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:01 crc kubenswrapper[4789]: I1122 07:57:01.882072 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:01Z","lastTransitionTime":"2025-11-22T07:57:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:01 crc kubenswrapper[4789]: I1122 07:57:01.964615 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 07:57:01 crc kubenswrapper[4789]: E1122 07:57:01.964878 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 07:57:01 crc kubenswrapper[4789]: I1122 07:57:01.984691 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:01 crc kubenswrapper[4789]: I1122 07:57:01.984777 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:01 crc kubenswrapper[4789]: I1122 07:57:01.984789 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:01 crc kubenswrapper[4789]: I1122 07:57:01.984802 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:01 crc kubenswrapper[4789]: I1122 07:57:01.984813 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:01Z","lastTransitionTime":"2025-11-22T07:57:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:02 crc kubenswrapper[4789]: I1122 07:57:02.087276 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:02 crc kubenswrapper[4789]: I1122 07:57:02.087323 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:02 crc kubenswrapper[4789]: I1122 07:57:02.087332 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:02 crc kubenswrapper[4789]: I1122 07:57:02.087347 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:02 crc kubenswrapper[4789]: I1122 07:57:02.087357 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:02Z","lastTransitionTime":"2025-11-22T07:57:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:02 crc kubenswrapper[4789]: I1122 07:57:02.190011 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:02 crc kubenswrapper[4789]: I1122 07:57:02.190048 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:02 crc kubenswrapper[4789]: I1122 07:57:02.190058 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:02 crc kubenswrapper[4789]: I1122 07:57:02.190072 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:02 crc kubenswrapper[4789]: I1122 07:57:02.190081 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:02Z","lastTransitionTime":"2025-11-22T07:57:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:02 crc kubenswrapper[4789]: I1122 07:57:02.293358 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:02 crc kubenswrapper[4789]: I1122 07:57:02.293398 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:02 crc kubenswrapper[4789]: I1122 07:57:02.293408 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:02 crc kubenswrapper[4789]: I1122 07:57:02.293428 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:02 crc kubenswrapper[4789]: I1122 07:57:02.293440 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:02Z","lastTransitionTime":"2025-11-22T07:57:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:02 crc kubenswrapper[4789]: I1122 07:57:02.395593 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:02 crc kubenswrapper[4789]: I1122 07:57:02.395653 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:02 crc kubenswrapper[4789]: I1122 07:57:02.395666 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:02 crc kubenswrapper[4789]: I1122 07:57:02.395690 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:02 crc kubenswrapper[4789]: I1122 07:57:02.395706 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:02Z","lastTransitionTime":"2025-11-22T07:57:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:02 crc kubenswrapper[4789]: I1122 07:57:02.498938 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:02 crc kubenswrapper[4789]: I1122 07:57:02.498993 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:02 crc kubenswrapper[4789]: I1122 07:57:02.499005 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:02 crc kubenswrapper[4789]: I1122 07:57:02.499025 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:02 crc kubenswrapper[4789]: I1122 07:57:02.499039 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:02Z","lastTransitionTime":"2025-11-22T07:57:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:02 crc kubenswrapper[4789]: I1122 07:57:02.601487 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:02 crc kubenswrapper[4789]: I1122 07:57:02.601533 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:02 crc kubenswrapper[4789]: I1122 07:57:02.601542 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:02 crc kubenswrapper[4789]: I1122 07:57:02.601558 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:02 crc kubenswrapper[4789]: I1122 07:57:02.601569 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:02Z","lastTransitionTime":"2025-11-22T07:57:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:02 crc kubenswrapper[4789]: I1122 07:57:02.704392 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:02 crc kubenswrapper[4789]: I1122 07:57:02.704434 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:02 crc kubenswrapper[4789]: I1122 07:57:02.704443 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:02 crc kubenswrapper[4789]: I1122 07:57:02.704460 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:02 crc kubenswrapper[4789]: I1122 07:57:02.704468 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:02Z","lastTransitionTime":"2025-11-22T07:57:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:02 crc kubenswrapper[4789]: I1122 07:57:02.807307 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:02 crc kubenswrapper[4789]: I1122 07:57:02.807348 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:02 crc kubenswrapper[4789]: I1122 07:57:02.807361 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:02 crc kubenswrapper[4789]: I1122 07:57:02.807376 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:02 crc kubenswrapper[4789]: I1122 07:57:02.807386 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:02Z","lastTransitionTime":"2025-11-22T07:57:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:02 crc kubenswrapper[4789]: I1122 07:57:02.909192 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:02 crc kubenswrapper[4789]: I1122 07:57:02.909248 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:02 crc kubenswrapper[4789]: I1122 07:57:02.909256 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:02 crc kubenswrapper[4789]: I1122 07:57:02.909268 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:02 crc kubenswrapper[4789]: I1122 07:57:02.909276 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:02Z","lastTransitionTime":"2025-11-22T07:57:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:02 crc kubenswrapper[4789]: I1122 07:57:02.965166 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 07:57:02 crc kubenswrapper[4789]: I1122 07:57:02.965191 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 07:57:02 crc kubenswrapper[4789]: E1122 07:57:02.965326 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 07:57:02 crc kubenswrapper[4789]: E1122 07:57:02.965418 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 07:57:03 crc kubenswrapper[4789]: I1122 07:57:03.011579 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:03 crc kubenswrapper[4789]: I1122 07:57:03.011627 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:03 crc kubenswrapper[4789]: I1122 07:57:03.011637 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:03 crc kubenswrapper[4789]: I1122 07:57:03.011655 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:03 crc kubenswrapper[4789]: I1122 07:57:03.011670 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:03Z","lastTransitionTime":"2025-11-22T07:57:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:03 crc kubenswrapper[4789]: I1122 07:57:03.114783 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:03 crc kubenswrapper[4789]: I1122 07:57:03.114856 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:03 crc kubenswrapper[4789]: I1122 07:57:03.114874 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:03 crc kubenswrapper[4789]: I1122 07:57:03.114898 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:03 crc kubenswrapper[4789]: I1122 07:57:03.114911 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:03Z","lastTransitionTime":"2025-11-22T07:57:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:03 crc kubenswrapper[4789]: I1122 07:57:03.218169 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:03 crc kubenswrapper[4789]: I1122 07:57:03.218204 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:03 crc kubenswrapper[4789]: I1122 07:57:03.218216 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:03 crc kubenswrapper[4789]: I1122 07:57:03.218230 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:03 crc kubenswrapper[4789]: I1122 07:57:03.218241 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:03Z","lastTransitionTime":"2025-11-22T07:57:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:03 crc kubenswrapper[4789]: I1122 07:57:03.321294 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:03 crc kubenswrapper[4789]: I1122 07:57:03.321437 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:03 crc kubenswrapper[4789]: I1122 07:57:03.321458 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:03 crc kubenswrapper[4789]: I1122 07:57:03.321958 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:03 crc kubenswrapper[4789]: I1122 07:57:03.322228 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:03Z","lastTransitionTime":"2025-11-22T07:57:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:03 crc kubenswrapper[4789]: I1122 07:57:03.425218 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:03 crc kubenswrapper[4789]: I1122 07:57:03.425259 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:03 crc kubenswrapper[4789]: I1122 07:57:03.425269 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:03 crc kubenswrapper[4789]: I1122 07:57:03.425305 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:03 crc kubenswrapper[4789]: I1122 07:57:03.425316 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:03Z","lastTransitionTime":"2025-11-22T07:57:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:03 crc kubenswrapper[4789]: I1122 07:57:03.527687 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:03 crc kubenswrapper[4789]: I1122 07:57:03.527735 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:03 crc kubenswrapper[4789]: I1122 07:57:03.527764 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:03 crc kubenswrapper[4789]: I1122 07:57:03.527779 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:03 crc kubenswrapper[4789]: I1122 07:57:03.527790 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:03Z","lastTransitionTime":"2025-11-22T07:57:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:03 crc kubenswrapper[4789]: I1122 07:57:03.630381 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:03 crc kubenswrapper[4789]: I1122 07:57:03.630441 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:03 crc kubenswrapper[4789]: I1122 07:57:03.630456 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:03 crc kubenswrapper[4789]: I1122 07:57:03.630475 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:03 crc kubenswrapper[4789]: I1122 07:57:03.630490 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:03Z","lastTransitionTime":"2025-11-22T07:57:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:03 crc kubenswrapper[4789]: I1122 07:57:03.658094 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tps6j"] Nov 22 07:57:03 crc kubenswrapper[4789]: I1122 07:57:03.658732 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tps6j" Nov 22 07:57:03 crc kubenswrapper[4789]: I1122 07:57:03.660499 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Nov 22 07:57:03 crc kubenswrapper[4789]: I1122 07:57:03.662139 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Nov 22 07:57:03 crc kubenswrapper[4789]: I1122 07:57:03.677465 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66f74155-c96e-4faa-876f-62b8b5e6b3bc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53953e73ae4a11a47b7a5332786b97d9f129d9576e66aa2664836ec5f3786683\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65813cc305e2105a598eb0b701dd079f1f26db8cc1b2e67679e7086c2ba3630f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1d47e3e0a1939863a1ae9e1f52262d8fad029cd90552c21e73a98cdf0a803ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9d09f0deafe8d22df512d2acd120ce10eec182173daaf5029ab72f848ceecd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c410501c102d72819d29aa0036b6270b1ce9804bba84092161b2c2d8ebaa885a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0755dd1f1e7b1fda23c8973b56521ab15b62b9eb76f9b5604110e04e2f25b422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0755dd1f1e7b1fda23c8973b56521ab15b62b9eb76f9b5604110e04e2f25b422\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T07:56:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T07:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://680b49efb88db68b75efc6ca47cdfa449c1cf688d277b3e1eaa75cb466e90480\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://680b49efb88db68b75efc6ca47cdfa449c1cf688d277b3e1eaa75cb466e90480\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T07:56:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T07:56:32Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c5b1cf28577e60931184499a22559f30b0bdf5dab0b6cb542c948fec5d9c54cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5b1cf28577e60931184499a22559f30b0bdf5dab0b6cb542c948fec5d9c54cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T07:56:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T07:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:29Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:03 crc kubenswrapper[4789]: I1122 07:57:03.688694 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:03 crc kubenswrapper[4789]: I1122 07:57:03.696970 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:03 crc kubenswrapper[4789]: I1122 07:57:03.713821 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0cb4cf51-62d6-4f32-a8b4-4116555f3870\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qmxss\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:03 crc kubenswrapper[4789]: I1122 07:57:03.716096 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b7f118eb-f1c8-44c6-a7a0-0e950f5c8e47-env-overrides\") pod \"ovnkube-control-plane-749d76644c-tps6j\" (UID: \"b7f118eb-f1c8-44c6-a7a0-0e950f5c8e47\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tps6j" Nov 22 07:57:03 crc kubenswrapper[4789]: I1122 07:57:03.716147 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b7f118eb-f1c8-44c6-a7a0-0e950f5c8e47-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-tps6j\" (UID: \"b7f118eb-f1c8-44c6-a7a0-0e950f5c8e47\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tps6j" Nov 22 07:57:03 crc kubenswrapper[4789]: I1122 07:57:03.716180 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6q8vz\" (UniqueName: \"kubernetes.io/projected/b7f118eb-f1c8-44c6-a7a0-0e950f5c8e47-kube-api-access-6q8vz\") pod \"ovnkube-control-plane-749d76644c-tps6j\" (UID: \"b7f118eb-f1c8-44c6-a7a0-0e950f5c8e47\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tps6j" Nov 22 07:57:03 crc kubenswrapper[4789]: I1122 07:57:03.716196 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b7f118eb-f1c8-44c6-a7a0-0e950f5c8e47-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-tps6j\" (UID: \"b7f118eb-f1c8-44c6-a7a0-0e950f5c8e47\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tps6j" Nov 22 07:57:03 crc kubenswrapper[4789]: I1122 07:57:03.723084 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tps6j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7f118eb-f1c8-44c6-a7a0-0e950f5c8e47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6q8vz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6q8vz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:57:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tps6j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:03 crc kubenswrapper[4789]: I1122 07:57:03.732575 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:03 crc kubenswrapper[4789]: I1122 07:57:03.733316 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:03 crc kubenswrapper[4789]: I1122 07:57:03.733364 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:03 crc kubenswrapper[4789]: I1122 07:57:03.733374 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:03 crc kubenswrapper[4789]: I1122 07:57:03.733394 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:03 crc kubenswrapper[4789]: I1122 07:57:03.733403 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:03Z","lastTransitionTime":"2025-11-22T07:57:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:03 crc kubenswrapper[4789]: I1122 07:57:03.739946 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9f74f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"29ad5e00-4918-4718-9d49-013b12621f7d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mnz6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9f74f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:03 crc kubenswrapper[4789]: I1122 07:57:03.750858 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2wdp7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2fdee89f-559b-4f4e-8d20-6ecc3aafd22e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxvhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxvhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxvhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxvhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxvhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxvhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxvhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2wdp7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:03 crc kubenswrapper[4789]: I1122 07:57:03.761163 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:03 crc kubenswrapper[4789]: I1122 07:57:03.772507 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:03 crc kubenswrapper[4789]: I1122 07:57:03.781230 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58aa8071-7ecc-4692-8789-2db04ec70647\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfjgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfjgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pxvrm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:03 crc kubenswrapper[4789]: I1122 07:57:03.790265 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vrkrh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"371275b1-1d13-4f3c-8190-8c59c080fef5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:54Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:54Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8w4z2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vrkrh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:03 crc kubenswrapper[4789]: I1122 07:57:03.802330 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d081b29e-edd7-4fb0-a2e5-431ee8491bd0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43d9e10254883b6491bef617cf9f0b319171bc1fbc49c7770ed9630a8a1ae2cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://223c2300bb6afb8812a72eb7605015b0473f59046565cdf39fb9b9f0cf6634bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca5af01b2e23b2bcac23ed767bc7dcd2ac0be1010ee5e675b3b44333ee337180\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8e79bfb6172c894868a8e1b47ac701ba4fdaca1fc6b1582f7e70807aadfdf1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:29Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:03 crc kubenswrapper[4789]: I1122 07:57:03.812165 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:03 crc kubenswrapper[4789]: I1122 07:57:03.816661 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b7f118eb-f1c8-44c6-a7a0-0e950f5c8e47-env-overrides\") pod \"ovnkube-control-plane-749d76644c-tps6j\" (UID: \"b7f118eb-f1c8-44c6-a7a0-0e950f5c8e47\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tps6j" Nov 22 07:57:03 crc kubenswrapper[4789]: I1122 07:57:03.816712 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b7f118eb-f1c8-44c6-a7a0-0e950f5c8e47-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-tps6j\" (UID: \"b7f118eb-f1c8-44c6-a7a0-0e950f5c8e47\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tps6j" Nov 22 07:57:03 crc kubenswrapper[4789]: I1122 07:57:03.816778 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6q8vz\" (UniqueName: \"kubernetes.io/projected/b7f118eb-f1c8-44c6-a7a0-0e950f5c8e47-kube-api-access-6q8vz\") pod \"ovnkube-control-plane-749d76644c-tps6j\" (UID: \"b7f118eb-f1c8-44c6-a7a0-0e950f5c8e47\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tps6j" Nov 22 07:57:03 crc kubenswrapper[4789]: I1122 07:57:03.816806 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b7f118eb-f1c8-44c6-a7a0-0e950f5c8e47-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-tps6j\" (UID: \"b7f118eb-f1c8-44c6-a7a0-0e950f5c8e47\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tps6j" Nov 22 07:57:03 crc kubenswrapper[4789]: I1122 07:57:03.818074 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b7f118eb-f1c8-44c6-a7a0-0e950f5c8e47-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-tps6j\" (UID: \"b7f118eb-f1c8-44c6-a7a0-0e950f5c8e47\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tps6j" Nov 22 07:57:03 crc kubenswrapper[4789]: I1122 07:57:03.818307 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b7f118eb-f1c8-44c6-a7a0-0e950f5c8e47-env-overrides\") pod \"ovnkube-control-plane-749d76644c-tps6j\" (UID: \"b7f118eb-f1c8-44c6-a7a0-0e950f5c8e47\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tps6j" Nov 22 07:57:03 crc kubenswrapper[4789]: I1122 07:57:03.820899 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b7f118eb-f1c8-44c6-a7a0-0e950f5c8e47-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-tps6j\" (UID: \"b7f118eb-f1c8-44c6-a7a0-0e950f5c8e47\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tps6j" Nov 22 07:57:03 crc kubenswrapper[4789]: I1122 07:57:03.823522 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rks5j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5613c209-59a0-4b8a-ac8b-5c633bc1bdc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wj8sc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rks5j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:03 crc kubenswrapper[4789]: I1122 07:57:03.833379 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6q8vz\" (UniqueName: \"kubernetes.io/projected/b7f118eb-f1c8-44c6-a7a0-0e950f5c8e47-kube-api-access-6q8vz\") pod \"ovnkube-control-plane-749d76644c-tps6j\" (UID: \"b7f118eb-f1c8-44c6-a7a0-0e950f5c8e47\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tps6j" Nov 22 07:57:03 crc kubenswrapper[4789]: I1122 07:57:03.835502 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:03 crc kubenswrapper[4789]: I1122 07:57:03.835668 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:03 crc kubenswrapper[4789]: I1122 07:57:03.835777 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:03 crc kubenswrapper[4789]: I1122 07:57:03.835889 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:03 crc kubenswrapper[4789]: I1122 07:57:03.835996 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:03Z","lastTransitionTime":"2025-11-22T07:57:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:03 crc kubenswrapper[4789]: I1122 07:57:03.839571 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93be48ce-7d87-448b-b03e-ab6736cfc34c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba91466edcf49f73751622f98632b3040a046ccd8a42ba8ecee7ab08688a8c24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b6666e0ce42a46f669505659298a19931b1bf7c253ae6a97c6acfd80a9e717f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://251e72e1cd0f3e0ee195bacf33ab38765d1dd943adf4395f9d6dd282bda024cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://098c31b822e7c2351f1142af36b942d67ff8262896e28022c96a187d40bd0ddb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://098c31b822e7c2351f1142af36b942d67ff8262896e28022c96a187d40bd0ddb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"message\\\":\\\"le observer\\\\nW1122 07:56:50.641735 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1122 07:56:50.642792 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 07:56:50.643945 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3588433834/tls.crt::/tmp/serving-cert-3588433834/tls.key\\\\\\\"\\\\nI1122 07:56:50.949040 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 07:56:50.952068 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 07:56:50.952092 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 07:56:50.952116 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 07:56:50.952121 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 07:56:50.958549 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1122 07:56:50.958565 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1122 07:56:50.958579 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 07:56:50.958585 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 07:56:50.958591 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 07:56:50.958596 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1122 07:56:50.958600 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 07:56:50.958605 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1122 07:56:50.962189 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T07:56:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://263dce4c3f4f3fe388b21d4bb55a9fc7d956c3cca7da34ee00f470d88ef501f1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac2b567fcebc517d3fb4366aebce03351183ea03ce62f86c2abdf5b95e20b34a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac2b567fcebc517d3fb4366aebce03351183ea03ce62f86c2abdf5b95e20b34a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T07:56:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T07:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:03 crc kubenswrapper[4789]: I1122 07:57:03.939085 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:03 crc kubenswrapper[4789]: I1122 07:57:03.939140 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:03 crc kubenswrapper[4789]: I1122 07:57:03.939153 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:03 crc kubenswrapper[4789]: I1122 07:57:03.939170 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:03 crc kubenswrapper[4789]: I1122 07:57:03.939183 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:03Z","lastTransitionTime":"2025-11-22T07:57:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:03 crc kubenswrapper[4789]: I1122 07:57:03.964966 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 07:57:03 crc kubenswrapper[4789]: E1122 07:57:03.965305 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 07:57:03 crc kubenswrapper[4789]: I1122 07:57:03.973635 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tps6j" Nov 22 07:57:04 crc kubenswrapper[4789]: I1122 07:57:04.042714 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:04 crc kubenswrapper[4789]: I1122 07:57:04.042775 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:04 crc kubenswrapper[4789]: I1122 07:57:04.042798 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:04 crc kubenswrapper[4789]: I1122 07:57:04.042816 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:04 crc kubenswrapper[4789]: I1122 07:57:04.042827 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:04Z","lastTransitionTime":"2025-11-22T07:57:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:04 crc kubenswrapper[4789]: I1122 07:57:04.145246 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:04 crc kubenswrapper[4789]: I1122 07:57:04.145291 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:04 crc kubenswrapper[4789]: I1122 07:57:04.145303 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:04 crc kubenswrapper[4789]: I1122 07:57:04.145319 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:04 crc kubenswrapper[4789]: I1122 07:57:04.145329 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:04Z","lastTransitionTime":"2025-11-22T07:57:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:04 crc kubenswrapper[4789]: I1122 07:57:04.159977 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tps6j" event={"ID":"b7f118eb-f1c8-44c6-a7a0-0e950f5c8e47","Type":"ContainerStarted","Data":"82796c542e888ea082088d95608b293bf5a05d02ba723a98a9fe5f997037dcb0"} Nov 22 07:57:04 crc kubenswrapper[4789]: I1122 07:57:04.247179 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:04 crc kubenswrapper[4789]: I1122 07:57:04.247230 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:04 crc kubenswrapper[4789]: I1122 07:57:04.247241 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:04 crc kubenswrapper[4789]: I1122 07:57:04.247259 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:04 crc kubenswrapper[4789]: I1122 07:57:04.247272 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:04Z","lastTransitionTime":"2025-11-22T07:57:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:04 crc kubenswrapper[4789]: I1122 07:57:04.350110 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:04 crc kubenswrapper[4789]: I1122 07:57:04.350161 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:04 crc kubenswrapper[4789]: I1122 07:57:04.350173 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:04 crc kubenswrapper[4789]: I1122 07:57:04.350189 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:04 crc kubenswrapper[4789]: I1122 07:57:04.350202 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:04Z","lastTransitionTime":"2025-11-22T07:57:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:04 crc kubenswrapper[4789]: I1122 07:57:04.452921 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:04 crc kubenswrapper[4789]: I1122 07:57:04.453432 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:04 crc kubenswrapper[4789]: I1122 07:57:04.453451 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:04 crc kubenswrapper[4789]: I1122 07:57:04.453476 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:04 crc kubenswrapper[4789]: I1122 07:57:04.453493 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:04Z","lastTransitionTime":"2025-11-22T07:57:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:04 crc kubenswrapper[4789]: I1122 07:57:04.556446 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:04 crc kubenswrapper[4789]: I1122 07:57:04.556503 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:04 crc kubenswrapper[4789]: I1122 07:57:04.556520 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:04 crc kubenswrapper[4789]: I1122 07:57:04.556536 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:04 crc kubenswrapper[4789]: I1122 07:57:04.556545 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:04Z","lastTransitionTime":"2025-11-22T07:57:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:04 crc kubenswrapper[4789]: I1122 07:57:04.659553 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:04 crc kubenswrapper[4789]: I1122 07:57:04.659594 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:04 crc kubenswrapper[4789]: I1122 07:57:04.659621 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:04 crc kubenswrapper[4789]: I1122 07:57:04.659639 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:04 crc kubenswrapper[4789]: I1122 07:57:04.659651 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:04Z","lastTransitionTime":"2025-11-22T07:57:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:04 crc kubenswrapper[4789]: I1122 07:57:04.762446 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:04 crc kubenswrapper[4789]: I1122 07:57:04.762485 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:04 crc kubenswrapper[4789]: I1122 07:57:04.762499 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:04 crc kubenswrapper[4789]: I1122 07:57:04.762515 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:04 crc kubenswrapper[4789]: I1122 07:57:04.762526 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:04Z","lastTransitionTime":"2025-11-22T07:57:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:04 crc kubenswrapper[4789]: I1122 07:57:04.865888 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:04 crc kubenswrapper[4789]: I1122 07:57:04.865954 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:04 crc kubenswrapper[4789]: I1122 07:57:04.865963 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:04 crc kubenswrapper[4789]: I1122 07:57:04.865980 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:04 crc kubenswrapper[4789]: I1122 07:57:04.865991 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:04Z","lastTransitionTime":"2025-11-22T07:57:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:04 crc kubenswrapper[4789]: I1122 07:57:04.965528 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 07:57:04 crc kubenswrapper[4789]: I1122 07:57:04.965790 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 07:57:04 crc kubenswrapper[4789]: E1122 07:57:04.965811 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 07:57:04 crc kubenswrapper[4789]: E1122 07:57:04.966280 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 07:57:04 crc kubenswrapper[4789]: I1122 07:57:04.968377 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:04 crc kubenswrapper[4789]: I1122 07:57:04.968414 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:04 crc kubenswrapper[4789]: I1122 07:57:04.968427 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:04 crc kubenswrapper[4789]: I1122 07:57:04.968444 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:04 crc kubenswrapper[4789]: I1122 07:57:04.968458 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:04Z","lastTransitionTime":"2025-11-22T07:57:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.071283 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.071315 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.071323 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.071336 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.071345 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:05Z","lastTransitionTime":"2025-11-22T07:57:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.165022 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tps6j" event={"ID":"b7f118eb-f1c8-44c6-a7a0-0e950f5c8e47","Type":"ContainerStarted","Data":"1f10fa30e2fcad3c91a3cc323797ff15355f5ec68e42660aee3587c8b2161a92"} Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.165070 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tps6j" event={"ID":"b7f118eb-f1c8-44c6-a7a0-0e950f5c8e47","Type":"ContainerStarted","Data":"b1bd2d93c66e740ad46549c4d36f463be72975f230fc71c52fba21d211a3d132"} Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.167558 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-9f74f" event={"ID":"29ad5e00-4918-4718-9d49-013b12621f7d","Type":"ContainerStarted","Data":"c5b1f947f92448ec5c74bda8d4e53fc56ae353887c51580368e07546a4d398b6"} Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.175772 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.176905 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.176929 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.176938 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.176952 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.176962 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:05Z","lastTransitionTime":"2025-11-22T07:57:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.187160 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.206594 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0cb4cf51-62d6-4f32-a8b4-4116555f3870\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qmxss\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.218729 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tps6j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7f118eb-f1c8-44c6-a7a0-0e950f5c8e47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1bd2d93c66e740ad46549c4d36f463be72975f230fc71c52fba21d211a3d132\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6q8vz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f10fa30e2fcad3c91a3cc323797ff15355f5ec68e42660aee3587c8b2161a92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6q8vz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:57:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tps6j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.236766 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66f74155-c96e-4faa-876f-62b8b5e6b3bc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53953e73ae4a11a47b7a5332786b97d9f129d9576e66aa2664836ec5f3786683\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65813cc305e2105a598eb0b701dd079f1f26db8cc1b2e67679e7086c2ba3630f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1d47e3e0a1939863a1ae9e1f52262d8fad029cd90552c21e73a98cdf0a803ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9d09f0deafe8d22df512d2acd120ce10eec182173daaf5029ab72f848ceecd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c410501c102d72819d29aa0036b6270b1ce9804bba84092161b2c2d8ebaa885a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0755dd1f1e7b1fda23c8973b56521ab15b62b9eb76f9b5604110e04e2f25b422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0755dd1f1e7b1fda23c8973b56521ab15b62b9eb76f9b5604110e04e2f25b422\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T07:56:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T07:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://680b49efb88db68b75efc6ca47cdfa449c1cf688d277b3e1eaa75cb466e90480\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://680b49efb88db68b75efc6ca47cdfa449c1cf688d277b3e1eaa75cb466e90480\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T07:56:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T07:56:32Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c5b1cf28577e60931184499a22559f30b0bdf5dab0b6cb542c948fec5d9c54cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5b1cf28577e60931184499a22559f30b0bdf5dab0b6cb542c948fec5d9c54cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T07:56:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T07:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:29Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.244658 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9f74f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"29ad5e00-4918-4718-9d49-013b12621f7d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mnz6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9f74f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.256135 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2wdp7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2fdee89f-559b-4f4e-8d20-6ecc3aafd22e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxvhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxvhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxvhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxvhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxvhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxvhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxvhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2wdp7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.267150 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.280604 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.280656 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.280669 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.280692 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.280710 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:05Z","lastTransitionTime":"2025-11-22T07:57:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.283249 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.293690 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.302863 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58aa8071-7ecc-4692-8789-2db04ec70647\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfjgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfjgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pxvrm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.311409 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vrkrh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"371275b1-1d13-4f3c-8190-8c59c080fef5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:54Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:54Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8w4z2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vrkrh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.319937 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d081b29e-edd7-4fb0-a2e5-431ee8491bd0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43d9e10254883b6491bef617cf9f0b319171bc1fbc49c7770ed9630a8a1ae2cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://223c2300bb6afb8812a72eb7605015b0473f59046565cdf39fb9b9f0cf6634bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca5af01b2e23b2bcac23ed767bc7dcd2ac0be1010ee5e675b3b44333ee337180\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8e79bfb6172c894868a8e1b47ac701ba4fdaca1fc6b1582f7e70807aadfdf1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:29Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.330594 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rks5j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5613c209-59a0-4b8a-ac8b-5c633bc1bdc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wj8sc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rks5j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.341407 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93be48ce-7d87-448b-b03e-ab6736cfc34c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba91466edcf49f73751622f98632b3040a046ccd8a42ba8ecee7ab08688a8c24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b6666e0ce42a46f669505659298a19931b1bf7c253ae6a97c6acfd80a9e717f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://251e72e1cd0f3e0ee195bacf33ab38765d1dd943adf4395f9d6dd282bda024cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://098c31b822e7c2351f1142af36b942d67ff8262896e28022c96a187d40bd0ddb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://098c31b822e7c2351f1142af36b942d67ff8262896e28022c96a187d40bd0ddb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"message\\\":\\\"le observer\\\\nW1122 07:56:50.641735 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1122 07:56:50.642792 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 07:56:50.643945 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3588433834/tls.crt::/tmp/serving-cert-3588433834/tls.key\\\\\\\"\\\\nI1122 07:56:50.949040 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 07:56:50.952068 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 07:56:50.952092 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 07:56:50.952116 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 07:56:50.952121 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 07:56:50.958549 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1122 07:56:50.958565 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1122 07:56:50.958579 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 07:56:50.958585 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 07:56:50.958591 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 07:56:50.958596 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1122 07:56:50.958600 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 07:56:50.958605 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1122 07:56:50.962189 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T07:56:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://263dce4c3f4f3fe388b21d4bb55a9fc7d956c3cca7da34ee00f470d88ef501f1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac2b567fcebc517d3fb4366aebce03351183ea03ce62f86c2abdf5b95e20b34a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac2b567fcebc517d3fb4366aebce03351183ea03ce62f86c2abdf5b95e20b34a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T07:56:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T07:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.350824 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.361573 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.371110 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rks5j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5613c209-59a0-4b8a-ac8b-5c633bc1bdc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wj8sc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rks5j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.380914 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93be48ce-7d87-448b-b03e-ab6736cfc34c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba91466edcf49f73751622f98632b3040a046ccd8a42ba8ecee7ab08688a8c24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b6666e0ce42a46f669505659298a19931b1bf7c253ae6a97c6acfd80a9e717f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://251e72e1cd0f3e0ee195bacf33ab38765d1dd943adf4395f9d6dd282bda024cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://098c31b822e7c2351f1142af36b942d67ff8262896e28022c96a187d40bd0ddb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://098c31b822e7c2351f1142af36b942d67ff8262896e28022c96a187d40bd0ddb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"message\\\":\\\"le observer\\\\nW1122 07:56:50.641735 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1122 07:56:50.642792 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 07:56:50.643945 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3588433834/tls.crt::/tmp/serving-cert-3588433834/tls.key\\\\\\\"\\\\nI1122 07:56:50.949040 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 07:56:50.952068 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 07:56:50.952092 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 07:56:50.952116 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 07:56:50.952121 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 07:56:50.958549 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1122 07:56:50.958565 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1122 07:56:50.958579 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 07:56:50.958585 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 07:56:50.958591 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 07:56:50.958596 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1122 07:56:50.958600 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 07:56:50.958605 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1122 07:56:50.962189 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T07:56:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://263dce4c3f4f3fe388b21d4bb55a9fc7d956c3cca7da34ee00f470d88ef501f1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac2b567fcebc517d3fb4366aebce03351183ea03ce62f86c2abdf5b95e20b34a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac2b567fcebc517d3fb4366aebce03351183ea03ce62f86c2abdf5b95e20b34a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T07:56:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T07:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.382914 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.382954 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.382974 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.382995 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.383010 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:05Z","lastTransitionTime":"2025-11-22T07:57:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.391053 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-q7rc4"] Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.391609 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-q7rc4" Nov 22 07:57:05 crc kubenswrapper[4789]: E1122 07:57:05.391691 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-q7rc4" podUID="8adfba92-180a-4f0a-adc0-d1e2bdd62dba" Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.398314 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66f74155-c96e-4faa-876f-62b8b5e6b3bc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53953e73ae4a11a47b7a5332786b97d9f129d9576e66aa2664836ec5f3786683\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65813cc305e2105a598eb0b701dd079f1f26db8cc1b2e67679e7086c2ba3630f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1d47e3e0a1939863a1ae9e1f52262d8fad029cd90552c21e73a98cdf0a803ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9d09f0deafe8d22df512d2acd120ce10eec182173daaf5029ab72f848ceecd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c410501c102d72819d29aa0036b6270b1ce9804bba84092161b2c2d8ebaa885a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0755dd1f1e7b1fda23c8973b56521ab15b62b9eb76f9b5604110e04e2f25b422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0755dd1f1e7b1fda23c8973b56521ab15b62b9eb76f9b5604110e04e2f25b422\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T07:56:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T07:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://680b49efb88db68b75efc6ca47cdfa449c1cf688d277b3e1eaa75cb466e90480\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://680b49efb88db68b75efc6ca47cdfa449c1cf688d277b3e1eaa75cb466e90480\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T07:56:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T07:56:32Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c5b1cf28577e60931184499a22559f30b0bdf5dab0b6cb542c948fec5d9c54cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5b1cf28577e60931184499a22559f30b0bdf5dab0b6cb542c948fec5d9c54cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T07:56:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T07:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:29Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.408487 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.416969 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.431340 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0cb4cf51-62d6-4f32-a8b4-4116555f3870\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qmxss\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.439040 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tps6j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7f118eb-f1c8-44c6-a7a0-0e950f5c8e47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1bd2d93c66e740ad46549c4d36f463be72975f230fc71c52fba21d211a3d132\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6q8vz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f10fa30e2fcad3c91a3cc323797ff15355f5ec68e42660aee3587c8b2161a92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6q8vz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:57:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tps6j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.452825 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.461445 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9f74f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"29ad5e00-4918-4718-9d49-013b12621f7d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5b1f947f92448ec5c74bda8d4e53fc56ae353887c51580368e07546a4d398b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mnz6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9f74f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.474231 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2wdp7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2fdee89f-559b-4f4e-8d20-6ecc3aafd22e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxvhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxvhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxvhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxvhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxvhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxvhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxvhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2wdp7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.483548 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.485170 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.485220 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.485232 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.485249 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.485261 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:05Z","lastTransitionTime":"2025-11-22T07:57:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.495345 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.506590 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58aa8071-7ecc-4692-8789-2db04ec70647\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfjgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfjgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pxvrm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.513396 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vrkrh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"371275b1-1d13-4f3c-8190-8c59c080fef5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:54Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:54Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8w4z2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vrkrh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.522520 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d081b29e-edd7-4fb0-a2e5-431ee8491bd0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43d9e10254883b6491bef617cf9f0b319171bc1fbc49c7770ed9630a8a1ae2cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://223c2300bb6afb8812a72eb7605015b0473f59046565cdf39fb9b9f0cf6634bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca5af01b2e23b2bcac23ed767bc7dcd2ac0be1010ee5e675b3b44333ee337180\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8e79bfb6172c894868a8e1b47ac701ba4fdaca1fc6b1582f7e70807aadfdf1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:29Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.535524 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2qtmr\" (UniqueName: \"kubernetes.io/projected/8adfba92-180a-4f0a-adc0-d1e2bdd62dba-kube-api-access-2qtmr\") pod \"network-metrics-daemon-q7rc4\" (UID: \"8adfba92-180a-4f0a-adc0-d1e2bdd62dba\") " pod="openshift-multus/network-metrics-daemon-q7rc4" Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.535581 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8adfba92-180a-4f0a-adc0-d1e2bdd62dba-metrics-certs\") pod \"network-metrics-daemon-q7rc4\" (UID: \"8adfba92-180a-4f0a-adc0-d1e2bdd62dba\") " pod="openshift-multus/network-metrics-daemon-q7rc4" Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.536616 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93be48ce-7d87-448b-b03e-ab6736cfc34c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba91466edcf49f73751622f98632b3040a046ccd8a42ba8ecee7ab08688a8c24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b6666e0ce42a46f669505659298a19931b1bf7c253ae6a97c6acfd80a9e717f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://251e72e1cd0f3e0ee195bacf33ab38765d1dd943adf4395f9d6dd282bda024cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://098c31b822e7c2351f1142af36b942d67ff8262896e28022c96a187d40bd0ddb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://098c31b822e7c2351f1142af36b942d67ff8262896e28022c96a187d40bd0ddb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"message\\\":\\\"le observer\\\\nW1122 07:56:50.641735 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1122 07:56:50.642792 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 07:56:50.643945 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3588433834/tls.crt::/tmp/serving-cert-3588433834/tls.key\\\\\\\"\\\\nI1122 07:56:50.949040 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 07:56:50.952068 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 07:56:50.952092 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 07:56:50.952116 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 07:56:50.952121 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 07:56:50.958549 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1122 07:56:50.958565 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1122 07:56:50.958579 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 07:56:50.958585 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 07:56:50.958591 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 07:56:50.958596 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1122 07:56:50.958600 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 07:56:50.958605 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1122 07:56:50.962189 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T07:56:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://263dce4c3f4f3fe388b21d4bb55a9fc7d956c3cca7da34ee00f470d88ef501f1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac2b567fcebc517d3fb4366aebce03351183ea03ce62f86c2abdf5b95e20b34a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac2b567fcebc517d3fb4366aebce03351183ea03ce62f86c2abdf5b95e20b34a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T07:56:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T07:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.547967 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.556577 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rks5j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5613c209-59a0-4b8a-ac8b-5c633bc1bdc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wj8sc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rks5j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.572145 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0cb4cf51-62d6-4f32-a8b4-4116555f3870\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qmxss\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.579661 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tps6j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7f118eb-f1c8-44c6-a7a0-0e950f5c8e47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1bd2d93c66e740ad46549c4d36f463be72975f230fc71c52fba21d211a3d132\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6q8vz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f10fa30e2fcad3c91a3cc323797ff15355f5ec68e42660aee3587c8b2161a92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6q8vz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:57:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tps6j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.586531 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-q7rc4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8adfba92-180a-4f0a-adc0-d1e2bdd62dba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qtmr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qtmr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:57:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-q7rc4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.587391 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.587450 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.587465 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.587481 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.587492 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:05Z","lastTransitionTime":"2025-11-22T07:57:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.601541 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66f74155-c96e-4faa-876f-62b8b5e6b3bc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53953e73ae4a11a47b7a5332786b97d9f129d9576e66aa2664836ec5f3786683\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65813cc305e2105a598eb0b701dd079f1f26db8cc1b2e67679e7086c2ba3630f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1d47e3e0a1939863a1ae9e1f52262d8fad029cd90552c21e73a98cdf0a803ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9d09f0deafe8d22df512d2acd120ce10eec182173daaf5029ab72f848ceecd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c410501c102d72819d29aa0036b6270b1ce9804bba84092161b2c2d8ebaa885a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0755dd1f1e7b1fda23c8973b56521ab15b62b9eb76f9b5604110e04e2f25b422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0755dd1f1e7b1fda23c8973b56521ab15b62b9eb76f9b5604110e04e2f25b422\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T07:56:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T07:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://680b49efb88db68b75efc6ca47cdfa449c1cf688d277b3e1eaa75cb466e90480\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://680b49efb88db68b75efc6ca47cdfa449c1cf688d277b3e1eaa75cb466e90480\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T07:56:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T07:56:32Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c5b1cf28577e60931184499a22559f30b0bdf5dab0b6cb542c948fec5d9c54cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5b1cf28577e60931184499a22559f30b0bdf5dab0b6cb542c948fec5d9c54cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T07:56:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T07:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:29Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.611037 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.619574 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.628371 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.636960 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2qtmr\" (UniqueName: \"kubernetes.io/projected/8adfba92-180a-4f0a-adc0-d1e2bdd62dba-kube-api-access-2qtmr\") pod \"network-metrics-daemon-q7rc4\" (UID: \"8adfba92-180a-4f0a-adc0-d1e2bdd62dba\") " pod="openshift-multus/network-metrics-daemon-q7rc4" Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.637068 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8adfba92-180a-4f0a-adc0-d1e2bdd62dba-metrics-certs\") pod \"network-metrics-daemon-q7rc4\" (UID: \"8adfba92-180a-4f0a-adc0-d1e2bdd62dba\") " pod="openshift-multus/network-metrics-daemon-q7rc4" Nov 22 07:57:05 crc kubenswrapper[4789]: E1122 07:57:05.637207 4789 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 22 07:57:05 crc kubenswrapper[4789]: E1122 07:57:05.637272 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8adfba92-180a-4f0a-adc0-d1e2bdd62dba-metrics-certs podName:8adfba92-180a-4f0a-adc0-d1e2bdd62dba nodeName:}" failed. No retries permitted until 2025-11-22 07:57:06.137254548 +0000 UTC m=+40.371654821 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8adfba92-180a-4f0a-adc0-d1e2bdd62dba-metrics-certs") pod "network-metrics-daemon-q7rc4" (UID: "8adfba92-180a-4f0a-adc0-d1e2bdd62dba") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.638365 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.646416 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9f74f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"29ad5e00-4918-4718-9d49-013b12621f7d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5b1f947f92448ec5c74bda8d4e53fc56ae353887c51580368e07546a4d398b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mnz6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9f74f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.656563 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2wdp7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2fdee89f-559b-4f4e-8d20-6ecc3aafd22e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxvhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxvhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxvhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxvhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxvhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxvhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxvhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2wdp7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.659462 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2qtmr\" (UniqueName: \"kubernetes.io/projected/8adfba92-180a-4f0a-adc0-d1e2bdd62dba-kube-api-access-2qtmr\") pod \"network-metrics-daemon-q7rc4\" (UID: \"8adfba92-180a-4f0a-adc0-d1e2bdd62dba\") " pod="openshift-multus/network-metrics-daemon-q7rc4" Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.665950 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vrkrh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"371275b1-1d13-4f3c-8190-8c59c080fef5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:54Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:54Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8w4z2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vrkrh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.674739 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d081b29e-edd7-4fb0-a2e5-431ee8491bd0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43d9e10254883b6491bef617cf9f0b319171bc1fbc49c7770ed9630a8a1ae2cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://223c2300bb6afb8812a72eb7605015b0473f59046565cdf39fb9b9f0cf6634bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca5af01b2e23b2bcac23ed767bc7dcd2ac0be1010ee5e675b3b44333ee337180\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8e79bfb6172c894868a8e1b47ac701ba4fdaca1fc6b1582f7e70807aadfdf1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:29Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.684475 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.689727 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.689940 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.690024 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.690127 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.690215 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:05Z","lastTransitionTime":"2025-11-22T07:57:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.696371 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58aa8071-7ecc-4692-8789-2db04ec70647\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfjgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfjgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pxvrm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.793319 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.793361 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.793373 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.793392 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.793405 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:05Z","lastTransitionTime":"2025-11-22T07:57:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.896393 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.896477 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.896503 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.896537 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.896560 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:05Z","lastTransitionTime":"2025-11-22T07:57:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.965061 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 07:57:05 crc kubenswrapper[4789]: E1122 07:57:05.965572 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.999603 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.999645 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.999659 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.999678 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:05 crc kubenswrapper[4789]: I1122 07:57:05.999693 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:05Z","lastTransitionTime":"2025-11-22T07:57:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:06 crc kubenswrapper[4789]: I1122 07:57:06.103045 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:06 crc kubenswrapper[4789]: I1122 07:57:06.103087 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:06 crc kubenswrapper[4789]: I1122 07:57:06.103098 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:06 crc kubenswrapper[4789]: I1122 07:57:06.103113 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:06 crc kubenswrapper[4789]: I1122 07:57:06.103125 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:06Z","lastTransitionTime":"2025-11-22T07:57:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:06 crc kubenswrapper[4789]: I1122 07:57:06.141892 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8adfba92-180a-4f0a-adc0-d1e2bdd62dba-metrics-certs\") pod \"network-metrics-daemon-q7rc4\" (UID: \"8adfba92-180a-4f0a-adc0-d1e2bdd62dba\") " pod="openshift-multus/network-metrics-daemon-q7rc4" Nov 22 07:57:06 crc kubenswrapper[4789]: E1122 07:57:06.142165 4789 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 22 07:57:06 crc kubenswrapper[4789]: E1122 07:57:06.142287 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8adfba92-180a-4f0a-adc0-d1e2bdd62dba-metrics-certs podName:8adfba92-180a-4f0a-adc0-d1e2bdd62dba nodeName:}" failed. No retries permitted until 2025-11-22 07:57:07.142260938 +0000 UTC m=+41.376661301 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8adfba92-180a-4f0a-adc0-d1e2bdd62dba-metrics-certs") pod "network-metrics-daemon-q7rc4" (UID: "8adfba92-180a-4f0a-adc0-d1e2bdd62dba") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 22 07:57:06 crc kubenswrapper[4789]: I1122 07:57:06.171000 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"19b75892cebd683f74bda91e8123010033c2e9012d3da2703e1391d9e21e27d8"} Nov 22 07:57:06 crc kubenswrapper[4789]: I1122 07:57:06.173210 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"a9cec09dda3f96b4326a7167240730336d426d1b112de39723463adea918e2c9"} Nov 22 07:57:06 crc kubenswrapper[4789]: I1122 07:57:06.174645 4789 generic.go:334] "Generic (PLEG): container finished" podID="2fdee89f-559b-4f4e-8d20-6ecc3aafd22e" containerID="a12a4ef1ebac0fb5977a6e9df35c1a1c1a5fbbcf98b58069ed1bc4cccb581e48" exitCode=0 Nov 22 07:57:06 crc kubenswrapper[4789]: I1122 07:57:06.174934 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2wdp7" event={"ID":"2fdee89f-559b-4f4e-8d20-6ecc3aafd22e","Type":"ContainerDied","Data":"a12a4ef1ebac0fb5977a6e9df35c1a1c1a5fbbcf98b58069ed1bc4cccb581e48"} Nov 22 07:57:06 crc kubenswrapper[4789]: I1122 07:57:06.182557 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rks5j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5613c209-59a0-4b8a-ac8b-5c633bc1bdc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wj8sc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rks5j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:06 crc kubenswrapper[4789]: I1122 07:57:06.202007 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93be48ce-7d87-448b-b03e-ab6736cfc34c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba91466edcf49f73751622f98632b3040a046ccd8a42ba8ecee7ab08688a8c24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b6666e0ce42a46f669505659298a19931b1bf7c253ae6a97c6acfd80a9e717f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://251e72e1cd0f3e0ee195bacf33ab38765d1dd943adf4395f9d6dd282bda024cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://098c31b822e7c2351f1142af36b942d67ff8262896e28022c96a187d40bd0ddb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://098c31b822e7c2351f1142af36b942d67ff8262896e28022c96a187d40bd0ddb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"message\\\":\\\"le observer\\\\nW1122 07:56:50.641735 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1122 07:56:50.642792 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 07:56:50.643945 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3588433834/tls.crt::/tmp/serving-cert-3588433834/tls.key\\\\\\\"\\\\nI1122 07:56:50.949040 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 07:56:50.952068 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 07:56:50.952092 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 07:56:50.952116 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 07:56:50.952121 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 07:56:50.958549 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1122 07:56:50.958565 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1122 07:56:50.958579 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 07:56:50.958585 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 07:56:50.958591 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 07:56:50.958596 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1122 07:56:50.958600 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 07:56:50.958605 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1122 07:56:50.962189 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T07:56:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://263dce4c3f4f3fe388b21d4bb55a9fc7d956c3cca7da34ee00f470d88ef501f1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac2b567fcebc517d3fb4366aebce03351183ea03ce62f86c2abdf5b95e20b34a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac2b567fcebc517d3fb4366aebce03351183ea03ce62f86c2abdf5b95e20b34a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T07:56:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T07:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:06 crc kubenswrapper[4789]: I1122 07:57:06.205819 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:06 crc kubenswrapper[4789]: I1122 07:57:06.205856 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:06 crc kubenswrapper[4789]: I1122 07:57:06.205867 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:06 crc kubenswrapper[4789]: I1122 07:57:06.205884 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:06 crc kubenswrapper[4789]: I1122 07:57:06.205896 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:06Z","lastTransitionTime":"2025-11-22T07:57:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:06 crc kubenswrapper[4789]: I1122 07:57:06.213856 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:06 crc kubenswrapper[4789]: I1122 07:57:06.224526 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:06 crc kubenswrapper[4789]: I1122 07:57:06.233710 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:06 crc kubenswrapper[4789]: I1122 07:57:06.251987 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0cb4cf51-62d6-4f32-a8b4-4116555f3870\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qmxss\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:06 crc kubenswrapper[4789]: I1122 07:57:06.263618 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tps6j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7f118eb-f1c8-44c6-a7a0-0e950f5c8e47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1bd2d93c66e740ad46549c4d36f463be72975f230fc71c52fba21d211a3d132\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6q8vz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f10fa30e2fcad3c91a3cc323797ff15355f5ec68e42660aee3587c8b2161a92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6q8vz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:57:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tps6j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:06 crc kubenswrapper[4789]: I1122 07:57:06.274576 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-q7rc4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8adfba92-180a-4f0a-adc0-d1e2bdd62dba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qtmr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qtmr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:57:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-q7rc4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:06 crc kubenswrapper[4789]: I1122 07:57:06.296658 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66f74155-c96e-4faa-876f-62b8b5e6b3bc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53953e73ae4a11a47b7a5332786b97d9f129d9576e66aa2664836ec5f3786683\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65813cc305e2105a598eb0b701dd079f1f26db8cc1b2e67679e7086c2ba3630f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1d47e3e0a1939863a1ae9e1f52262d8fad029cd90552c21e73a98cdf0a803ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9d09f0deafe8d22df512d2acd120ce10eec182173daaf5029ab72f848ceecd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c410501c102d72819d29aa0036b6270b1ce9804bba84092161b2c2d8ebaa885a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0755dd1f1e7b1fda23c8973b56521ab15b62b9eb76f9b5604110e04e2f25b422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0755dd1f1e7b1fda23c8973b56521ab15b62b9eb76f9b5604110e04e2f25b422\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T07:56:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T07:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://680b49efb88db68b75efc6ca47cdfa449c1cf688d277b3e1eaa75cb466e90480\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://680b49efb88db68b75efc6ca47cdfa449c1cf688d277b3e1eaa75cb466e90480\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T07:56:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T07:56:32Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c5b1cf28577e60931184499a22559f30b0bdf5dab0b6cb542c948fec5d9c54cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5b1cf28577e60931184499a22559f30b0bdf5dab0b6cb542c948fec5d9c54cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T07:56:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T07:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:29Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:06 crc kubenswrapper[4789]: I1122 07:57:06.312575 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9f74f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"29ad5e00-4918-4718-9d49-013b12621f7d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5b1f947f92448ec5c74bda8d4e53fc56ae353887c51580368e07546a4d398b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mnz6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9f74f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:06 crc kubenswrapper[4789]: I1122 07:57:06.315331 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:06 crc kubenswrapper[4789]: I1122 07:57:06.315370 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:06 crc kubenswrapper[4789]: I1122 07:57:06.315383 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:06 crc kubenswrapper[4789]: I1122 07:57:06.315400 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:06 crc kubenswrapper[4789]: I1122 07:57:06.315432 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:06Z","lastTransitionTime":"2025-11-22T07:57:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:06 crc kubenswrapper[4789]: I1122 07:57:06.332052 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2wdp7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2fdee89f-559b-4f4e-8d20-6ecc3aafd22e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxvhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxvhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxvhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxvhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxvhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxvhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxvhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2wdp7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:06 crc kubenswrapper[4789]: I1122 07:57:06.352061 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:06 crc kubenswrapper[4789]: I1122 07:57:06.369856 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:57:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19b75892cebd683f74bda91e8123010033c2e9012d3da2703e1391d9e21e27d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:57:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:06 crc kubenswrapper[4789]: I1122 07:57:06.381550 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:06 crc kubenswrapper[4789]: I1122 07:57:06.391385 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58aa8071-7ecc-4692-8789-2db04ec70647\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfjgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfjgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pxvrm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:06 crc kubenswrapper[4789]: I1122 07:57:06.399058 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vrkrh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"371275b1-1d13-4f3c-8190-8c59c080fef5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:54Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:54Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8w4z2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vrkrh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:06 crc kubenswrapper[4789]: I1122 07:57:06.411052 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d081b29e-edd7-4fb0-a2e5-431ee8491bd0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43d9e10254883b6491bef617cf9f0b319171bc1fbc49c7770ed9630a8a1ae2cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://223c2300bb6afb8812a72eb7605015b0473f59046565cdf39fb9b9f0cf6634bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca5af01b2e23b2bcac23ed767bc7dcd2ac0be1010ee5e675b3b44333ee337180\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8e79bfb6172c894868a8e1b47ac701ba4fdaca1fc6b1582f7e70807aadfdf1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:29Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:06 crc kubenswrapper[4789]: I1122 07:57:06.417724 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:06 crc kubenswrapper[4789]: I1122 07:57:06.417783 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:06 crc kubenswrapper[4789]: I1122 07:57:06.417792 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:06 crc kubenswrapper[4789]: I1122 07:57:06.417806 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:06 crc kubenswrapper[4789]: I1122 07:57:06.417816 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:06Z","lastTransitionTime":"2025-11-22T07:57:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:06 crc kubenswrapper[4789]: I1122 07:57:06.420623 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:06 crc kubenswrapper[4789]: I1122 07:57:06.431860 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:57:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:57:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19b75892cebd683f74bda91e8123010033c2e9012d3da2703e1391d9e21e27d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:57:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:06 crc kubenswrapper[4789]: I1122 07:57:06.439902 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9f74f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"29ad5e00-4918-4718-9d49-013b12621f7d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5b1f947f92448ec5c74bda8d4e53fc56ae353887c51580368e07546a4d398b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mnz6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9f74f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:06 crc kubenswrapper[4789]: I1122 07:57:06.451421 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2wdp7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2fdee89f-559b-4f4e-8d20-6ecc3aafd22e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxvhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a12a4ef1ebac0fb5977a6e9df35c1a1c1a5fbbcf98b58069ed1bc4cccb581e48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a12a4ef1ebac0fb5977a6e9df35c1a1c1a5fbbcf98b58069ed1bc4cccb581e48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T07:57:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T07:57:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxvhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxvhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxvhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxvhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxvhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxvhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2wdp7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:06 crc kubenswrapper[4789]: I1122 07:57:06.451516 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 07:57:06 crc kubenswrapper[4789]: E1122 07:57:06.451724 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 07:57:22.451707906 +0000 UTC m=+56.686108179 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:06 crc kubenswrapper[4789]: I1122 07:57:06.461824 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d081b29e-edd7-4fb0-a2e5-431ee8491bd0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43d9e10254883b6491bef617cf9f0b319171bc1fbc49c7770ed9630a8a1ae2cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://223c2300bb6afb8812a72eb7605015b0473f59046565cdf39fb9b9f0cf6634bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca5af01b2e23b2bcac23ed767bc7dcd2ac0be1010ee5e675b3b44333ee337180\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8e79bfb6172c894868a8e1b47ac701ba4fdaca1fc6b1582f7e70807aadfdf1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:29Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:06 crc kubenswrapper[4789]: I1122 07:57:06.473159 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:06 crc kubenswrapper[4789]: I1122 07:57:06.482380 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58aa8071-7ecc-4692-8789-2db04ec70647\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfjgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfjgw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-pxvrm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:06 crc kubenswrapper[4789]: I1122 07:57:06.490250 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vrkrh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"371275b1-1d13-4f3c-8190-8c59c080fef5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:54Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:54Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8w4z2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vrkrh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:06 crc kubenswrapper[4789]: I1122 07:57:06.502507 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93be48ce-7d87-448b-b03e-ab6736cfc34c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba91466edcf49f73751622f98632b3040a046ccd8a42ba8ecee7ab08688a8c24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b6666e0ce42a46f669505659298a19931b1bf7c253ae6a97c6acfd80a9e717f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://251e72e1cd0f3e0ee195bacf33ab38765d1dd943adf4395f9d6dd282bda024cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://098c31b822e7c2351f1142af36b942d67ff8262896e28022c96a187d40bd0ddb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://098c31b822e7c2351f1142af36b942d67ff8262896e28022c96a187d40bd0ddb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"message\\\":\\\"le observer\\\\nW1122 07:56:50.641735 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1122 07:56:50.642792 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 07:56:50.643945 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3588433834/tls.crt::/tmp/serving-cert-3588433834/tls.key\\\\\\\"\\\\nI1122 07:56:50.949040 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 07:56:50.952068 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 07:56:50.952092 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 07:56:50.952116 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 07:56:50.952121 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 07:56:50.958549 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1122 07:56:50.958565 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1122 07:56:50.958579 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 07:56:50.958585 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 07:56:50.958591 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 07:56:50.958596 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1122 07:56:50.958600 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 07:56:50.958605 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1122 07:56:50.962189 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T07:56:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://263dce4c3f4f3fe388b21d4bb55a9fc7d956c3cca7da34ee00f470d88ef501f1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac2b567fcebc517d3fb4366aebce03351183ea03ce62f86c2abdf5b95e20b34a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac2b567fcebc517d3fb4366aebce03351183ea03ce62f86c2abdf5b95e20b34a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T07:56:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T07:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:06 crc kubenswrapper[4789]: I1122 07:57:06.514357 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:06 crc kubenswrapper[4789]: I1122 07:57:06.519897 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:06 crc kubenswrapper[4789]: I1122 07:57:06.519944 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:06 crc kubenswrapper[4789]: I1122 07:57:06.519954 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:06 crc kubenswrapper[4789]: I1122 07:57:06.519970 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:06 crc kubenswrapper[4789]: I1122 07:57:06.519983 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:06Z","lastTransitionTime":"2025-11-22T07:57:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:06 crc kubenswrapper[4789]: I1122 07:57:06.525462 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rks5j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5613c209-59a0-4b8a-ac8b-5c633bc1bdc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wj8sc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:52Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rks5j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:06 crc kubenswrapper[4789]: I1122 07:57:06.544775 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66f74155-c96e-4faa-876f-62b8b5e6b3bc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53953e73ae4a11a47b7a5332786b97d9f129d9576e66aa2664836ec5f3786683\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65813cc305e2105a598eb0b701dd079f1f26db8cc1b2e67679e7086c2ba3630f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1d47e3e0a1939863a1ae9e1f52262d8fad029cd90552c21e73a98cdf0a803ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9d09f0deafe8d22df512d2acd120ce10eec182173daaf5029ab72f848ceecd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c410501c102d72819d29aa0036b6270b1ce9804bba84092161b2c2d8ebaa885a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0755dd1f1e7b1fda23c8973b56521ab15b62b9eb76f9b5604110e04e2f25b422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0755dd1f1e7b1fda23c8973b56521ab15b62b9eb76f9b5604110e04e2f25b422\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T07:56:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T07:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://680b49efb88db68b75efc6ca47cdfa449c1cf688d277b3e1eaa75cb466e90480\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://680b49efb88db68b75efc6ca47cdfa449c1cf688d277b3e1eaa75cb466e90480\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T07:56:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T07:56:32Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c5b1cf28577e60931184499a22559f30b0bdf5dab0b6cb542c948fec5d9c54cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5b1cf28577e60931184499a22559f30b0bdf5dab0b6cb542c948fec5d9c54cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T07:56:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T07:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:29Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:06 crc kubenswrapper[4789]: I1122 07:57:06.552446 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 07:57:06 crc kubenswrapper[4789]: I1122 07:57:06.552513 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 07:57:06 crc kubenswrapper[4789]: I1122 07:57:06.552543 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 07:57:06 crc kubenswrapper[4789]: I1122 07:57:06.552714 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 07:57:06 crc kubenswrapper[4789]: E1122 07:57:06.552661 4789 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 22 07:57:06 crc kubenswrapper[4789]: E1122 07:57:06.552794 4789 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 22 07:57:06 crc kubenswrapper[4789]: E1122 07:57:06.552842 4789 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 22 07:57:06 crc kubenswrapper[4789]: E1122 07:57:06.552865 4789 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 07:57:06 crc kubenswrapper[4789]: E1122 07:57:06.552665 4789 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 22 07:57:06 crc kubenswrapper[4789]: E1122 07:57:06.552821 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-22 07:57:22.552804172 +0000 UTC m=+56.787204445 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 22 07:57:06 crc kubenswrapper[4789]: E1122 07:57:06.552918 4789 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 22 07:57:06 crc kubenswrapper[4789]: E1122 07:57:06.552961 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-22 07:57:22.552937266 +0000 UTC m=+56.787337539 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 07:57:06 crc kubenswrapper[4789]: E1122 07:57:06.552955 4789 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 22 07:57:06 crc kubenswrapper[4789]: E1122 07:57:06.552989 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-22 07:57:22.552979047 +0000 UTC m=+56.787379320 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 22 07:57:06 crc kubenswrapper[4789]: E1122 07:57:06.552992 4789 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 07:57:06 crc kubenswrapper[4789]: E1122 07:57:06.553074 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-22 07:57:22.553048689 +0000 UTC m=+56.787448962 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 07:57:06 crc kubenswrapper[4789]: I1122 07:57:06.556884 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:06 crc kubenswrapper[4789]: I1122 07:57:06.568203 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:06 crc kubenswrapper[4789]: I1122 07:57:06.586113 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0cb4cf51-62d6-4f32-a8b4-4116555f3870\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qmxss\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:06 crc kubenswrapper[4789]: I1122 07:57:06.594929 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tps6j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7f118eb-f1c8-44c6-a7a0-0e950f5c8e47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1bd2d93c66e740ad46549c4d36f463be72975f230fc71c52fba21d211a3d132\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6q8vz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f10fa30e2fcad3c91a3cc323797ff15355f5ec68e42660aee3587c8b2161a92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6q8vz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:57:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tps6j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:06 crc kubenswrapper[4789]: I1122 07:57:06.605624 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-q7rc4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8adfba92-180a-4f0a-adc0-d1e2bdd62dba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:57:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:57:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qtmr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qtmr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:57:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-q7rc4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 07:57:06 crc kubenswrapper[4789]: I1122 07:57:06.622203 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:06 crc kubenswrapper[4789]: I1122 07:57:06.622240 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:06 crc kubenswrapper[4789]: I1122 07:57:06.622253 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:06 crc kubenswrapper[4789]: I1122 07:57:06.622269 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:06 crc kubenswrapper[4789]: I1122 07:57:06.622280 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:06Z","lastTransitionTime":"2025-11-22T07:57:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:06 crc kubenswrapper[4789]: I1122 07:57:06.724978 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:06 crc kubenswrapper[4789]: I1122 07:57:06.725024 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:06 crc kubenswrapper[4789]: I1122 07:57:06.725035 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:06 crc kubenswrapper[4789]: I1122 07:57:06.725054 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:06 crc kubenswrapper[4789]: I1122 07:57:06.725063 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:06Z","lastTransitionTime":"2025-11-22T07:57:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:06 crc kubenswrapper[4789]: I1122 07:57:06.827345 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:06 crc kubenswrapper[4789]: I1122 07:57:06.827384 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:06 crc kubenswrapper[4789]: I1122 07:57:06.827396 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:06 crc kubenswrapper[4789]: I1122 07:57:06.827413 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:06 crc kubenswrapper[4789]: I1122 07:57:06.827426 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:06Z","lastTransitionTime":"2025-11-22T07:57:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:06 crc kubenswrapper[4789]: I1122 07:57:06.930334 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:06 crc kubenswrapper[4789]: I1122 07:57:06.930402 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:06 crc kubenswrapper[4789]: I1122 07:57:06.930413 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:06 crc kubenswrapper[4789]: I1122 07:57:06.930434 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:06 crc kubenswrapper[4789]: I1122 07:57:06.930443 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:06Z","lastTransitionTime":"2025-11-22T07:57:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:06 crc kubenswrapper[4789]: I1122 07:57:06.964994 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 07:57:06 crc kubenswrapper[4789]: I1122 07:57:06.965104 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-q7rc4" Nov 22 07:57:06 crc kubenswrapper[4789]: E1122 07:57:06.965301 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 07:57:06 crc kubenswrapper[4789]: I1122 07:57:06.965676 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 07:57:06 crc kubenswrapper[4789]: E1122 07:57:06.965766 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 07:57:06 crc kubenswrapper[4789]: E1122 07:57:06.966778 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-q7rc4" podUID="8adfba92-180a-4f0a-adc0-d1e2bdd62dba" Nov 22 07:57:06 crc kubenswrapper[4789]: I1122 07:57:06.966896 4789 scope.go:117] "RemoveContainer" containerID="098c31b822e7c2351f1142af36b942d67ff8262896e28022c96a187d40bd0ddb" Nov 22 07:57:07 crc kubenswrapper[4789]: I1122 07:57:07.032962 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:07 crc kubenswrapper[4789]: I1122 07:57:07.032997 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:07 crc kubenswrapper[4789]: I1122 07:57:07.033009 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:07 crc kubenswrapper[4789]: I1122 07:57:07.033026 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:07 crc kubenswrapper[4789]: I1122 07:57:07.033038 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:07Z","lastTransitionTime":"2025-11-22T07:57:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:07 crc kubenswrapper[4789]: I1122 07:57:07.134593 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:07 crc kubenswrapper[4789]: I1122 07:57:07.134644 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:07 crc kubenswrapper[4789]: I1122 07:57:07.134653 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:07 crc kubenswrapper[4789]: I1122 07:57:07.134668 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:07 crc kubenswrapper[4789]: I1122 07:57:07.134677 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:07Z","lastTransitionTime":"2025-11-22T07:57:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:07 crc kubenswrapper[4789]: I1122 07:57:07.160560 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8adfba92-180a-4f0a-adc0-d1e2bdd62dba-metrics-certs\") pod \"network-metrics-daemon-q7rc4\" (UID: \"8adfba92-180a-4f0a-adc0-d1e2bdd62dba\") " pod="openshift-multus/network-metrics-daemon-q7rc4" Nov 22 07:57:07 crc kubenswrapper[4789]: E1122 07:57:07.160798 4789 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 22 07:57:07 crc kubenswrapper[4789]: E1122 07:57:07.160872 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8adfba92-180a-4f0a-adc0-d1e2bdd62dba-metrics-certs podName:8adfba92-180a-4f0a-adc0-d1e2bdd62dba nodeName:}" failed. No retries permitted until 2025-11-22 07:57:09.160853552 +0000 UTC m=+43.395253825 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8adfba92-180a-4f0a-adc0-d1e2bdd62dba-metrics-certs") pod "network-metrics-daemon-q7rc4" (UID: "8adfba92-180a-4f0a-adc0-d1e2bdd62dba") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 22 07:57:07 crc kubenswrapper[4789]: I1122 07:57:07.178897 4789 generic.go:334] "Generic (PLEG): container finished" podID="0cb4cf51-62d6-4f32-a8b4-4116555f3870" containerID="5ce8875ca32df64c155691ab953221d98fdd435cf3fff8df49a2524cb6a8fca3" exitCode=0 Nov 22 07:57:07 crc kubenswrapper[4789]: I1122 07:57:07.178975 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" event={"ID":"0cb4cf51-62d6-4f32-a8b4-4116555f3870","Type":"ContainerDied","Data":"5ce8875ca32df64c155691ab953221d98fdd435cf3fff8df49a2524cb6a8fca3"} Nov 22 07:57:07 crc kubenswrapper[4789]: I1122 07:57:07.181083 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-rks5j" event={"ID":"5613c209-59a0-4b8a-ac8b-5c633bc1bdc4","Type":"ContainerStarted","Data":"569b7f705aa7fd96b15a7cd1e36497d9f4901040399fa1f49c231af8753a51fc"} Nov 22 07:57:07 crc kubenswrapper[4789]: I1122 07:57:07.184617 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"ba81377a41b41982c4866c2fe4f49bff8eaa64090b780024222fe1f8c9f3c983"} Nov 22 07:57:07 crc kubenswrapper[4789]: I1122 07:57:07.187093 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2wdp7" event={"ID":"2fdee89f-559b-4f4e-8d20-6ecc3aafd22e","Type":"ContainerStarted","Data":"c4e91b6517e366dbc769c3a8172000bfb805655ff2cc19a69d892ac56021caee"} Nov 22 07:57:07 crc kubenswrapper[4789]: I1122 07:57:07.202407 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66f74155-c96e-4faa-876f-62b8b5e6b3bc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53953e73ae4a11a47b7a5332786b97d9f129d9576e66aa2664836ec5f3786683\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65813cc305e2105a598eb0b701dd079f1f26db8cc1b2e67679e7086c2ba3630f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1d47e3e0a1939863a1ae9e1f52262d8fad029cd90552c21e73a98cdf0a803ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9d09f0deafe8d22df512d2acd120ce10eec182173daaf5029ab72f848ceecd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c410501c102d72819d29aa0036b6270b1ce9804bba84092161b2c2d8ebaa885a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:56:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0755dd1f1e7b1fda23c8973b56521ab15b62b9eb76f9b5604110e04e2f25b422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0755dd1f1e7b1fda23c8973b56521ab15b62b9eb76f9b5604110e04e2f25b422\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T07:56:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T07:56:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://680b49efb88db68b75efc6ca47cdfa449c1cf688d277b3e1eaa75cb466e90480\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://680b49efb88db68b75efc6ca47cdfa449c1cf688d277b3e1eaa75cb466e90480\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T07:56:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T07:56:32Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c5b1cf28577e60931184499a22559f30b0bdf5dab0b6cb542c948fec5d9c54cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5b1cf28577e60931184499a22559f30b0bdf5dab0b6cb542c948fec5d9c54cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T07:56:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T07:56:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:29Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T07:57:07Z is after 2025-08-24T17:21:41Z" Nov 22 07:57:07 crc kubenswrapper[4789]: I1122 07:57:07.215759 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T07:57:07Z is after 2025-08-24T17:21:41Z" Nov 22 07:57:07 crc kubenswrapper[4789]: I1122 07:57:07.232993 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:50Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T07:57:07Z is after 2025-08-24T17:21:41Z" Nov 22 07:57:07 crc kubenswrapper[4789]: I1122 07:57:07.236806 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:07 crc kubenswrapper[4789]: I1122 07:57:07.236854 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:07 crc kubenswrapper[4789]: I1122 07:57:07.236865 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:07 crc kubenswrapper[4789]: I1122 07:57:07.236881 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:07 crc kubenswrapper[4789]: I1122 07:57:07.236891 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:07Z","lastTransitionTime":"2025-11-22T07:57:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:07 crc kubenswrapper[4789]: I1122 07:57:07.252256 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0cb4cf51-62d6-4f32-a8b4-4116555f3870\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:56:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7d59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:56:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qmxss\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T07:57:07Z is after 2025-08-24T17:21:41Z" Nov 22 07:57:07 crc kubenswrapper[4789]: I1122 07:57:07.268266 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tps6j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7f118eb-f1c8-44c6-a7a0-0e950f5c8e47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T07:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1bd2d93c66e740ad46549c4d36f463be72975f230fc71c52fba21d211a3d132\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6q8vz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f10fa30e2fcad3c91a3cc323797ff15355f5ec68e42660aee3587c8b2161a92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T07:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6q8vz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T07:57:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tps6j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T07:57:07Z is after 2025-08-24T17:21:41Z" Nov 22 07:57:07 crc kubenswrapper[4789]: I1122 07:57:07.338454 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:07 crc kubenswrapper[4789]: I1122 07:57:07.338493 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:07 crc kubenswrapper[4789]: I1122 07:57:07.338504 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:07 crc kubenswrapper[4789]: I1122 07:57:07.338522 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:07 crc kubenswrapper[4789]: I1122 07:57:07.338534 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:07Z","lastTransitionTime":"2025-11-22T07:57:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:07 crc kubenswrapper[4789]: I1122 07:57:07.343394 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-9f74f" podStartSLOduration=17.343377645 podStartE2EDuration="17.343377645s" podCreationTimestamp="2025-11-22 07:56:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 07:57:07.343295153 +0000 UTC m=+41.577695426" watchObservedRunningTime="2025-11-22 07:57:07.343377645 +0000 UTC m=+41.577777918" Nov 22 07:57:07 crc kubenswrapper[4789]: I1122 07:57:07.397556 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=15.397529616 podStartE2EDuration="15.397529616s" podCreationTimestamp="2025-11-22 07:56:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 07:57:07.382023248 +0000 UTC m=+41.616423521" watchObservedRunningTime="2025-11-22 07:57:07.397529616 +0000 UTC m=+41.631929889" Nov 22 07:57:07 crc kubenswrapper[4789]: I1122 07:57:07.440851 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:07 crc kubenswrapper[4789]: I1122 07:57:07.441168 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:07 crc kubenswrapper[4789]: I1122 07:57:07.441246 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:07 crc kubenswrapper[4789]: I1122 07:57:07.441336 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:07 crc kubenswrapper[4789]: I1122 07:57:07.441427 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:07Z","lastTransitionTime":"2025-11-22T07:57:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:07 crc kubenswrapper[4789]: I1122 07:57:07.495470 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=14.495450257 podStartE2EDuration="14.495450257s" podCreationTimestamp="2025-11-22 07:56:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 07:57:07.492779925 +0000 UTC m=+41.727180208" watchObservedRunningTime="2025-11-22 07:57:07.495450257 +0000 UTC m=+41.729850540" Nov 22 07:57:07 crc kubenswrapper[4789]: I1122 07:57:07.495797 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-rks5j" podStartSLOduration=17.495790636 podStartE2EDuration="17.495790636s" podCreationTimestamp="2025-11-22 07:56:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 07:57:07.467550234 +0000 UTC m=+41.701950527" watchObservedRunningTime="2025-11-22 07:57:07.495790636 +0000 UTC m=+41.730190909" Nov 22 07:57:07 crc kubenswrapper[4789]: I1122 07:57:07.543959 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:07 crc kubenswrapper[4789]: I1122 07:57:07.544008 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:07 crc kubenswrapper[4789]: I1122 07:57:07.544017 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:07 crc kubenswrapper[4789]: I1122 07:57:07.544032 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:07 crc kubenswrapper[4789]: I1122 07:57:07.544042 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:07Z","lastTransitionTime":"2025-11-22T07:57:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:07 crc kubenswrapper[4789]: I1122 07:57:07.560681 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tps6j" podStartSLOduration=16.560654806 podStartE2EDuration="16.560654806s" podCreationTimestamp="2025-11-22 07:56:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 07:57:07.560159442 +0000 UTC m=+41.794559715" watchObservedRunningTime="2025-11-22 07:57:07.560654806 +0000 UTC m=+41.795055079" Nov 22 07:57:07 crc kubenswrapper[4789]: I1122 07:57:07.646295 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:07 crc kubenswrapper[4789]: I1122 07:57:07.646729 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:07 crc kubenswrapper[4789]: I1122 07:57:07.646740 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:07 crc kubenswrapper[4789]: I1122 07:57:07.646774 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:07 crc kubenswrapper[4789]: I1122 07:57:07.646788 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:07Z","lastTransitionTime":"2025-11-22T07:57:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:07 crc kubenswrapper[4789]: I1122 07:57:07.749715 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:07 crc kubenswrapper[4789]: I1122 07:57:07.750088 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:07 crc kubenswrapper[4789]: I1122 07:57:07.750181 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:07 crc kubenswrapper[4789]: I1122 07:57:07.750289 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:07 crc kubenswrapper[4789]: I1122 07:57:07.750362 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:07Z","lastTransitionTime":"2025-11-22T07:57:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:07 crc kubenswrapper[4789]: I1122 07:57:07.853581 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:07 crc kubenswrapper[4789]: I1122 07:57:07.853637 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:07 crc kubenswrapper[4789]: I1122 07:57:07.853654 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:07 crc kubenswrapper[4789]: I1122 07:57:07.853678 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:07 crc kubenswrapper[4789]: I1122 07:57:07.853689 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:07Z","lastTransitionTime":"2025-11-22T07:57:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:07 crc kubenswrapper[4789]: I1122 07:57:07.957219 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:07 crc kubenswrapper[4789]: I1122 07:57:07.957260 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:07 crc kubenswrapper[4789]: I1122 07:57:07.957268 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:07 crc kubenswrapper[4789]: I1122 07:57:07.957284 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:07 crc kubenswrapper[4789]: I1122 07:57:07.957294 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:07Z","lastTransitionTime":"2025-11-22T07:57:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:07 crc kubenswrapper[4789]: I1122 07:57:07.964566 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 07:57:07 crc kubenswrapper[4789]: E1122 07:57:07.965349 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 07:57:08 crc kubenswrapper[4789]: I1122 07:57:08.062829 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:08 crc kubenswrapper[4789]: I1122 07:57:08.062876 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:08 crc kubenswrapper[4789]: I1122 07:57:08.062891 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:08 crc kubenswrapper[4789]: I1122 07:57:08.062909 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:08 crc kubenswrapper[4789]: I1122 07:57:08.062925 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:08Z","lastTransitionTime":"2025-11-22T07:57:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:08 crc kubenswrapper[4789]: I1122 07:57:08.168457 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:08 crc kubenswrapper[4789]: I1122 07:57:08.168498 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:08 crc kubenswrapper[4789]: I1122 07:57:08.168510 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:08 crc kubenswrapper[4789]: I1122 07:57:08.168534 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:08 crc kubenswrapper[4789]: I1122 07:57:08.168545 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:08Z","lastTransitionTime":"2025-11-22T07:57:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:08 crc kubenswrapper[4789]: I1122 07:57:08.192491 4789 generic.go:334] "Generic (PLEG): container finished" podID="2fdee89f-559b-4f4e-8d20-6ecc3aafd22e" containerID="c4e91b6517e366dbc769c3a8172000bfb805655ff2cc19a69d892ac56021caee" exitCode=0 Nov 22 07:57:08 crc kubenswrapper[4789]: I1122 07:57:08.192531 4789 generic.go:334] "Generic (PLEG): container finished" podID="2fdee89f-559b-4f4e-8d20-6ecc3aafd22e" containerID="baa8603de8553972e5990b28b26ad3774eee39cc246a9c943f2372b361882629" exitCode=0 Nov 22 07:57:08 crc kubenswrapper[4789]: I1122 07:57:08.192566 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2wdp7" event={"ID":"2fdee89f-559b-4f4e-8d20-6ecc3aafd22e","Type":"ContainerDied","Data":"c4e91b6517e366dbc769c3a8172000bfb805655ff2cc19a69d892ac56021caee"} Nov 22 07:57:08 crc kubenswrapper[4789]: I1122 07:57:08.192603 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2wdp7" event={"ID":"2fdee89f-559b-4f4e-8d20-6ecc3aafd22e","Type":"ContainerDied","Data":"baa8603de8553972e5990b28b26ad3774eee39cc246a9c943f2372b361882629"} Nov 22 07:57:08 crc kubenswrapper[4789]: I1122 07:57:08.193921 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"b96c165f33b0114221d22d44ce7e05eecbb17dcb944252e9be188734307e39ac"} Nov 22 07:57:08 crc kubenswrapper[4789]: I1122 07:57:08.195737 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Nov 22 07:57:08 crc kubenswrapper[4789]: I1122 07:57:08.197665 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"da07fe384bbf1a978e802661c0a3bf8079bef6bbce76956c00725322110a899c"} Nov 22 07:57:08 crc kubenswrapper[4789]: I1122 07:57:08.198016 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 22 07:57:08 crc kubenswrapper[4789]: I1122 07:57:08.199583 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" event={"ID":"58aa8071-7ecc-4692-8789-2db04ec70647","Type":"ContainerStarted","Data":"f09e6b1fb73b8c9251c8f302c0780c94671efbe22b36b4a44eca3bd855222ecb"} Nov 22 07:57:08 crc kubenswrapper[4789]: I1122 07:57:08.199630 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" event={"ID":"58aa8071-7ecc-4692-8789-2db04ec70647","Type":"ContainerStarted","Data":"8f2ef4eceebb3af5fedfd9bd9b5e59a01bbba909dc1d449d76acf701bd82428a"} Nov 22 07:57:08 crc kubenswrapper[4789]: I1122 07:57:08.271079 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:08 crc kubenswrapper[4789]: I1122 07:57:08.271130 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:08 crc kubenswrapper[4789]: I1122 07:57:08.271142 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:08 crc kubenswrapper[4789]: I1122 07:57:08.271165 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:08 crc kubenswrapper[4789]: I1122 07:57:08.271177 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:08Z","lastTransitionTime":"2025-11-22T07:57:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:08 crc kubenswrapper[4789]: I1122 07:57:08.272327 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=17.272313401 podStartE2EDuration="17.272313401s" podCreationTimestamp="2025-11-22 07:56:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 07:57:08.259201586 +0000 UTC m=+42.493601879" watchObservedRunningTime="2025-11-22 07:57:08.272313401 +0000 UTC m=+42.506713674" Nov 22 07:57:08 crc kubenswrapper[4789]: I1122 07:57:08.320948 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podStartSLOduration=18.320928512 podStartE2EDuration="18.320928512s" podCreationTimestamp="2025-11-22 07:56:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 07:57:08.320074509 +0000 UTC m=+42.554474792" watchObservedRunningTime="2025-11-22 07:57:08.320928512 +0000 UTC m=+42.555328775" Nov 22 07:57:08 crc kubenswrapper[4789]: I1122 07:57:08.375546 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:08 crc kubenswrapper[4789]: I1122 07:57:08.375736 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:08 crc kubenswrapper[4789]: I1122 07:57:08.375839 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:08 crc kubenswrapper[4789]: I1122 07:57:08.375856 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:08 crc kubenswrapper[4789]: I1122 07:57:08.375869 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:08Z","lastTransitionTime":"2025-11-22T07:57:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:08 crc kubenswrapper[4789]: I1122 07:57:08.479732 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:08 crc kubenswrapper[4789]: I1122 07:57:08.479804 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:08 crc kubenswrapper[4789]: I1122 07:57:08.479823 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:08 crc kubenswrapper[4789]: I1122 07:57:08.479843 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:08 crc kubenswrapper[4789]: I1122 07:57:08.479855 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:08Z","lastTransitionTime":"2025-11-22T07:57:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:08 crc kubenswrapper[4789]: I1122 07:57:08.582486 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:08 crc kubenswrapper[4789]: I1122 07:57:08.582529 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:08 crc kubenswrapper[4789]: I1122 07:57:08.582538 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:08 crc kubenswrapper[4789]: I1122 07:57:08.582554 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:08 crc kubenswrapper[4789]: I1122 07:57:08.582564 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:08Z","lastTransitionTime":"2025-11-22T07:57:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:08 crc kubenswrapper[4789]: I1122 07:57:08.685324 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:08 crc kubenswrapper[4789]: I1122 07:57:08.685365 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:08 crc kubenswrapper[4789]: I1122 07:57:08.685376 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:08 crc kubenswrapper[4789]: I1122 07:57:08.685391 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:08 crc kubenswrapper[4789]: I1122 07:57:08.685402 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:08Z","lastTransitionTime":"2025-11-22T07:57:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:08 crc kubenswrapper[4789]: I1122 07:57:08.788043 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:08 crc kubenswrapper[4789]: I1122 07:57:08.788079 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:08 crc kubenswrapper[4789]: I1122 07:57:08.788088 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:08 crc kubenswrapper[4789]: I1122 07:57:08.788104 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:08 crc kubenswrapper[4789]: I1122 07:57:08.788113 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:08Z","lastTransitionTime":"2025-11-22T07:57:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:08 crc kubenswrapper[4789]: I1122 07:57:08.890259 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:08 crc kubenswrapper[4789]: I1122 07:57:08.890325 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:08 crc kubenswrapper[4789]: I1122 07:57:08.890335 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:08 crc kubenswrapper[4789]: I1122 07:57:08.890352 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:08 crc kubenswrapper[4789]: I1122 07:57:08.890363 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:08Z","lastTransitionTime":"2025-11-22T07:57:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:08 crc kubenswrapper[4789]: I1122 07:57:08.964867 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-q7rc4" Nov 22 07:57:08 crc kubenswrapper[4789]: I1122 07:57:08.964904 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 07:57:08 crc kubenswrapper[4789]: E1122 07:57:08.965398 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-q7rc4" podUID="8adfba92-180a-4f0a-adc0-d1e2bdd62dba" Nov 22 07:57:08 crc kubenswrapper[4789]: E1122 07:57:08.965520 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 07:57:08 crc kubenswrapper[4789]: I1122 07:57:08.964991 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 07:57:08 crc kubenswrapper[4789]: E1122 07:57:08.965665 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 07:57:08 crc kubenswrapper[4789]: I1122 07:57:08.992005 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:08 crc kubenswrapper[4789]: I1122 07:57:08.992049 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:08 crc kubenswrapper[4789]: I1122 07:57:08.992062 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:08 crc kubenswrapper[4789]: I1122 07:57:08.992079 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:08 crc kubenswrapper[4789]: I1122 07:57:08.992090 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:08Z","lastTransitionTime":"2025-11-22T07:57:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:09 crc kubenswrapper[4789]: I1122 07:57:09.095062 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:09 crc kubenswrapper[4789]: I1122 07:57:09.095118 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:09 crc kubenswrapper[4789]: I1122 07:57:09.095130 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:09 crc kubenswrapper[4789]: I1122 07:57:09.095147 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:09 crc kubenswrapper[4789]: I1122 07:57:09.095162 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:09Z","lastTransitionTime":"2025-11-22T07:57:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:09 crc kubenswrapper[4789]: I1122 07:57:09.178960 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8adfba92-180a-4f0a-adc0-d1e2bdd62dba-metrics-certs\") pod \"network-metrics-daemon-q7rc4\" (UID: \"8adfba92-180a-4f0a-adc0-d1e2bdd62dba\") " pod="openshift-multus/network-metrics-daemon-q7rc4" Nov 22 07:57:09 crc kubenswrapper[4789]: E1122 07:57:09.179089 4789 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 22 07:57:09 crc kubenswrapper[4789]: E1122 07:57:09.179150 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8adfba92-180a-4f0a-adc0-d1e2bdd62dba-metrics-certs podName:8adfba92-180a-4f0a-adc0-d1e2bdd62dba nodeName:}" failed. No retries permitted until 2025-11-22 07:57:13.179135208 +0000 UTC m=+47.413535481 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8adfba92-180a-4f0a-adc0-d1e2bdd62dba-metrics-certs") pod "network-metrics-daemon-q7rc4" (UID: "8adfba92-180a-4f0a-adc0-d1e2bdd62dba") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 22 07:57:09 crc kubenswrapper[4789]: I1122 07:57:09.198337 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:09 crc kubenswrapper[4789]: I1122 07:57:09.198386 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:09 crc kubenswrapper[4789]: I1122 07:57:09.198397 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:09 crc kubenswrapper[4789]: I1122 07:57:09.198416 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:09 crc kubenswrapper[4789]: I1122 07:57:09.198428 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:09Z","lastTransitionTime":"2025-11-22T07:57:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:09 crc kubenswrapper[4789]: I1122 07:57:09.206207 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" event={"ID":"0cb4cf51-62d6-4f32-a8b4-4116555f3870","Type":"ContainerStarted","Data":"3740d073f60216d0f7bcd8e638a89cf9f3fe351cd7c21ffaaa67a65a64546c33"} Nov 22 07:57:09 crc kubenswrapper[4789]: I1122 07:57:09.206245 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" event={"ID":"0cb4cf51-62d6-4f32-a8b4-4116555f3870","Type":"ContainerStarted","Data":"f21f329a56796c8ad0319e24f24b6bb5f1d0f34426dd16613b9fe2819b16f69f"} Nov 22 07:57:09 crc kubenswrapper[4789]: I1122 07:57:09.206256 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" event={"ID":"0cb4cf51-62d6-4f32-a8b4-4116555f3870","Type":"ContainerStarted","Data":"a42be6036a735dcc45137b6268cfd915cb077dea0c25fdacc35c454aae4fd3a4"} Nov 22 07:57:09 crc kubenswrapper[4789]: I1122 07:57:09.206266 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" event={"ID":"0cb4cf51-62d6-4f32-a8b4-4116555f3870","Type":"ContainerStarted","Data":"dc46c530762e95475561ad5c2b62c516fd6a204aa8ba1010358a8ad977d02aa1"} Nov 22 07:57:09 crc kubenswrapper[4789]: I1122 07:57:09.206275 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" event={"ID":"0cb4cf51-62d6-4f32-a8b4-4116555f3870","Type":"ContainerStarted","Data":"16f5881811857276dcb1ad168c209a633185f75938f2d03755a8ae41e3b8e528"} Nov 22 07:57:09 crc kubenswrapper[4789]: I1122 07:57:09.206285 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" event={"ID":"0cb4cf51-62d6-4f32-a8b4-4116555f3870","Type":"ContainerStarted","Data":"266632a371608ebd031caa58818577320156f15bcb53b8beb701f785a6b57f5a"} Nov 22 07:57:09 crc kubenswrapper[4789]: I1122 07:57:09.207213 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-vrkrh" event={"ID":"371275b1-1d13-4f3c-8190-8c59c080fef5","Type":"ContainerStarted","Data":"b70d551a08ead82caca33b0019a0a391502fc9202b20820c49cfce57b1c80259"} Nov 22 07:57:09 crc kubenswrapper[4789]: I1122 07:57:09.209600 4789 generic.go:334] "Generic (PLEG): container finished" podID="2fdee89f-559b-4f4e-8d20-6ecc3aafd22e" containerID="529e32af6976e4eb79af9aed268c18d0dafe451236f9ecff83db3aca0278a567" exitCode=0 Nov 22 07:57:09 crc kubenswrapper[4789]: I1122 07:57:09.209637 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2wdp7" event={"ID":"2fdee89f-559b-4f4e-8d20-6ecc3aafd22e","Type":"ContainerDied","Data":"529e32af6976e4eb79af9aed268c18d0dafe451236f9ecff83db3aca0278a567"} Nov 22 07:57:09 crc kubenswrapper[4789]: I1122 07:57:09.300916 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:09 crc kubenswrapper[4789]: I1122 07:57:09.300955 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:09 crc kubenswrapper[4789]: I1122 07:57:09.300966 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:09 crc kubenswrapper[4789]: I1122 07:57:09.301022 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:09 crc kubenswrapper[4789]: I1122 07:57:09.301034 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:09Z","lastTransitionTime":"2025-11-22T07:57:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:09 crc kubenswrapper[4789]: I1122 07:57:09.403741 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:09 crc kubenswrapper[4789]: I1122 07:57:09.403807 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:09 crc kubenswrapper[4789]: I1122 07:57:09.403820 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:09 crc kubenswrapper[4789]: I1122 07:57:09.403860 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:09 crc kubenswrapper[4789]: I1122 07:57:09.403875 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:09Z","lastTransitionTime":"2025-11-22T07:57:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:09 crc kubenswrapper[4789]: I1122 07:57:09.506138 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:09 crc kubenswrapper[4789]: I1122 07:57:09.506176 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:09 crc kubenswrapper[4789]: I1122 07:57:09.506188 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:09 crc kubenswrapper[4789]: I1122 07:57:09.506203 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:09 crc kubenswrapper[4789]: I1122 07:57:09.506215 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:09Z","lastTransitionTime":"2025-11-22T07:57:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:09 crc kubenswrapper[4789]: I1122 07:57:09.608591 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:09 crc kubenswrapper[4789]: I1122 07:57:09.608636 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:09 crc kubenswrapper[4789]: I1122 07:57:09.608648 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:09 crc kubenswrapper[4789]: I1122 07:57:09.608663 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:09 crc kubenswrapper[4789]: I1122 07:57:09.608676 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:09Z","lastTransitionTime":"2025-11-22T07:57:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:09 crc kubenswrapper[4789]: I1122 07:57:09.711402 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:09 crc kubenswrapper[4789]: I1122 07:57:09.711871 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:09 crc kubenswrapper[4789]: I1122 07:57:09.711889 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:09 crc kubenswrapper[4789]: I1122 07:57:09.711910 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:09 crc kubenswrapper[4789]: I1122 07:57:09.711925 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:09Z","lastTransitionTime":"2025-11-22T07:57:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:09 crc kubenswrapper[4789]: I1122 07:57:09.814441 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:09 crc kubenswrapper[4789]: I1122 07:57:09.814482 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:09 crc kubenswrapper[4789]: I1122 07:57:09.814491 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:09 crc kubenswrapper[4789]: I1122 07:57:09.814505 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:09 crc kubenswrapper[4789]: I1122 07:57:09.814514 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:09Z","lastTransitionTime":"2025-11-22T07:57:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:09 crc kubenswrapper[4789]: I1122 07:57:09.916307 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:09 crc kubenswrapper[4789]: I1122 07:57:09.916350 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:09 crc kubenswrapper[4789]: I1122 07:57:09.916360 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:09 crc kubenswrapper[4789]: I1122 07:57:09.916375 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:09 crc kubenswrapper[4789]: I1122 07:57:09.916383 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:09Z","lastTransitionTime":"2025-11-22T07:57:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:09 crc kubenswrapper[4789]: I1122 07:57:09.965208 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 07:57:09 crc kubenswrapper[4789]: E1122 07:57:09.965333 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 07:57:10 crc kubenswrapper[4789]: I1122 07:57:10.019095 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:10 crc kubenswrapper[4789]: I1122 07:57:10.019178 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:10 crc kubenswrapper[4789]: I1122 07:57:10.019205 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:10 crc kubenswrapper[4789]: I1122 07:57:10.019237 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:10 crc kubenswrapper[4789]: I1122 07:57:10.019260 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:10Z","lastTransitionTime":"2025-11-22T07:57:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:10 crc kubenswrapper[4789]: I1122 07:57:10.122248 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:10 crc kubenswrapper[4789]: I1122 07:57:10.122288 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:10 crc kubenswrapper[4789]: I1122 07:57:10.122302 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:10 crc kubenswrapper[4789]: I1122 07:57:10.122320 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:10 crc kubenswrapper[4789]: I1122 07:57:10.122329 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:10Z","lastTransitionTime":"2025-11-22T07:57:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:10 crc kubenswrapper[4789]: I1122 07:57:10.216411 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2wdp7" event={"ID":"2fdee89f-559b-4f4e-8d20-6ecc3aafd22e","Type":"ContainerStarted","Data":"82725c7fd689cdf0722c6d82967386baac0ff54fb45d47eec762b29eb099a0e3"} Nov 22 07:57:10 crc kubenswrapper[4789]: I1122 07:57:10.224858 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:10 crc kubenswrapper[4789]: I1122 07:57:10.225261 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:10 crc kubenswrapper[4789]: I1122 07:57:10.225271 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:10 crc kubenswrapper[4789]: I1122 07:57:10.225286 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:10 crc kubenswrapper[4789]: I1122 07:57:10.225295 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:10Z","lastTransitionTime":"2025-11-22T07:57:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:10 crc kubenswrapper[4789]: I1122 07:57:10.231191 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-vrkrh" podStartSLOduration=20.231172455 podStartE2EDuration="20.231172455s" podCreationTimestamp="2025-11-22 07:56:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 07:57:10.230712633 +0000 UTC m=+44.465112906" watchObservedRunningTime="2025-11-22 07:57:10.231172455 +0000 UTC m=+44.465572768" Nov 22 07:57:10 crc kubenswrapper[4789]: I1122 07:57:10.328171 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:10 crc kubenswrapper[4789]: I1122 07:57:10.328219 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:10 crc kubenswrapper[4789]: I1122 07:57:10.328227 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:10 crc kubenswrapper[4789]: I1122 07:57:10.328241 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:10 crc kubenswrapper[4789]: I1122 07:57:10.328253 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:10Z","lastTransitionTime":"2025-11-22T07:57:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:10 crc kubenswrapper[4789]: I1122 07:57:10.430839 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:10 crc kubenswrapper[4789]: I1122 07:57:10.430889 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:10 crc kubenswrapper[4789]: I1122 07:57:10.430900 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:10 crc kubenswrapper[4789]: I1122 07:57:10.430925 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:10 crc kubenswrapper[4789]: I1122 07:57:10.430936 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:10Z","lastTransitionTime":"2025-11-22T07:57:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:10 crc kubenswrapper[4789]: I1122 07:57:10.533552 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:10 crc kubenswrapper[4789]: I1122 07:57:10.533596 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:10 crc kubenswrapper[4789]: I1122 07:57:10.533608 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:10 crc kubenswrapper[4789]: I1122 07:57:10.533623 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:10 crc kubenswrapper[4789]: I1122 07:57:10.533980 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:10Z","lastTransitionTime":"2025-11-22T07:57:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:10 crc kubenswrapper[4789]: I1122 07:57:10.636560 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:10 crc kubenswrapper[4789]: I1122 07:57:10.636593 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:10 crc kubenswrapper[4789]: I1122 07:57:10.636602 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:10 crc kubenswrapper[4789]: I1122 07:57:10.636618 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:10 crc kubenswrapper[4789]: I1122 07:57:10.636627 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:10Z","lastTransitionTime":"2025-11-22T07:57:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:10 crc kubenswrapper[4789]: I1122 07:57:10.739326 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:10 crc kubenswrapper[4789]: I1122 07:57:10.739400 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:10 crc kubenswrapper[4789]: I1122 07:57:10.739429 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:10 crc kubenswrapper[4789]: I1122 07:57:10.739451 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:10 crc kubenswrapper[4789]: I1122 07:57:10.739465 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:10Z","lastTransitionTime":"2025-11-22T07:57:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:10 crc kubenswrapper[4789]: I1122 07:57:10.905575 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:10 crc kubenswrapper[4789]: I1122 07:57:10.905627 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:10 crc kubenswrapper[4789]: I1122 07:57:10.905659 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:10 crc kubenswrapper[4789]: I1122 07:57:10.905676 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:10 crc kubenswrapper[4789]: I1122 07:57:10.905686 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:10Z","lastTransitionTime":"2025-11-22T07:57:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:10 crc kubenswrapper[4789]: I1122 07:57:10.964962 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 07:57:10 crc kubenswrapper[4789]: I1122 07:57:10.965081 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-q7rc4" Nov 22 07:57:10 crc kubenswrapper[4789]: I1122 07:57:10.965161 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 07:57:10 crc kubenswrapper[4789]: E1122 07:57:10.965282 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 07:57:10 crc kubenswrapper[4789]: E1122 07:57:10.965429 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-q7rc4" podUID="8adfba92-180a-4f0a-adc0-d1e2bdd62dba" Nov 22 07:57:10 crc kubenswrapper[4789]: E1122 07:57:10.965543 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 07:57:11 crc kubenswrapper[4789]: I1122 07:57:11.007878 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:11 crc kubenswrapper[4789]: I1122 07:57:11.007920 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:11 crc kubenswrapper[4789]: I1122 07:57:11.007931 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:11 crc kubenswrapper[4789]: I1122 07:57:11.007949 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:11 crc kubenswrapper[4789]: I1122 07:57:11.007960 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:11Z","lastTransitionTime":"2025-11-22T07:57:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:11 crc kubenswrapper[4789]: I1122 07:57:11.110050 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:11 crc kubenswrapper[4789]: I1122 07:57:11.110089 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:11 crc kubenswrapper[4789]: I1122 07:57:11.110098 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:11 crc kubenswrapper[4789]: I1122 07:57:11.110111 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:11 crc kubenswrapper[4789]: I1122 07:57:11.110120 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:11Z","lastTransitionTime":"2025-11-22T07:57:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:11 crc kubenswrapper[4789]: I1122 07:57:11.212331 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:11 crc kubenswrapper[4789]: I1122 07:57:11.212376 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:11 crc kubenswrapper[4789]: I1122 07:57:11.212387 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:11 crc kubenswrapper[4789]: I1122 07:57:11.212402 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:11 crc kubenswrapper[4789]: I1122 07:57:11.212412 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:11Z","lastTransitionTime":"2025-11-22T07:57:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:11 crc kubenswrapper[4789]: I1122 07:57:11.223655 4789 generic.go:334] "Generic (PLEG): container finished" podID="2fdee89f-559b-4f4e-8d20-6ecc3aafd22e" containerID="82725c7fd689cdf0722c6d82967386baac0ff54fb45d47eec762b29eb099a0e3" exitCode=0 Nov 22 07:57:11 crc kubenswrapper[4789]: I1122 07:57:11.223730 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2wdp7" event={"ID":"2fdee89f-559b-4f4e-8d20-6ecc3aafd22e","Type":"ContainerDied","Data":"82725c7fd689cdf0722c6d82967386baac0ff54fb45d47eec762b29eb099a0e3"} Nov 22 07:57:11 crc kubenswrapper[4789]: I1122 07:57:11.234278 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" event={"ID":"0cb4cf51-62d6-4f32-a8b4-4116555f3870","Type":"ContainerStarted","Data":"5fd3bfe1a04103047c632d783bc804e6866ca3c2f012da7653ad20d277c3a9d0"} Nov 22 07:57:11 crc kubenswrapper[4789]: I1122 07:57:11.315815 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:11 crc kubenswrapper[4789]: I1122 07:57:11.315856 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:11 crc kubenswrapper[4789]: I1122 07:57:11.315869 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:11 crc kubenswrapper[4789]: I1122 07:57:11.315885 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:11 crc kubenswrapper[4789]: I1122 07:57:11.315897 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:11Z","lastTransitionTime":"2025-11-22T07:57:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:11 crc kubenswrapper[4789]: I1122 07:57:11.418660 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:11 crc kubenswrapper[4789]: I1122 07:57:11.418711 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:11 crc kubenswrapper[4789]: I1122 07:57:11.418724 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:11 crc kubenswrapper[4789]: I1122 07:57:11.418740 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:11 crc kubenswrapper[4789]: I1122 07:57:11.418765 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:11Z","lastTransitionTime":"2025-11-22T07:57:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:11 crc kubenswrapper[4789]: I1122 07:57:11.521897 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:11 crc kubenswrapper[4789]: I1122 07:57:11.521931 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:11 crc kubenswrapper[4789]: I1122 07:57:11.521940 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:11 crc kubenswrapper[4789]: I1122 07:57:11.521952 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:11 crc kubenswrapper[4789]: I1122 07:57:11.521961 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:11Z","lastTransitionTime":"2025-11-22T07:57:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:11 crc kubenswrapper[4789]: I1122 07:57:11.625271 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:11 crc kubenswrapper[4789]: I1122 07:57:11.625607 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:11 crc kubenswrapper[4789]: I1122 07:57:11.625619 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:11 crc kubenswrapper[4789]: I1122 07:57:11.625636 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:11 crc kubenswrapper[4789]: I1122 07:57:11.625649 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:11Z","lastTransitionTime":"2025-11-22T07:57:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:11 crc kubenswrapper[4789]: I1122 07:57:11.727937 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:11 crc kubenswrapper[4789]: I1122 07:57:11.727984 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:11 crc kubenswrapper[4789]: I1122 07:57:11.727997 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:11 crc kubenswrapper[4789]: I1122 07:57:11.728017 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:11 crc kubenswrapper[4789]: I1122 07:57:11.728032 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:11Z","lastTransitionTime":"2025-11-22T07:57:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:11 crc kubenswrapper[4789]: I1122 07:57:11.830905 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:11 crc kubenswrapper[4789]: I1122 07:57:11.830958 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:11 crc kubenswrapper[4789]: I1122 07:57:11.830969 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:11 crc kubenswrapper[4789]: I1122 07:57:11.830986 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:11 crc kubenswrapper[4789]: I1122 07:57:11.830999 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:11Z","lastTransitionTime":"2025-11-22T07:57:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:11 crc kubenswrapper[4789]: I1122 07:57:11.933842 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:11 crc kubenswrapper[4789]: I1122 07:57:11.933909 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:11 crc kubenswrapper[4789]: I1122 07:57:11.933925 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:11 crc kubenswrapper[4789]: I1122 07:57:11.933954 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:11 crc kubenswrapper[4789]: I1122 07:57:11.933973 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:11Z","lastTransitionTime":"2025-11-22T07:57:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:11 crc kubenswrapper[4789]: I1122 07:57:11.964878 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 07:57:11 crc kubenswrapper[4789]: E1122 07:57:11.965109 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 07:57:12 crc kubenswrapper[4789]: I1122 07:57:12.036666 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:12 crc kubenswrapper[4789]: I1122 07:57:12.036745 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:12 crc kubenswrapper[4789]: I1122 07:57:12.036821 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:12 crc kubenswrapper[4789]: I1122 07:57:12.036846 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:12 crc kubenswrapper[4789]: I1122 07:57:12.036863 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:12Z","lastTransitionTime":"2025-11-22T07:57:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:12 crc kubenswrapper[4789]: I1122 07:57:12.139430 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:12 crc kubenswrapper[4789]: I1122 07:57:12.139467 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:12 crc kubenswrapper[4789]: I1122 07:57:12.139477 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:12 crc kubenswrapper[4789]: I1122 07:57:12.139499 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:12 crc kubenswrapper[4789]: I1122 07:57:12.139509 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:12Z","lastTransitionTime":"2025-11-22T07:57:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:12 crc kubenswrapper[4789]: I1122 07:57:12.216593 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 07:57:12 crc kubenswrapper[4789]: I1122 07:57:12.216643 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 07:57:12 crc kubenswrapper[4789]: I1122 07:57:12.216651 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 07:57:12 crc kubenswrapper[4789]: I1122 07:57:12.216674 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 07:57:12 crc kubenswrapper[4789]: I1122 07:57:12.216687 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T07:57:12Z","lastTransitionTime":"2025-11-22T07:57:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 07:57:12 crc kubenswrapper[4789]: I1122 07:57:12.243247 4789 generic.go:334] "Generic (PLEG): container finished" podID="2fdee89f-559b-4f4e-8d20-6ecc3aafd22e" containerID="3aa86b0ca4662fc86494ea0a9c8812aebe6d08e8f0289592f039f7acbad35dfe" exitCode=0 Nov 22 07:57:12 crc kubenswrapper[4789]: I1122 07:57:12.243298 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2wdp7" event={"ID":"2fdee89f-559b-4f4e-8d20-6ecc3aafd22e","Type":"ContainerDied","Data":"3aa86b0ca4662fc86494ea0a9c8812aebe6d08e8f0289592f039f7acbad35dfe"} Nov 22 07:57:12 crc kubenswrapper[4789]: I1122 07:57:12.258294 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-vs5fm"] Nov 22 07:57:12 crc kubenswrapper[4789]: I1122 07:57:12.258804 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-vs5fm" Nov 22 07:57:12 crc kubenswrapper[4789]: I1122 07:57:12.260509 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Nov 22 07:57:12 crc kubenswrapper[4789]: I1122 07:57:12.260856 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Nov 22 07:57:12 crc kubenswrapper[4789]: I1122 07:57:12.261509 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Nov 22 07:57:12 crc kubenswrapper[4789]: I1122 07:57:12.261837 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Nov 22 07:57:12 crc kubenswrapper[4789]: I1122 07:57:12.413380 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/262d696f-4681-438c-97a8-159b61f1b412-service-ca\") pod \"cluster-version-operator-5c965bbfc6-vs5fm\" (UID: \"262d696f-4681-438c-97a8-159b61f1b412\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-vs5fm" Nov 22 07:57:12 crc kubenswrapper[4789]: I1122 07:57:12.413432 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/262d696f-4681-438c-97a8-159b61f1b412-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-vs5fm\" (UID: \"262d696f-4681-438c-97a8-159b61f1b412\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-vs5fm" Nov 22 07:57:12 crc kubenswrapper[4789]: I1122 07:57:12.413473 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/262d696f-4681-438c-97a8-159b61f1b412-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-vs5fm\" (UID: \"262d696f-4681-438c-97a8-159b61f1b412\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-vs5fm" Nov 22 07:57:12 crc kubenswrapper[4789]: I1122 07:57:12.413489 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/262d696f-4681-438c-97a8-159b61f1b412-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-vs5fm\" (UID: \"262d696f-4681-438c-97a8-159b61f1b412\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-vs5fm" Nov 22 07:57:12 crc kubenswrapper[4789]: I1122 07:57:12.413526 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/262d696f-4681-438c-97a8-159b61f1b412-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-vs5fm\" (UID: \"262d696f-4681-438c-97a8-159b61f1b412\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-vs5fm" Nov 22 07:57:12 crc kubenswrapper[4789]: I1122 07:57:12.515302 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/262d696f-4681-438c-97a8-159b61f1b412-service-ca\") pod \"cluster-version-operator-5c965bbfc6-vs5fm\" (UID: \"262d696f-4681-438c-97a8-159b61f1b412\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-vs5fm" Nov 22 07:57:12 crc kubenswrapper[4789]: I1122 07:57:12.515362 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/262d696f-4681-438c-97a8-159b61f1b412-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-vs5fm\" (UID: \"262d696f-4681-438c-97a8-159b61f1b412\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-vs5fm" Nov 22 07:57:12 crc kubenswrapper[4789]: I1122 07:57:12.515415 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/262d696f-4681-438c-97a8-159b61f1b412-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-vs5fm\" (UID: \"262d696f-4681-438c-97a8-159b61f1b412\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-vs5fm" Nov 22 07:57:12 crc kubenswrapper[4789]: I1122 07:57:12.515428 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/262d696f-4681-438c-97a8-159b61f1b412-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-vs5fm\" (UID: \"262d696f-4681-438c-97a8-159b61f1b412\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-vs5fm" Nov 22 07:57:12 crc kubenswrapper[4789]: I1122 07:57:12.515461 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/262d696f-4681-438c-97a8-159b61f1b412-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-vs5fm\" (UID: \"262d696f-4681-438c-97a8-159b61f1b412\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-vs5fm" Nov 22 07:57:12 crc kubenswrapper[4789]: I1122 07:57:12.515484 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/262d696f-4681-438c-97a8-159b61f1b412-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-vs5fm\" (UID: \"262d696f-4681-438c-97a8-159b61f1b412\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-vs5fm" Nov 22 07:57:12 crc kubenswrapper[4789]: I1122 07:57:12.515520 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/262d696f-4681-438c-97a8-159b61f1b412-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-vs5fm\" (UID: \"262d696f-4681-438c-97a8-159b61f1b412\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-vs5fm" Nov 22 07:57:12 crc kubenswrapper[4789]: I1122 07:57:12.517300 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/262d696f-4681-438c-97a8-159b61f1b412-service-ca\") pod \"cluster-version-operator-5c965bbfc6-vs5fm\" (UID: \"262d696f-4681-438c-97a8-159b61f1b412\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-vs5fm" Nov 22 07:57:12 crc kubenswrapper[4789]: I1122 07:57:12.522733 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/262d696f-4681-438c-97a8-159b61f1b412-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-vs5fm\" (UID: \"262d696f-4681-438c-97a8-159b61f1b412\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-vs5fm" Nov 22 07:57:12 crc kubenswrapper[4789]: I1122 07:57:12.535261 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/262d696f-4681-438c-97a8-159b61f1b412-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-vs5fm\" (UID: \"262d696f-4681-438c-97a8-159b61f1b412\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-vs5fm" Nov 22 07:57:12 crc kubenswrapper[4789]: I1122 07:57:12.571682 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-vs5fm" Nov 22 07:57:12 crc kubenswrapper[4789]: W1122 07:57:12.595478 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod262d696f_4681_438c_97a8_159b61f1b412.slice/crio-8c1e2b5b7fee7200e709afbe982e0d930c8f196c263ffd3684f8b9d84bf0cabb WatchSource:0}: Error finding container 8c1e2b5b7fee7200e709afbe982e0d930c8f196c263ffd3684f8b9d84bf0cabb: Status 404 returned error can't find the container with id 8c1e2b5b7fee7200e709afbe982e0d930c8f196c263ffd3684f8b9d84bf0cabb Nov 22 07:57:12 crc kubenswrapper[4789]: I1122 07:57:12.964127 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-q7rc4" Nov 22 07:57:12 crc kubenswrapper[4789]: I1122 07:57:12.964185 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 07:57:12 crc kubenswrapper[4789]: I1122 07:57:12.964128 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 07:57:12 crc kubenswrapper[4789]: E1122 07:57:12.964258 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-q7rc4" podUID="8adfba92-180a-4f0a-adc0-d1e2bdd62dba" Nov 22 07:57:12 crc kubenswrapper[4789]: E1122 07:57:12.964877 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 07:57:12 crc kubenswrapper[4789]: E1122 07:57:12.964961 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 07:57:13 crc kubenswrapper[4789]: I1122 07:57:13.221384 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8adfba92-180a-4f0a-adc0-d1e2bdd62dba-metrics-certs\") pod \"network-metrics-daemon-q7rc4\" (UID: \"8adfba92-180a-4f0a-adc0-d1e2bdd62dba\") " pod="openshift-multus/network-metrics-daemon-q7rc4" Nov 22 07:57:13 crc kubenswrapper[4789]: E1122 07:57:13.221660 4789 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 22 07:57:13 crc kubenswrapper[4789]: E1122 07:57:13.222013 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8adfba92-180a-4f0a-adc0-d1e2bdd62dba-metrics-certs podName:8adfba92-180a-4f0a-adc0-d1e2bdd62dba nodeName:}" failed. No retries permitted until 2025-11-22 07:57:21.221995364 +0000 UTC m=+55.456395637 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8adfba92-180a-4f0a-adc0-d1e2bdd62dba-metrics-certs") pod "network-metrics-daemon-q7rc4" (UID: "8adfba92-180a-4f0a-adc0-d1e2bdd62dba") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 22 07:57:13 crc kubenswrapper[4789]: I1122 07:57:13.250385 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2wdp7" event={"ID":"2fdee89f-559b-4f4e-8d20-6ecc3aafd22e","Type":"ContainerStarted","Data":"0ce61841f0ebdf9564a46631cab5de5e6e41937a5e6994aed98ca78166c65f41"} Nov 22 07:57:13 crc kubenswrapper[4789]: I1122 07:57:13.252186 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-vs5fm" event={"ID":"262d696f-4681-438c-97a8-159b61f1b412","Type":"ContainerStarted","Data":"34981a7dd72cf93fad1ae59a427f39751631e8b81cf8e7c49015bfc73c3bb21f"} Nov 22 07:57:13 crc kubenswrapper[4789]: I1122 07:57:13.252219 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-vs5fm" event={"ID":"262d696f-4681-438c-97a8-159b61f1b412","Type":"ContainerStarted","Data":"8c1e2b5b7fee7200e709afbe982e0d930c8f196c263ffd3684f8b9d84bf0cabb"} Nov 22 07:57:13 crc kubenswrapper[4789]: I1122 07:57:13.284267 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-2wdp7" podStartSLOduration=23.284246093 podStartE2EDuration="23.284246093s" podCreationTimestamp="2025-11-22 07:56:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 07:57:13.271985583 +0000 UTC m=+47.506385896" watchObservedRunningTime="2025-11-22 07:57:13.284246093 +0000 UTC m=+47.518646366" Nov 22 07:57:13 crc kubenswrapper[4789]: I1122 07:57:13.284944 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-vs5fm" podStartSLOduration=23.284940222 podStartE2EDuration="23.284940222s" podCreationTimestamp="2025-11-22 07:56:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 07:57:13.284539491 +0000 UTC m=+47.518939764" watchObservedRunningTime="2025-11-22 07:57:13.284940222 +0000 UTC m=+47.519340495" Nov 22 07:57:13 crc kubenswrapper[4789]: I1122 07:57:13.964518 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 07:57:13 crc kubenswrapper[4789]: E1122 07:57:13.964658 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 07:57:14 crc kubenswrapper[4789]: I1122 07:57:14.262607 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" event={"ID":"0cb4cf51-62d6-4f32-a8b4-4116555f3870","Type":"ContainerStarted","Data":"c99a248e9a7e8df9e1be7cc7742476be75b4344274665da471fbf3022a343f28"} Nov 22 07:57:14 crc kubenswrapper[4789]: I1122 07:57:14.298274 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" podStartSLOduration=23.298249033 podStartE2EDuration="23.298249033s" podCreationTimestamp="2025-11-22 07:56:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 07:57:14.298237843 +0000 UTC m=+48.532638126" watchObservedRunningTime="2025-11-22 07:57:14.298249033 +0000 UTC m=+48.532649346" Nov 22 07:57:14 crc kubenswrapper[4789]: I1122 07:57:14.964893 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 07:57:14 crc kubenswrapper[4789]: I1122 07:57:14.965319 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 07:57:14 crc kubenswrapper[4789]: I1122 07:57:14.965325 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-q7rc4" Nov 22 07:57:14 crc kubenswrapper[4789]: E1122 07:57:14.965520 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 07:57:14 crc kubenswrapper[4789]: E1122 07:57:14.966402 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-q7rc4" podUID="8adfba92-180a-4f0a-adc0-d1e2bdd62dba" Nov 22 07:57:14 crc kubenswrapper[4789]: E1122 07:57:14.966536 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 07:57:15 crc kubenswrapper[4789]: I1122 07:57:15.266197 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" Nov 22 07:57:15 crc kubenswrapper[4789]: I1122 07:57:15.266247 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" Nov 22 07:57:15 crc kubenswrapper[4789]: I1122 07:57:15.266262 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" Nov 22 07:57:15 crc kubenswrapper[4789]: I1122 07:57:15.319353 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" Nov 22 07:57:15 crc kubenswrapper[4789]: I1122 07:57:15.319645 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" Nov 22 07:57:15 crc kubenswrapper[4789]: I1122 07:57:15.956258 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-q7rc4"] Nov 22 07:57:15 crc kubenswrapper[4789]: I1122 07:57:15.956348 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-q7rc4" Nov 22 07:57:15 crc kubenswrapper[4789]: E1122 07:57:15.956432 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-q7rc4" podUID="8adfba92-180a-4f0a-adc0-d1e2bdd62dba" Nov 22 07:57:15 crc kubenswrapper[4789]: I1122 07:57:15.964204 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 07:57:15 crc kubenswrapper[4789]: E1122 07:57:15.964387 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 07:57:16 crc kubenswrapper[4789]: I1122 07:57:16.965189 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 07:57:16 crc kubenswrapper[4789]: I1122 07:57:16.965208 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 07:57:16 crc kubenswrapper[4789]: E1122 07:57:16.965428 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 07:57:16 crc kubenswrapper[4789]: E1122 07:57:16.965474 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 07:57:17 crc kubenswrapper[4789]: I1122 07:57:17.726944 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 22 07:57:17 crc kubenswrapper[4789]: I1122 07:57:17.964388 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-q7rc4" Nov 22 07:57:17 crc kubenswrapper[4789]: I1122 07:57:17.964477 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 07:57:17 crc kubenswrapper[4789]: E1122 07:57:17.965133 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-q7rc4" podUID="8adfba92-180a-4f0a-adc0-d1e2bdd62dba" Nov 22 07:57:17 crc kubenswrapper[4789]: E1122 07:57:17.965208 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.726988 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.727164 4789 kubelet_node_status.go:538] "Fast updating node status as it just became ready" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.758532 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6jw6b"] Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.759133 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6jw6b" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.759157 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-t5d22"] Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.759497 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-t5d22" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.760715 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-v4wls"] Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.761342 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-v4wls" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.761459 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wjmrv"] Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.762191 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wjmrv" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.762261 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-qwhkx"] Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.762595 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-qwhkx" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.763270 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-fpw88"] Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.763616 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fpw88" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.769917 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.770165 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Nov 22 07:57:18 crc kubenswrapper[4789]: W1122 07:57:18.770177 4789 reflector.go:561] object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw": failed to list *v1.Secret: secrets "openshift-controller-manager-operator-dockercfg-vw8fw" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-controller-manager-operator": no relationship found between node 'crc' and this object Nov 22 07:57:18 crc kubenswrapper[4789]: E1122 07:57:18.770216 4789 reflector.go:158] "Unhandled Error" err="object-\"openshift-controller-manager-operator\"/\"openshift-controller-manager-operator-dockercfg-vw8fw\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"openshift-controller-manager-operator-dockercfg-vw8fw\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-controller-manager-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.770252 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-tdpht"] Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.770506 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.770720 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-tdpht" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.770900 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.770897 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.770963 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.771132 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.771283 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.771734 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.771941 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.772121 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-9lzll"] Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.772190 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.772377 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.772490 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.772522 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.772558 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-9lzll" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.772572 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.772646 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.772669 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.772714 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.772804 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 22 07:57:18 crc kubenswrapper[4789]: W1122 07:57:18.772851 4789 reflector.go:561] object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert": failed to list *v1.Secret: secrets "openshift-controller-manager-operator-serving-cert" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-controller-manager-operator": no relationship found between node 'crc' and this object Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.772873 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.772804 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 22 07:57:18 crc kubenswrapper[4789]: E1122 07:57:18.772903 4789 reflector.go:158] "Unhandled Error" err="object-\"openshift-controller-manager-operator\"/\"openshift-controller-manager-operator-serving-cert\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"openshift-controller-manager-operator-serving-cert\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-controller-manager-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.772989 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.773046 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.773113 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.773152 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.773157 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.776963 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.778347 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-ksdn9"] Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.778824 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-ksdn9" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.779013 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-pb98x"] Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.779367 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-pb98x" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.780357 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ckmwl"] Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.780598 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ckmwl" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.784071 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.786899 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.788156 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.788629 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.791080 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.796930 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.797045 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.797096 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.797153 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.797276 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.797806 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.799080 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.799962 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.800208 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-4s4sh"] Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.800357 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.800408 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.801182 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.813053 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.813203 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.813795 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-4s4sh" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.814478 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.815282 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-wnfzj"] Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.815845 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-wnfzj" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.816813 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.816812 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.817162 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-bsrcz"] Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.817593 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.817819 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.818137 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-bsrcz" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.821797 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-frls8"] Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.822579 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-frls8" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.828065 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.832000 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.832206 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.832290 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.834448 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-kdt8l"] Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.834819 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.834966 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-vh974"] Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.835242 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.835371 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.835484 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.835593 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-vh974" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.835978 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-kdt8l" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.839997 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.844136 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.844382 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.844543 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.844985 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.845227 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.845264 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.845359 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.845424 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.845460 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.845542 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.845562 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.845678 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.845689 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.845796 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.845826 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.845829 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.845957 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.846001 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.846063 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.846161 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.845907 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.846327 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.846249 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.854256 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.855860 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.856184 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-4fbj7"] Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.857503 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396625-nrhcv"] Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.858258 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396625-nrhcv" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.858556 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4fbj7" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.876129 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-spnzw"] Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.876869 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-spnzw" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.880469 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.880571 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.880614 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.881925 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.883060 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-tnmrh"] Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.883495 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.883645 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.883729 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.883733 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-fb49f"] Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.883830 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-tnmrh" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.884539 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-fb49f" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.885046 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.885382 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.885688 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.885833 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.885917 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.886391 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0fb11c40-1e46-42a7-b384-626786bc4e3e-config\") pod \"openshift-apiserver-operator-796bbdcf4f-wjmrv\" (UID: \"0fb11c40-1e46-42a7-b384-626786bc4e3e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wjmrv" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.886449 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/79e3d4d9-ec8e-4578-871c-a336ad22a2ac-config\") pod \"console-operator-58897d9998-9lzll\" (UID: \"79e3d4d9-ec8e-4578-871c-a336ad22a2ac\") " pod="openshift-console-operator/console-operator-58897d9998-9lzll" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.886482 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/ff518a19-6241-4c65-b2a0-bee6ab939a6c-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-6jw6b\" (UID: \"ff518a19-6241-4c65-b2a0-bee6ab939a6c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6jw6b" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.886509 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ca00755-c6ae-42d2-833e-fc31b6a480df-config\") pod \"authentication-operator-69f744f599-ksdn9\" (UID: \"1ca00755-c6ae-42d2-833e-fc31b6a480df\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ksdn9" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.886529 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xh6c5\" (UniqueName: \"kubernetes.io/projected/9b861750-2517-4cf3-9c78-930480711af0-kube-api-access-xh6c5\") pod \"console-f9d7485db-pb98x\" (UID: \"9b861750-2517-4cf3-9c78-930480711af0\") " pod="openshift-console/console-f9d7485db-pb98x" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.886569 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sp9jx\" (UniqueName: \"kubernetes.io/projected/be26df1a-a3b3-411b-9e84-df356f3a7af3-kube-api-access-sp9jx\") pod \"machine-approver-56656f9798-v4wls\" (UID: \"be26df1a-a3b3-411b-9e84-df356f3a7af3\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-v4wls" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.886591 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ce67cd3e-c2d2-4562-bc26-0a68bce1a1f0-serving-cert\") pod \"openshift-config-operator-7777fb866f-qwhkx\" (UID: \"ce67cd3e-c2d2-4562-bc26-0a68bce1a1f0\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-qwhkx" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.886611 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/ce67cd3e-c2d2-4562-bc26-0a68bce1a1f0-available-featuregates\") pod \"openshift-config-operator-7777fb866f-qwhkx\" (UID: \"ce67cd3e-c2d2-4562-bc26-0a68bce1a1f0\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-qwhkx" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.886634 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/23967110-3d55-4767-9fa9-230138c92e42-config\") pod \"controller-manager-879f6c89f-4s4sh\" (UID: \"23967110-3d55-4767-9fa9-230138c92e42\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4s4sh" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.886662 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/23967110-3d55-4767-9fa9-230138c92e42-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-4s4sh\" (UID: \"23967110-3d55-4767-9fa9-230138c92e42\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4s4sh" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.886683 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ffef360f-8678-442d-9b27-b8de7646f1a3-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-t5d22\" (UID: \"ffef360f-8678-442d-9b27-b8de7646f1a3\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-t5d22" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.886703 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1ca00755-c6ae-42d2-833e-fc31b6a480df-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-ksdn9\" (UID: \"1ca00755-c6ae-42d2-833e-fc31b6a480df\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ksdn9" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.886716 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0fb11c40-1e46-42a7-b384-626786bc4e3e-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-wjmrv\" (UID: \"0fb11c40-1e46-42a7-b384-626786bc4e3e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wjmrv" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.886732 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5btzg\" (UniqueName: \"kubernetes.io/projected/ff518a19-6241-4c65-b2a0-bee6ab939a6c-kube-api-access-5btzg\") pod \"cluster-samples-operator-665b6dd947-6jw6b\" (UID: \"ff518a19-6241-4c65-b2a0-bee6ab939a6c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6jw6b" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.886762 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d4d9ab07-53b6-4aef-bcac-c0140d870370-serving-cert\") pod \"route-controller-manager-6576b87f9c-fpw88\" (UID: \"d4d9ab07-53b6-4aef-bcac-c0140d870370\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fpw88" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.886776 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/23967110-3d55-4767-9fa9-230138c92e42-serving-cert\") pod \"controller-manager-879f6c89f-4s4sh\" (UID: \"23967110-3d55-4767-9fa9-230138c92e42\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4s4sh" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.886789 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1ca00755-c6ae-42d2-833e-fc31b6a480df-service-ca-bundle\") pod \"authentication-operator-69f744f599-ksdn9\" (UID: \"1ca00755-c6ae-42d2-833e-fc31b6a480df\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ksdn9" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.886804 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/79e3d4d9-ec8e-4578-871c-a336ad22a2ac-trusted-ca\") pod \"console-operator-58897d9998-9lzll\" (UID: \"79e3d4d9-ec8e-4578-871c-a336ad22a2ac\") " pod="openshift-console-operator/console-operator-58897d9998-9lzll" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.886818 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/be26df1a-a3b3-411b-9e84-df356f3a7af3-config\") pod \"machine-approver-56656f9798-v4wls\" (UID: \"be26df1a-a3b3-411b-9e84-df356f3a7af3\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-v4wls" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.886844 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9b861750-2517-4cf3-9c78-930480711af0-service-ca\") pod \"console-f9d7485db-pb98x\" (UID: \"9b861750-2517-4cf3-9c78-930480711af0\") " pod="openshift-console/console-f9d7485db-pb98x" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.886858 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1ca00755-c6ae-42d2-833e-fc31b6a480df-serving-cert\") pod \"authentication-operator-69f744f599-ksdn9\" (UID: \"1ca00755-c6ae-42d2-833e-fc31b6a480df\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ksdn9" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.886874 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qsdgk\" (UniqueName: \"kubernetes.io/projected/79e3d4d9-ec8e-4578-871c-a336ad22a2ac-kube-api-access-qsdgk\") pod \"console-operator-58897d9998-9lzll\" (UID: \"79e3d4d9-ec8e-4578-871c-a336ad22a2ac\") " pod="openshift-console-operator/console-operator-58897d9998-9lzll" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.886888 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ffef360f-8678-442d-9b27-b8de7646f1a3-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-t5d22\" (UID: \"ffef360f-8678-442d-9b27-b8de7646f1a3\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-t5d22" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.886901 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/23967110-3d55-4767-9fa9-230138c92e42-client-ca\") pod \"controller-manager-879f6c89f-4s4sh\" (UID: \"23967110-3d55-4767-9fa9-230138c92e42\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4s4sh" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.886914 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9b861750-2517-4cf3-9c78-930480711af0-trusted-ca-bundle\") pod \"console-f9d7485db-pb98x\" (UID: \"9b861750-2517-4cf3-9c78-930480711af0\") " pod="openshift-console/console-f9d7485db-pb98x" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.886930 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bpjnq\" (UniqueName: \"kubernetes.io/projected/d4d9ab07-53b6-4aef-bcac-c0140d870370-kube-api-access-bpjnq\") pod \"route-controller-manager-6576b87f9c-fpw88\" (UID: \"d4d9ab07-53b6-4aef-bcac-c0140d870370\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fpw88" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.886948 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bb5dq\" (UniqueName: \"kubernetes.io/projected/23967110-3d55-4767-9fa9-230138c92e42-kube-api-access-bb5dq\") pod \"controller-manager-879f6c89f-4s4sh\" (UID: \"23967110-3d55-4767-9fa9-230138c92e42\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4s4sh" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.886963 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/79e3d4d9-ec8e-4578-871c-a336ad22a2ac-serving-cert\") pod \"console-operator-58897d9998-9lzll\" (UID: \"79e3d4d9-ec8e-4578-871c-a336ad22a2ac\") " pod="openshift-console-operator/console-operator-58897d9998-9lzll" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.886978 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d4d9ab07-53b6-4aef-bcac-c0140d870370-client-ca\") pod \"route-controller-manager-6576b87f9c-fpw88\" (UID: \"d4d9ab07-53b6-4aef-bcac-c0140d870370\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fpw88" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.886995 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/be26df1a-a3b3-411b-9e84-df356f3a7af3-auth-proxy-config\") pod \"machine-approver-56656f9798-v4wls\" (UID: \"be26df1a-a3b3-411b-9e84-df356f3a7af3\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-v4wls" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.887010 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/9b861750-2517-4cf3-9c78-930480711af0-console-serving-cert\") pod \"console-f9d7485db-pb98x\" (UID: \"9b861750-2517-4cf3-9c78-930480711af0\") " pod="openshift-console/console-f9d7485db-pb98x" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.887023 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/9b861750-2517-4cf3-9c78-930480711af0-oauth-serving-cert\") pod \"console-f9d7485db-pb98x\" (UID: \"9b861750-2517-4cf3-9c78-930480711af0\") " pod="openshift-console/console-f9d7485db-pb98x" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.887036 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/be26df1a-a3b3-411b-9e84-df356f3a7af3-machine-approver-tls\") pod \"machine-approver-56656f9798-v4wls\" (UID: \"be26df1a-a3b3-411b-9e84-df356f3a7af3\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-v4wls" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.887053 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d4d9ab07-53b6-4aef-bcac-c0140d870370-config\") pod \"route-controller-manager-6576b87f9c-fpw88\" (UID: \"d4d9ab07-53b6-4aef-bcac-c0140d870370\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fpw88" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.887067 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kb79q\" (UniqueName: \"kubernetes.io/projected/ce67cd3e-c2d2-4562-bc26-0a68bce1a1f0-kube-api-access-kb79q\") pod \"openshift-config-operator-7777fb866f-qwhkx\" (UID: \"ce67cd3e-c2d2-4562-bc26-0a68bce1a1f0\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-qwhkx" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.887088 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/9b861750-2517-4cf3-9c78-930480711af0-console-oauth-config\") pod \"console-f9d7485db-pb98x\" (UID: \"9b861750-2517-4cf3-9c78-930480711af0\") " pod="openshift-console/console-f9d7485db-pb98x" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.887107 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/9b861750-2517-4cf3-9c78-930480711af0-console-config\") pod \"console-f9d7485db-pb98x\" (UID: \"9b861750-2517-4cf3-9c78-930480711af0\") " pod="openshift-console/console-f9d7485db-pb98x" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.887129 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tsm8d\" (UniqueName: \"kubernetes.io/projected/c0beaf3d-a9eb-47cc-bb53-392e75959282-kube-api-access-tsm8d\") pod \"cluster-image-registry-operator-dc59b4c8b-ckmwl\" (UID: \"c0beaf3d-a9eb-47cc-bb53-392e75959282\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ckmwl" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.887153 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h8d7k\" (UniqueName: \"kubernetes.io/projected/0fb11c40-1e46-42a7-b384-626786bc4e3e-kube-api-access-h8d7k\") pod \"openshift-apiserver-operator-796bbdcf4f-wjmrv\" (UID: \"0fb11c40-1e46-42a7-b384-626786bc4e3e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wjmrv" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.887173 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c0beaf3d-a9eb-47cc-bb53-392e75959282-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-ckmwl\" (UID: \"c0beaf3d-a9eb-47cc-bb53-392e75959282\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ckmwl" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.887187 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rv7fq\" (UniqueName: \"kubernetes.io/projected/ffef360f-8678-442d-9b27-b8de7646f1a3-kube-api-access-rv7fq\") pod \"openshift-controller-manager-operator-756b6f6bc6-t5d22\" (UID: \"ffef360f-8678-442d-9b27-b8de7646f1a3\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-t5d22" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.887206 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c0beaf3d-a9eb-47cc-bb53-392e75959282-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-ckmwl\" (UID: \"c0beaf3d-a9eb-47cc-bb53-392e75959282\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ckmwl" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.887230 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7dqck\" (UniqueName: \"kubernetes.io/projected/83ac1d78-cc4c-48b7-b975-2f88802e2d17-kube-api-access-7dqck\") pod \"downloads-7954f5f757-tdpht\" (UID: \"83ac1d78-cc4c-48b7-b975-2f88802e2d17\") " pod="openshift-console/downloads-7954f5f757-tdpht" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.887229 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-fg7wj"] Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.887249 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/c0beaf3d-a9eb-47cc-bb53-392e75959282-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-ckmwl\" (UID: \"c0beaf3d-a9eb-47cc-bb53-392e75959282\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ckmwl" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.887275 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hmtsb\" (UniqueName: \"kubernetes.io/projected/1ca00755-c6ae-42d2-833e-fc31b6a480df-kube-api-access-hmtsb\") pod \"authentication-operator-69f744f599-ksdn9\" (UID: \"1ca00755-c6ae-42d2-833e-fc31b6a480df\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ksdn9" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.888042 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-fg7wj" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.888399 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2m4cb"] Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.888831 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2m4cb" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.890662 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.892614 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.894127 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9n9mc"] Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.895096 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.895417 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9n9mc" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.904680 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.904822 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.904714 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.906374 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-h6f5v"] Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.906981 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-h6f5v" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.908796 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-x5b9n"] Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.913157 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.913976 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-x5b9n" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.921152 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-zl87n"] Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.923111 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.924621 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-xlf6m"] Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.929293 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-hszd6"] Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.929934 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-xlf6m" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.930022 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zl87n" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.931453 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-hszd6" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.931987 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-d4l6q"] Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.932841 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-d4l6q" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.946093 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-77svl"] Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.946103 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.947087 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-77svl" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.949835 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-fr4zx"] Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.950554 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-fr4zx" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.950787 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-l4q7j"] Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.951499 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.952036 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2dxsq"] Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.952730 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2dxsq" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.954859 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-sszl6"] Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.955786 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-sszl6" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.956511 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-k5c2k"] Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.957598 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-k5c2k" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.959826 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-t5d22"] Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.961714 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-whfpb"] Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.963180 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-whfpb" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.963541 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wjmrv"] Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.964134 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.964472 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.964597 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/cni-sysctl-allowlist-ds-gcnts"] Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.965103 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/cni-sysctl-allowlist-ds-gcnts" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.966649 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-fpw88"] Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.967511 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6jw6b"] Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.970468 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ckmwl"] Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.971689 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-9lzll"] Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.972637 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-tdpht"] Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.973829 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-qwhkx"] Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.977011 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-kdt8l"] Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.978412 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-frls8"] Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.980490 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-qh9np"] Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.981288 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-qh9np" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.981633 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-qc925"] Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.981816 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.982623 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-qc925" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.987257 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-n27w4"] Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.987966 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/23967110-3d55-4767-9fa9-230138c92e42-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-4s4sh\" (UID: \"23967110-3d55-4767-9fa9-230138c92e42\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4s4sh" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.988020 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd618778-dc82-4797-b737-14231eefea15-config\") pod \"service-ca-operator-777779d784-fg7wj\" (UID: \"dd618778-dc82-4797-b737-14231eefea15\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-fg7wj" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.988051 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/3cfa3a64-3e7a-4336-9ae0-86822095a1b4-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-fb49f\" (UID: \"3cfa3a64-3e7a-4336-9ae0-86822095a1b4\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-fb49f" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.988082 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0fb11c40-1e46-42a7-b384-626786bc4e3e-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-wjmrv\" (UID: \"0fb11c40-1e46-42a7-b384-626786bc4e3e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wjmrv" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.988113 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/d086acbe-d982-4880-9538-04a184cb4148-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-spnzw\" (UID: \"d086acbe-d982-4880-9538-04a184cb4148\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-spnzw" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.988136 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/7dc5fba4-2ba8-4fa2-a76f-0ad8d1196838-etcd-client\") pod \"apiserver-7bbb656c7d-frls8\" (UID: \"7dc5fba4-2ba8-4fa2-a76f-0ad8d1196838\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-frls8" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.988192 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/55cac199-8203-40b3-8a4f-c571017d998b-audit-policies\") pod \"oauth-openshift-558db77b4-wnfzj\" (UID: \"55cac199-8203-40b3-8a4f-c571017d998b\") " pod="openshift-authentication/oauth-openshift-558db77b4-wnfzj" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.988373 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/55cac199-8203-40b3-8a4f-c571017d998b-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-wnfzj\" (UID: \"55cac199-8203-40b3-8a4f-c571017d998b\") " pod="openshift-authentication/oauth-openshift-558db77b4-wnfzj" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.988456 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/55cac199-8203-40b3-8a4f-c571017d998b-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-wnfzj\" (UID: \"55cac199-8203-40b3-8a4f-c571017d998b\") " pod="openshift-authentication/oauth-openshift-558db77b4-wnfzj" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.988485 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7dc5fba4-2ba8-4fa2-a76f-0ad8d1196838-audit-dir\") pod \"apiserver-7bbb656c7d-frls8\" (UID: \"7dc5fba4-2ba8-4fa2-a76f-0ad8d1196838\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-frls8" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.988505 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/55cac199-8203-40b3-8a4f-c571017d998b-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-wnfzj\" (UID: \"55cac199-8203-40b3-8a4f-c571017d998b\") " pod="openshift-authentication/oauth-openshift-558db77b4-wnfzj" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.988525 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/23967110-3d55-4767-9fa9-230138c92e42-serving-cert\") pod \"controller-manager-879f6c89f-4s4sh\" (UID: \"23967110-3d55-4767-9fa9-230138c92e42\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4s4sh" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.988542 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/79e3d4d9-ec8e-4578-871c-a336ad22a2ac-trusted-ca\") pod \"console-operator-58897d9998-9lzll\" (UID: \"79e3d4d9-ec8e-4578-871c-a336ad22a2ac\") " pod="openshift-console-operator/console-operator-58897d9998-9lzll" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.988571 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/352693a0-531b-4017-9582-5424c4691795-service-ca-bundle\") pod \"router-default-5444994796-x5b9n\" (UID: \"352693a0-531b-4017-9582-5424c4691795\") " pod="openshift-ingress/router-default-5444994796-x5b9n" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.988632 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/94294737-067b-4a4f-b59b-f9f987fa9127-audit\") pod \"apiserver-76f77b778f-vh974\" (UID: \"94294737-067b-4a4f-b59b-f9f987fa9127\") " pod="openshift-apiserver/apiserver-76f77b778f-vh974" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.988696 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/be26df1a-a3b3-411b-9e84-df356f3a7af3-config\") pod \"machine-approver-56656f9798-v4wls\" (UID: \"be26df1a-a3b3-411b-9e84-df356f3a7af3\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-v4wls" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.988798 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/55cac199-8203-40b3-8a4f-c571017d998b-audit-dir\") pod \"oauth-openshift-558db77b4-wnfzj\" (UID: \"55cac199-8203-40b3-8a4f-c571017d998b\") " pod="openshift-authentication/oauth-openshift-558db77b4-wnfzj" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.988823 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/dfa022a0-f86e-4f4b-99f5-22b84e9d6558-profile-collector-cert\") pod \"olm-operator-6b444d44fb-9n9mc\" (UID: \"dfa022a0-f86e-4f4b-99f5-22b84e9d6558\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9n9mc" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.988846 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/55cac199-8203-40b3-8a4f-c571017d998b-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-wnfzj\" (UID: \"55cac199-8203-40b3-8a4f-c571017d998b\") " pod="openshift-authentication/oauth-openshift-558db77b4-wnfzj" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.988867 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/cb9c37ae-6151-452f-b467-6b7d14bbb2f8-profile-collector-cert\") pod \"catalog-operator-68c6474976-h6f5v\" (UID: \"cb9c37ae-6151-452f-b467-6b7d14bbb2f8\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-h6f5v" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.989228 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d32dc4fd-3e62-402c-a1d9-8f2c83f2e334-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-tnmrh\" (UID: \"d32dc4fd-3e62-402c-a1d9-8f2c83f2e334\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-tnmrh" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.989256 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/94294737-067b-4a4f-b59b-f9f987fa9127-serving-cert\") pod \"apiserver-76f77b778f-vh974\" (UID: \"94294737-067b-4a4f-b59b-f9f987fa9127\") " pod="openshift-apiserver/apiserver-76f77b778f-vh974" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.989284 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7dc5fba4-2ba8-4fa2-a76f-0ad8d1196838-audit-policies\") pod \"apiserver-7bbb656c7d-frls8\" (UID: \"7dc5fba4-2ba8-4fa2-a76f-0ad8d1196838\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-frls8" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.989344 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bpjnq\" (UniqueName: \"kubernetes.io/projected/d4d9ab07-53b6-4aef-bcac-c0140d870370-kube-api-access-bpjnq\") pod \"route-controller-manager-6576b87f9c-fpw88\" (UID: \"d4d9ab07-53b6-4aef-bcac-c0140d870370\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fpw88" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.989365 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bb5dq\" (UniqueName: \"kubernetes.io/projected/23967110-3d55-4767-9fa9-230138c92e42-kube-api-access-bb5dq\") pod \"controller-manager-879f6c89f-4s4sh\" (UID: \"23967110-3d55-4767-9fa9-230138c92e42\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4s4sh" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.989384 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/79e3d4d9-ec8e-4578-871c-a336ad22a2ac-serving-cert\") pod \"console-operator-58897d9998-9lzll\" (UID: \"79e3d4d9-ec8e-4578-871c-a336ad22a2ac\") " pod="openshift-console-operator/console-operator-58897d9998-9lzll" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.989431 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/be26df1a-a3b3-411b-9e84-df356f3a7af3-config\") pod \"machine-approver-56656f9798-v4wls\" (UID: \"be26df1a-a3b3-411b-9e84-df356f3a7af3\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-v4wls" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.989434 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/55cac199-8203-40b3-8a4f-c571017d998b-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-wnfzj\" (UID: \"55cac199-8203-40b3-8a4f-c571017d998b\") " pod="openshift-authentication/oauth-openshift-558db77b4-wnfzj" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.989503 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-vh974"] Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.989588 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d4d9ab07-53b6-4aef-bcac-c0140d870370-client-ca\") pod \"route-controller-manager-6576b87f9c-fpw88\" (UID: \"d4d9ab07-53b6-4aef-bcac-c0140d870370\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fpw88" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.989628 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/9b861750-2517-4cf3-9c78-930480711af0-oauth-serving-cert\") pod \"console-f9d7485db-pb98x\" (UID: \"9b861750-2517-4cf3-9c78-930480711af0\") " pod="openshift-console/console-f9d7485db-pb98x" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.989647 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/be26df1a-a3b3-411b-9e84-df356f3a7af3-machine-approver-tls\") pod \"machine-approver-56656f9798-v4wls\" (UID: \"be26df1a-a3b3-411b-9e84-df356f3a7af3\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-v4wls" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.989667 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7dc5fba4-2ba8-4fa2-a76f-0ad8d1196838-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-frls8\" (UID: \"7dc5fba4-2ba8-4fa2-a76f-0ad8d1196838\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-frls8" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.989674 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-n27w4" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.989687 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-smhnq\" (UniqueName: \"kubernetes.io/projected/78d69d0e-bfac-4582-b339-b2e585de5eff-kube-api-access-smhnq\") pod \"machine-config-operator-74547568cd-4fbj7\" (UID: \"78d69d0e-bfac-4582-b339-b2e585de5eff\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4fbj7" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.989708 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/31bd2041-6cb8-4eb6-bb22-5efe6c6009ee-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-2m4cb\" (UID: \"31bd2041-6cb8-4eb6-bb22-5efe6c6009ee\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2m4cb" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.989727 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/9b861750-2517-4cf3-9c78-930480711af0-console-oauth-config\") pod \"console-f9d7485db-pb98x\" (UID: \"9b861750-2517-4cf3-9c78-930480711af0\") " pod="openshift-console/console-f9d7485db-pb98x" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.989746 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/94294737-067b-4a4f-b59b-f9f987fa9127-etcd-serving-ca\") pod \"apiserver-76f77b778f-vh974\" (UID: \"94294737-067b-4a4f-b59b-f9f987fa9127\") " pod="openshift-apiserver/apiserver-76f77b778f-vh974" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.989776 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tsm8d\" (UniqueName: \"kubernetes.io/projected/c0beaf3d-a9eb-47cc-bb53-392e75959282-kube-api-access-tsm8d\") pod \"cluster-image-registry-operator-dc59b4c8b-ckmwl\" (UID: \"c0beaf3d-a9eb-47cc-bb53-392e75959282\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ckmwl" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.989796 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kznfz\" (UniqueName: \"kubernetes.io/projected/ac3d1daa-cb6d-41fc-9798-e687d9b2a21d-kube-api-access-kznfz\") pod \"collect-profiles-29396625-nrhcv\" (UID: \"ac3d1daa-cb6d-41fc-9798-e687d9b2a21d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396625-nrhcv" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.989819 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cb90f2d2-3321-4118-a449-fdc9343b23e1-trusted-ca\") pod \"ingress-operator-5b745b69d9-bsrcz\" (UID: \"cb90f2d2-3321-4118-a449-fdc9343b23e1\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-bsrcz" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.989838 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/55cac199-8203-40b3-8a4f-c571017d998b-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-wnfzj\" (UID: \"55cac199-8203-40b3-8a4f-c571017d998b\") " pod="openshift-authentication/oauth-openshift-558db77b4-wnfzj" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.989855 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/31bd2041-6cb8-4eb6-bb22-5efe6c6009ee-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-2m4cb\" (UID: \"31bd2041-6cb8-4eb6-bb22-5efe6c6009ee\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2m4cb" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.989873 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/352693a0-531b-4017-9582-5424c4691795-stats-auth\") pod \"router-default-5444994796-x5b9n\" (UID: \"352693a0-531b-4017-9582-5424c4691795\") " pod="openshift-ingress/router-default-5444994796-x5b9n" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.989892 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rv7fq\" (UniqueName: \"kubernetes.io/projected/ffef360f-8678-442d-9b27-b8de7646f1a3-kube-api-access-rv7fq\") pod \"openshift-controller-manager-operator-756b6f6bc6-t5d22\" (UID: \"ffef360f-8678-442d-9b27-b8de7646f1a3\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-t5d22" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.989909 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c0beaf3d-a9eb-47cc-bb53-392e75959282-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-ckmwl\" (UID: \"c0beaf3d-a9eb-47cc-bb53-392e75959282\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ckmwl" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.989926 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7dqck\" (UniqueName: \"kubernetes.io/projected/83ac1d78-cc4c-48b7-b975-2f88802e2d17-kube-api-access-7dqck\") pod \"downloads-7954f5f757-tdpht\" (UID: \"83ac1d78-cc4c-48b7-b975-2f88802e2d17\") " pod="openshift-console/downloads-7954f5f757-tdpht" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.989946 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-42kgf\" (UniqueName: \"kubernetes.io/projected/cb90f2d2-3321-4118-a449-fdc9343b23e1-kube-api-access-42kgf\") pod \"ingress-operator-5b745b69d9-bsrcz\" (UID: \"cb90f2d2-3321-4118-a449-fdc9343b23e1\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-bsrcz" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.989963 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hmtsb\" (UniqueName: \"kubernetes.io/projected/1ca00755-c6ae-42d2-833e-fc31b6a480df-kube-api-access-hmtsb\") pod \"authentication-operator-69f744f599-ksdn9\" (UID: \"1ca00755-c6ae-42d2-833e-fc31b6a480df\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ksdn9" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.989990 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d32dc4fd-3e62-402c-a1d9-8f2c83f2e334-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-tnmrh\" (UID: \"d32dc4fd-3e62-402c-a1d9-8f2c83f2e334\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-tnmrh" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.990012 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/ff518a19-6241-4c65-b2a0-bee6ab939a6c-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-6jw6b\" (UID: \"ff518a19-6241-4c65-b2a0-bee6ab939a6c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6jw6b" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.990030 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ca00755-c6ae-42d2-833e-fc31b6a480df-config\") pod \"authentication-operator-69f744f599-ksdn9\" (UID: \"1ca00755-c6ae-42d2-833e-fc31b6a480df\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ksdn9" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.990048 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xh6c5\" (UniqueName: \"kubernetes.io/projected/9b861750-2517-4cf3-9c78-930480711af0-kube-api-access-xh6c5\") pod \"console-f9d7485db-pb98x\" (UID: \"9b861750-2517-4cf3-9c78-930480711af0\") " pod="openshift-console/console-f9d7485db-pb98x" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.990065 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/7dc5fba4-2ba8-4fa2-a76f-0ad8d1196838-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-frls8\" (UID: \"7dc5fba4-2ba8-4fa2-a76f-0ad8d1196838\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-frls8" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.990083 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bgr4n\" (UniqueName: \"kubernetes.io/projected/94294737-067b-4a4f-b59b-f9f987fa9127-kube-api-access-bgr4n\") pod \"apiserver-76f77b778f-vh974\" (UID: \"94294737-067b-4a4f-b59b-f9f987fa9127\") " pod="openshift-apiserver/apiserver-76f77b778f-vh974" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.990099 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/78d69d0e-bfac-4582-b339-b2e585de5eff-images\") pod \"machine-config-operator-74547568cd-4fbj7\" (UID: \"78d69d0e-bfac-4582-b339-b2e585de5eff\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4fbj7" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.990119 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sp9jx\" (UniqueName: \"kubernetes.io/projected/be26df1a-a3b3-411b-9e84-df356f3a7af3-kube-api-access-sp9jx\") pod \"machine-approver-56656f9798-v4wls\" (UID: \"be26df1a-a3b3-411b-9e84-df356f3a7af3\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-v4wls" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.990136 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/cb9c37ae-6151-452f-b467-6b7d14bbb2f8-srv-cert\") pod \"catalog-operator-68c6474976-h6f5v\" (UID: \"cb9c37ae-6151-452f-b467-6b7d14bbb2f8\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-h6f5v" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.990160 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1507f6ee-6a14-4d67-98b3-1e151b7949cc-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-kdt8l\" (UID: \"1507f6ee-6a14-4d67-98b3-1e151b7949cc\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-kdt8l" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.990183 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n6hh2\" (UniqueName: \"kubernetes.io/projected/352693a0-531b-4017-9582-5424c4691795-kube-api-access-n6hh2\") pod \"router-default-5444994796-x5b9n\" (UID: \"352693a0-531b-4017-9582-5424c4691795\") " pod="openshift-ingress/router-default-5444994796-x5b9n" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.990238 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ffef360f-8678-442d-9b27-b8de7646f1a3-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-t5d22\" (UID: \"ffef360f-8678-442d-9b27-b8de7646f1a3\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-t5d22" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.990257 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1ca00755-c6ae-42d2-833e-fc31b6a480df-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-ksdn9\" (UID: \"1ca00755-c6ae-42d2-833e-fc31b6a480df\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ksdn9" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.990273 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/cb90f2d2-3321-4118-a449-fdc9343b23e1-metrics-tls\") pod \"ingress-operator-5b745b69d9-bsrcz\" (UID: \"cb90f2d2-3321-4118-a449-fdc9343b23e1\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-bsrcz" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.990312 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/94294737-067b-4a4f-b59b-f9f987fa9127-trusted-ca-bundle\") pod \"apiserver-76f77b778f-vh974\" (UID: \"94294737-067b-4a4f-b59b-f9f987fa9127\") " pod="openshift-apiserver/apiserver-76f77b778f-vh974" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.990331 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1507f6ee-6a14-4d67-98b3-1e151b7949cc-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-kdt8l\" (UID: \"1507f6ee-6a14-4d67-98b3-1e151b7949cc\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-kdt8l" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.990351 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5btzg\" (UniqueName: \"kubernetes.io/projected/ff518a19-6241-4c65-b2a0-bee6ab939a6c-kube-api-access-5btzg\") pod \"cluster-samples-operator-665b6dd947-6jw6b\" (UID: \"ff518a19-6241-4c65-b2a0-bee6ab939a6c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6jw6b" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.990388 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d4d9ab07-53b6-4aef-bcac-c0140d870370-serving-cert\") pod \"route-controller-manager-6576b87f9c-fpw88\" (UID: \"d4d9ab07-53b6-4aef-bcac-c0140d870370\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fpw88" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.990395 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d4d9ab07-53b6-4aef-bcac-c0140d870370-client-ca\") pod \"route-controller-manager-6576b87f9c-fpw88\" (UID: \"d4d9ab07-53b6-4aef-bcac-c0140d870370\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fpw88" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.990406 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1ca00755-c6ae-42d2-833e-fc31b6a480df-service-ca-bundle\") pod \"authentication-operator-69f744f599-ksdn9\" (UID: \"1ca00755-c6ae-42d2-833e-fc31b6a480df\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ksdn9" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.990450 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/78d69d0e-bfac-4582-b339-b2e585de5eff-auth-proxy-config\") pod \"machine-config-operator-74547568cd-4fbj7\" (UID: \"78d69d0e-bfac-4582-b339-b2e585de5eff\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4fbj7" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.990474 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d32dc4fd-3e62-402c-a1d9-8f2c83f2e334-config\") pod \"kube-controller-manager-operator-78b949d7b-tnmrh\" (UID: \"d32dc4fd-3e62-402c-a1d9-8f2c83f2e334\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-tnmrh" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.990496 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/55cac199-8203-40b3-8a4f-c571017d998b-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-wnfzj\" (UID: \"55cac199-8203-40b3-8a4f-c571017d998b\") " pod="openshift-authentication/oauth-openshift-558db77b4-wnfzj" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.990522 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ac3d1daa-cb6d-41fc-9798-e687d9b2a21d-secret-volume\") pod \"collect-profiles-29396625-nrhcv\" (UID: \"ac3d1daa-cb6d-41fc-9798-e687d9b2a21d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396625-nrhcv" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.990538 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7dc5fba4-2ba8-4fa2-a76f-0ad8d1196838-serving-cert\") pod \"apiserver-7bbb656c7d-frls8\" (UID: \"7dc5fba4-2ba8-4fa2-a76f-0ad8d1196838\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-frls8" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.990554 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/dfa022a0-f86e-4f4b-99f5-22b84e9d6558-srv-cert\") pod \"olm-operator-6b444d44fb-9n9mc\" (UID: \"dfa022a0-f86e-4f4b-99f5-22b84e9d6558\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9n9mc" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.990571 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/55cac199-8203-40b3-8a4f-c571017d998b-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-wnfzj\" (UID: \"55cac199-8203-40b3-8a4f-c571017d998b\") " pod="openshift-authentication/oauth-openshift-558db77b4-wnfzj" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.990606 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9b861750-2517-4cf3-9c78-930480711af0-service-ca\") pod \"console-f9d7485db-pb98x\" (UID: \"9b861750-2517-4cf3-9c78-930480711af0\") " pod="openshift-console/console-f9d7485db-pb98x" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.990628 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vj4w5\" (UniqueName: \"kubernetes.io/projected/dd618778-dc82-4797-b737-14231eefea15-kube-api-access-vj4w5\") pod \"service-ca-operator-777779d784-fg7wj\" (UID: \"dd618778-dc82-4797-b737-14231eefea15\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-fg7wj" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.990650 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1ca00755-c6ae-42d2-833e-fc31b6a480df-serving-cert\") pod \"authentication-operator-69f744f599-ksdn9\" (UID: \"1ca00755-c6ae-42d2-833e-fc31b6a480df\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ksdn9" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.990669 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/7dc5fba4-2ba8-4fa2-a76f-0ad8d1196838-encryption-config\") pod \"apiserver-7bbb656c7d-frls8\" (UID: \"7dc5fba4-2ba8-4fa2-a76f-0ad8d1196838\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-frls8" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.990686 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/55cac199-8203-40b3-8a4f-c571017d998b-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-wnfzj\" (UID: \"55cac199-8203-40b3-8a4f-c571017d998b\") " pod="openshift-authentication/oauth-openshift-558db77b4-wnfzj" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.990707 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qsdgk\" (UniqueName: \"kubernetes.io/projected/79e3d4d9-ec8e-4578-871c-a336ad22a2ac-kube-api-access-qsdgk\") pod \"console-operator-58897d9998-9lzll\" (UID: \"79e3d4d9-ec8e-4578-871c-a336ad22a2ac\") " pod="openshift-console-operator/console-operator-58897d9998-9lzll" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.990725 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/94294737-067b-4a4f-b59b-f9f987fa9127-config\") pod \"apiserver-76f77b778f-vh974\" (UID: \"94294737-067b-4a4f-b59b-f9f987fa9127\") " pod="openshift-apiserver/apiserver-76f77b778f-vh974" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.990742 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/55cac199-8203-40b3-8a4f-c571017d998b-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-wnfzj\" (UID: \"55cac199-8203-40b3-8a4f-c571017d998b\") " pod="openshift-authentication/oauth-openshift-558db77b4-wnfzj" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.990778 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ffef360f-8678-442d-9b27-b8de7646f1a3-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-t5d22\" (UID: \"ffef360f-8678-442d-9b27-b8de7646f1a3\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-t5d22" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.990799 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/23967110-3d55-4767-9fa9-230138c92e42-client-ca\") pod \"controller-manager-879f6c89f-4s4sh\" (UID: \"23967110-3d55-4767-9fa9-230138c92e42\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4s4sh" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.990834 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9b861750-2517-4cf3-9c78-930480711af0-trusted-ca-bundle\") pod \"console-f9d7485db-pb98x\" (UID: \"9b861750-2517-4cf3-9c78-930480711af0\") " pod="openshift-console/console-f9d7485db-pb98x" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.990862 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/94294737-067b-4a4f-b59b-f9f987fa9127-node-pullsecrets\") pod \"apiserver-76f77b778f-vh974\" (UID: \"94294737-067b-4a4f-b59b-f9f987fa9127\") " pod="openshift-apiserver/apiserver-76f77b778f-vh974" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.990884 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ac3d1daa-cb6d-41fc-9798-e687d9b2a21d-config-volume\") pod \"collect-profiles-29396625-nrhcv\" (UID: \"ac3d1daa-cb6d-41fc-9798-e687d9b2a21d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396625-nrhcv" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.990907 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/be26df1a-a3b3-411b-9e84-df356f3a7af3-auth-proxy-config\") pod \"machine-approver-56656f9798-v4wls\" (UID: \"be26df1a-a3b3-411b-9e84-df356f3a7af3\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-v4wls" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.990926 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/9b861750-2517-4cf3-9c78-930480711af0-console-serving-cert\") pod \"console-f9d7485db-pb98x\" (UID: \"9b861750-2517-4cf3-9c78-930480711af0\") " pod="openshift-console/console-f9d7485db-pb98x" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.990948 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d086acbe-d982-4880-9538-04a184cb4148-config\") pod \"machine-api-operator-5694c8668f-spnzw\" (UID: \"d086acbe-d982-4880-9538-04a184cb4148\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-spnzw" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.990966 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/55cac199-8203-40b3-8a4f-c571017d998b-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-wnfzj\" (UID: \"55cac199-8203-40b3-8a4f-c571017d998b\") " pod="openshift-authentication/oauth-openshift-558db77b4-wnfzj" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.990979 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/79e3d4d9-ec8e-4578-871c-a336ad22a2ac-trusted-ca\") pod \"console-operator-58897d9998-9lzll\" (UID: \"79e3d4d9-ec8e-4578-871c-a336ad22a2ac\") " pod="openshift-console-operator/console-operator-58897d9998-9lzll" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.990990 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d4d9ab07-53b6-4aef-bcac-c0140d870370-config\") pod \"route-controller-manager-6576b87f9c-fpw88\" (UID: \"d4d9ab07-53b6-4aef-bcac-c0140d870370\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fpw88" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.991126 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kb79q\" (UniqueName: \"kubernetes.io/projected/ce67cd3e-c2d2-4562-bc26-0a68bce1a1f0-kube-api-access-kb79q\") pod \"openshift-config-operator-7777fb866f-qwhkx\" (UID: \"ce67cd3e-c2d2-4562-bc26-0a68bce1a1f0\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-qwhkx" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.991154 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/23967110-3d55-4767-9fa9-230138c92e42-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-4s4sh\" (UID: \"23967110-3d55-4767-9fa9-230138c92e42\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4s4sh" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.992628 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9b861750-2517-4cf3-9c78-930480711af0-service-ca\") pod \"console-f9d7485db-pb98x\" (UID: \"9b861750-2517-4cf3-9c78-930480711af0\") " pod="openshift-console/console-f9d7485db-pb98x" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.993522 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1ca00755-c6ae-42d2-833e-fc31b6a480df-service-ca-bundle\") pod \"authentication-operator-69f744f599-ksdn9\" (UID: \"1ca00755-c6ae-42d2-833e-fc31b6a480df\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ksdn9" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.994397 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ca00755-c6ae-42d2-833e-fc31b6a480df-config\") pod \"authentication-operator-69f744f599-ksdn9\" (UID: \"1ca00755-c6ae-42d2-833e-fc31b6a480df\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ksdn9" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.994738 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d4d9ab07-53b6-4aef-bcac-c0140d870370-config\") pod \"route-controller-manager-6576b87f9c-fpw88\" (UID: \"d4d9ab07-53b6-4aef-bcac-c0140d870370\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fpw88" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.994785 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/9b861750-2517-4cf3-9c78-930480711af0-oauth-serving-cert\") pod \"console-f9d7485db-pb98x\" (UID: \"9b861750-2517-4cf3-9c78-930480711af0\") " pod="openshift-console/console-f9d7485db-pb98x" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.994915 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ffef360f-8678-442d-9b27-b8de7646f1a3-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-t5d22\" (UID: \"ffef360f-8678-442d-9b27-b8de7646f1a3\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-t5d22" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.995004 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/be26df1a-a3b3-411b-9e84-df356f3a7af3-machine-approver-tls\") pod \"machine-approver-56656f9798-v4wls\" (UID: \"be26df1a-a3b3-411b-9e84-df356f3a7af3\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-v4wls" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.990350 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-bsrcz"] Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.995065 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-pb98x"] Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.995451 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1ca00755-c6ae-42d2-833e-fc31b6a480df-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-ksdn9\" (UID: \"1ca00755-c6ae-42d2-833e-fc31b6a480df\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ksdn9" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.995491 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-k5c2k"] Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.995522 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/23967110-3d55-4767-9fa9-230138c92e42-client-ca\") pod \"controller-manager-879f6c89f-4s4sh\" (UID: \"23967110-3d55-4767-9fa9-230138c92e42\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4s4sh" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.996183 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/23967110-3d55-4767-9fa9-230138c92e42-serving-cert\") pod \"controller-manager-879f6c89f-4s4sh\" (UID: \"23967110-3d55-4767-9fa9-230138c92e42\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4s4sh" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.996183 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9b861750-2517-4cf3-9c78-930480711af0-trusted-ca-bundle\") pod \"console-f9d7485db-pb98x\" (UID: \"9b861750-2517-4cf3-9c78-930480711af0\") " pod="openshift-console/console-f9d7485db-pb98x" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.996225 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/94294737-067b-4a4f-b59b-f9f987fa9127-image-import-ca\") pod \"apiserver-76f77b778f-vh974\" (UID: \"94294737-067b-4a4f-b59b-f9f987fa9127\") " pod="openshift-apiserver/apiserver-76f77b778f-vh974" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.996599 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/9b861750-2517-4cf3-9c78-930480711af0-console-config\") pod \"console-f9d7485db-pb98x\" (UID: \"9b861750-2517-4cf3-9c78-930480711af0\") " pod="openshift-console/console-f9d7485db-pb98x" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.996644 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h8d7k\" (UniqueName: \"kubernetes.io/projected/0fb11c40-1e46-42a7-b384-626786bc4e3e-kube-api-access-h8d7k\") pod \"openshift-apiserver-operator-796bbdcf4f-wjmrv\" (UID: \"0fb11c40-1e46-42a7-b384-626786bc4e3e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wjmrv" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.996690 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/94294737-067b-4a4f-b59b-f9f987fa9127-audit-dir\") pod \"apiserver-76f77b778f-vh974\" (UID: \"94294737-067b-4a4f-b59b-f9f987fa9127\") " pod="openshift-apiserver/apiserver-76f77b778f-vh974" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.996719 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wdcz2\" (UniqueName: \"kubernetes.io/projected/55cac199-8203-40b3-8a4f-c571017d998b-kube-api-access-wdcz2\") pod \"oauth-openshift-558db77b4-wnfzj\" (UID: \"55cac199-8203-40b3-8a4f-c571017d998b\") " pod="openshift-authentication/oauth-openshift-558db77b4-wnfzj" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.996764 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s86jn\" (UniqueName: \"kubernetes.io/projected/3cfa3a64-3e7a-4336-9ae0-86822095a1b4-kube-api-access-s86jn\") pod \"multus-admission-controller-857f4d67dd-fb49f\" (UID: \"3cfa3a64-3e7a-4336-9ae0-86822095a1b4\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-fb49f" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.996797 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c0beaf3d-a9eb-47cc-bb53-392e75959282-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-ckmwl\" (UID: \"c0beaf3d-a9eb-47cc-bb53-392e75959282\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ckmwl" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.996826 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kppbq\" (UniqueName: \"kubernetes.io/projected/d086acbe-d982-4880-9538-04a184cb4148-kube-api-access-kppbq\") pod \"machine-api-operator-5694c8668f-spnzw\" (UID: \"d086acbe-d982-4880-9538-04a184cb4148\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-spnzw" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.996856 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/31bd2041-6cb8-4eb6-bb22-5efe6c6009ee-config\") pod \"kube-apiserver-operator-766d6c64bb-2m4cb\" (UID: \"31bd2041-6cb8-4eb6-bb22-5efe6c6009ee\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2m4cb" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.996882 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/cb90f2d2-3321-4118-a449-fdc9343b23e1-bound-sa-token\") pod \"ingress-operator-5b745b69d9-bsrcz\" (UID: \"cb90f2d2-3321-4118-a449-fdc9343b23e1\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-bsrcz" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.996919 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2mgz5\" (UniqueName: \"kubernetes.io/projected/7dc5fba4-2ba8-4fa2-a76f-0ad8d1196838-kube-api-access-2mgz5\") pod \"apiserver-7bbb656c7d-frls8\" (UID: \"7dc5fba4-2ba8-4fa2-a76f-0ad8d1196838\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-frls8" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.996946 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/c0beaf3d-a9eb-47cc-bb53-392e75959282-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-ckmwl\" (UID: \"c0beaf3d-a9eb-47cc-bb53-392e75959282\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ckmwl" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.996977 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dd618778-dc82-4797-b737-14231eefea15-serving-cert\") pod \"service-ca-operator-777779d784-fg7wj\" (UID: \"dd618778-dc82-4797-b737-14231eefea15\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-fg7wj" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.997005 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/94294737-067b-4a4f-b59b-f9f987fa9127-encryption-config\") pod \"apiserver-76f77b778f-vh974\" (UID: \"94294737-067b-4a4f-b59b-f9f987fa9127\") " pod="openshift-apiserver/apiserver-76f77b778f-vh974" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.997030 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vzd9t\" (UniqueName: \"kubernetes.io/projected/cb9c37ae-6151-452f-b467-6b7d14bbb2f8-kube-api-access-vzd9t\") pod \"catalog-operator-68c6474976-h6f5v\" (UID: \"cb9c37ae-6151-452f-b467-6b7d14bbb2f8\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-h6f5v" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.997068 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0fb11c40-1e46-42a7-b384-626786bc4e3e-config\") pod \"openshift-apiserver-operator-796bbdcf4f-wjmrv\" (UID: \"0fb11c40-1e46-42a7-b384-626786bc4e3e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wjmrv" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.997094 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/94294737-067b-4a4f-b59b-f9f987fa9127-etcd-client\") pod \"apiserver-76f77b778f-vh974\" (UID: \"94294737-067b-4a4f-b59b-f9f987fa9127\") " pod="openshift-apiserver/apiserver-76f77b778f-vh974" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.997125 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/78d69d0e-bfac-4582-b339-b2e585de5eff-proxy-tls\") pod \"machine-config-operator-74547568cd-4fbj7\" (UID: \"78d69d0e-bfac-4582-b339-b2e585de5eff\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4fbj7" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.996656 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/79e3d4d9-ec8e-4578-871c-a336ad22a2ac-serving-cert\") pod \"console-operator-58897d9998-9lzll\" (UID: \"79e3d4d9-ec8e-4578-871c-a336ad22a2ac\") " pod="openshift-console-operator/console-operator-58897d9998-9lzll" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.997485 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/9b861750-2517-4cf3-9c78-930480711af0-console-config\") pod \"console-f9d7485db-pb98x\" (UID: \"9b861750-2517-4cf3-9c78-930480711af0\") " pod="openshift-console/console-f9d7485db-pb98x" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.998721 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/79e3d4d9-ec8e-4578-871c-a336ad22a2ac-config\") pod \"console-operator-58897d9998-9lzll\" (UID: \"79e3d4d9-ec8e-4578-871c-a336ad22a2ac\") " pod="openshift-console-operator/console-operator-58897d9998-9lzll" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.998831 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1507f6ee-6a14-4d67-98b3-1e151b7949cc-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-kdt8l\" (UID: \"1507f6ee-6a14-4d67-98b3-1e151b7949cc\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-kdt8l" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.998887 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/352693a0-531b-4017-9582-5424c4691795-default-certificate\") pod \"router-default-5444994796-x5b9n\" (UID: \"352693a0-531b-4017-9582-5424c4691795\") " pod="openshift-ingress/router-default-5444994796-x5b9n" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.998950 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ce67cd3e-c2d2-4562-bc26-0a68bce1a1f0-serving-cert\") pod \"openshift-config-operator-7777fb866f-qwhkx\" (UID: \"ce67cd3e-c2d2-4562-bc26-0a68bce1a1f0\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-qwhkx" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.998985 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/ce67cd3e-c2d2-4562-bc26-0a68bce1a1f0-available-featuregates\") pod \"openshift-config-operator-7777fb866f-qwhkx\" (UID: \"ce67cd3e-c2d2-4562-bc26-0a68bce1a1f0\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-qwhkx" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.999012 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/d086acbe-d982-4880-9538-04a184cb4148-images\") pod \"machine-api-operator-5694c8668f-spnzw\" (UID: \"d086acbe-d982-4880-9538-04a184cb4148\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-spnzw" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.999049 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jf27d\" (UniqueName: \"kubernetes.io/projected/dfa022a0-f86e-4f4b-99f5-22b84e9d6558-kube-api-access-jf27d\") pod \"olm-operator-6b444d44fb-9n9mc\" (UID: \"dfa022a0-f86e-4f4b-99f5-22b84e9d6558\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9n9mc" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.999080 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/352693a0-531b-4017-9582-5424c4691795-metrics-certs\") pod \"router-default-5444994796-x5b9n\" (UID: \"352693a0-531b-4017-9582-5424c4691795\") " pod="openshift-ingress/router-default-5444994796-x5b9n" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.999118 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/23967110-3d55-4767-9fa9-230138c92e42-config\") pod \"controller-manager-879f6c89f-4s4sh\" (UID: \"23967110-3d55-4767-9fa9-230138c92e42\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4s4sh" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.999283 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c0beaf3d-a9eb-47cc-bb53-392e75959282-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-ckmwl\" (UID: \"c0beaf3d-a9eb-47cc-bb53-392e75959282\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ckmwl" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.999463 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/79e3d4d9-ec8e-4578-871c-a336ad22a2ac-config\") pod \"console-operator-58897d9998-9lzll\" (UID: \"79e3d4d9-ec8e-4578-871c-a336ad22a2ac\") " pod="openshift-console-operator/console-operator-58897d9998-9lzll" Nov 22 07:57:18 crc kubenswrapper[4789]: I1122 07:57:18.999510 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/be26df1a-a3b3-411b-9e84-df356f3a7af3-auth-proxy-config\") pod \"machine-approver-56656f9798-v4wls\" (UID: \"be26df1a-a3b3-411b-9e84-df356f3a7af3\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-v4wls" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.000302 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d4d9ab07-53b6-4aef-bcac-c0140d870370-serving-cert\") pod \"route-controller-manager-6576b87f9c-fpw88\" (UID: \"d4d9ab07-53b6-4aef-bcac-c0140d870370\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fpw88" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.000456 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396625-nrhcv"] Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.000695 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1ca00755-c6ae-42d2-833e-fc31b6a480df-serving-cert\") pod \"authentication-operator-69f744f599-ksdn9\" (UID: \"1ca00755-c6ae-42d2-833e-fc31b6a480df\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ksdn9" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.000860 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/23967110-3d55-4767-9fa9-230138c92e42-config\") pod \"controller-manager-879f6c89f-4s4sh\" (UID: \"23967110-3d55-4767-9fa9-230138c92e42\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4s4sh" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.000899 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/9b861750-2517-4cf3-9c78-930480711af0-console-serving-cert\") pod \"console-f9d7485db-pb98x\" (UID: \"9b861750-2517-4cf3-9c78-930480711af0\") " pod="openshift-console/console-f9d7485db-pb98x" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.001384 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/9b861750-2517-4cf3-9c78-930480711af0-console-oauth-config\") pod \"console-f9d7485db-pb98x\" (UID: \"9b861750-2517-4cf3-9c78-930480711af0\") " pod="openshift-console/console-f9d7485db-pb98x" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.001660 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.002108 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/ce67cd3e-c2d2-4562-bc26-0a68bce1a1f0-available-featuregates\") pod \"openshift-config-operator-7777fb866f-qwhkx\" (UID: \"ce67cd3e-c2d2-4562-bc26-0a68bce1a1f0\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-qwhkx" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.002496 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0fb11c40-1e46-42a7-b384-626786bc4e3e-config\") pod \"openshift-apiserver-operator-796bbdcf4f-wjmrv\" (UID: \"0fb11c40-1e46-42a7-b384-626786bc4e3e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wjmrv" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.004085 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-tnmrh"] Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.004989 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/c0beaf3d-a9eb-47cc-bb53-392e75959282-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-ckmwl\" (UID: \"c0beaf3d-a9eb-47cc-bb53-392e75959282\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ckmwl" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.007225 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-4fbj7"] Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.008160 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0fb11c40-1e46-42a7-b384-626786bc4e3e-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-wjmrv\" (UID: \"0fb11c40-1e46-42a7-b384-626786bc4e3e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wjmrv" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.011948 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-fb49f"] Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.012542 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ce67cd3e-c2d2-4562-bc26-0a68bce1a1f0-serving-cert\") pod \"openshift-config-operator-7777fb866f-qwhkx\" (UID: \"ce67cd3e-c2d2-4562-bc26-0a68bce1a1f0\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-qwhkx" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.013202 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/ff518a19-6241-4c65-b2a0-bee6ab939a6c-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-6jw6b\" (UID: \"ff518a19-6241-4c65-b2a0-bee6ab939a6c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6jw6b" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.016772 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-4s4sh"] Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.019047 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-ksdn9"] Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.020484 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-whfpb"] Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.020551 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-fg7wj"] Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.021593 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-fr4zx"] Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.022672 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.023122 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-spnzw"] Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.024246 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-h6f5v"] Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.025610 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-zl87n"] Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.027313 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-xlf6m"] Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.028104 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-wnfzj"] Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.029271 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-hszd6"] Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.030376 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-77svl"] Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.031489 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2m4cb"] Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.032844 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2dxsq"] Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.034583 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-d4l6q"] Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.036132 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9n9mc"] Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.037851 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-l4q7j"] Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.039234 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-sszl6"] Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.040489 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-qc925"] Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.040996 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.041708 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-n27w4"] Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.042821 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-szb9d"] Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.043650 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-szb9d" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.043848 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-szb9d"] Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.061668 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.081965 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.100181 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/352693a0-531b-4017-9582-5424c4691795-metrics-certs\") pod \"router-default-5444994796-x5b9n\" (UID: \"352693a0-531b-4017-9582-5424c4691795\") " pod="openshift-ingress/router-default-5444994796-x5b9n" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.100375 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/75fa6283-acd8-41bc-9803-2f55119f6828-tuning-conf-dir\") pod \"cni-sysctl-allowlist-ds-gcnts\" (UID: \"75fa6283-acd8-41bc-9803-2f55119f6828\") " pod="openshift-multus/cni-sysctl-allowlist-ds-gcnts" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.100526 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/d086acbe-d982-4880-9538-04a184cb4148-images\") pod \"machine-api-operator-5694c8668f-spnzw\" (UID: \"d086acbe-d982-4880-9538-04a184cb4148\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-spnzw" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.100645 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jf27d\" (UniqueName: \"kubernetes.io/projected/dfa022a0-f86e-4f4b-99f5-22b84e9d6558-kube-api-access-jf27d\") pod \"olm-operator-6b444d44fb-9n9mc\" (UID: \"dfa022a0-f86e-4f4b-99f5-22b84e9d6558\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9n9mc" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.100745 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xk2ns\" (UniqueName: \"kubernetes.io/projected/c562523e-6eb7-4c72-8ba2-f968f171d387-kube-api-access-xk2ns\") pod \"machine-config-server-qh9np\" (UID: \"c562523e-6eb7-4c72-8ba2-f968f171d387\") " pod="openshift-machine-config-operator/machine-config-server-qh9np" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.100884 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/c562523e-6eb7-4c72-8ba2-f968f171d387-certs\") pod \"machine-config-server-qh9np\" (UID: \"c562523e-6eb7-4c72-8ba2-f968f171d387\") " pod="openshift-machine-config-operator/machine-config-server-qh9np" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.101018 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd618778-dc82-4797-b737-14231eefea15-config\") pod \"service-ca-operator-777779d784-fg7wj\" (UID: \"dd618778-dc82-4797-b737-14231eefea15\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-fg7wj" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.101176 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/3cfa3a64-3e7a-4336-9ae0-86822095a1b4-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-fb49f\" (UID: \"3cfa3a64-3e7a-4336-9ae0-86822095a1b4\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-fb49f" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.101304 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f8f3dfa-5fc0-4642-96c0-22b6f3081676-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-d4l6q\" (UID: \"7f8f3dfa-5fc0-4642-96c0-22b6f3081676\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-d4l6q" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.101425 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/d086acbe-d982-4880-9538-04a184cb4148-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-spnzw\" (UID: \"d086acbe-d982-4880-9538-04a184cb4148\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-spnzw" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.101530 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/7dc5fba4-2ba8-4fa2-a76f-0ad8d1196838-etcd-client\") pod \"apiserver-7bbb656c7d-frls8\" (UID: \"7dc5fba4-2ba8-4fa2-a76f-0ad8d1196838\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-frls8" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.101656 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/55cac199-8203-40b3-8a4f-c571017d998b-audit-policies\") pod \"oauth-openshift-558db77b4-wnfzj\" (UID: \"55cac199-8203-40b3-8a4f-c571017d998b\") " pod="openshift-authentication/oauth-openshift-558db77b4-wnfzj" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.101734 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.101880 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/55cac199-8203-40b3-8a4f-c571017d998b-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-wnfzj\" (UID: \"55cac199-8203-40b3-8a4f-c571017d998b\") " pod="openshift-authentication/oauth-openshift-558db77b4-wnfzj" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.102001 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7dc5fba4-2ba8-4fa2-a76f-0ad8d1196838-audit-dir\") pod \"apiserver-7bbb656c7d-frls8\" (UID: \"7dc5fba4-2ba8-4fa2-a76f-0ad8d1196838\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-frls8" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.102095 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7dc5fba4-2ba8-4fa2-a76f-0ad8d1196838-audit-dir\") pod \"apiserver-7bbb656c7d-frls8\" (UID: \"7dc5fba4-2ba8-4fa2-a76f-0ad8d1196838\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-frls8" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.102216 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/55cac199-8203-40b3-8a4f-c571017d998b-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-wnfzj\" (UID: \"55cac199-8203-40b3-8a4f-c571017d998b\") " pod="openshift-authentication/oauth-openshift-558db77b4-wnfzj" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.102291 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/55cac199-8203-40b3-8a4f-c571017d998b-audit-policies\") pod \"oauth-openshift-558db77b4-wnfzj\" (UID: \"55cac199-8203-40b3-8a4f-c571017d998b\") " pod="openshift-authentication/oauth-openshift-558db77b4-wnfzj" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.102419 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/94294737-067b-4a4f-b59b-f9f987fa9127-audit\") pod \"apiserver-76f77b778f-vh974\" (UID: \"94294737-067b-4a4f-b59b-f9f987fa9127\") " pod="openshift-apiserver/apiserver-76f77b778f-vh974" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.102544 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/55cac199-8203-40b3-8a4f-c571017d998b-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-wnfzj\" (UID: \"55cac199-8203-40b3-8a4f-c571017d998b\") " pod="openshift-authentication/oauth-openshift-558db77b4-wnfzj" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.102662 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/352693a0-531b-4017-9582-5424c4691795-service-ca-bundle\") pod \"router-default-5444994796-x5b9n\" (UID: \"352693a0-531b-4017-9582-5424c4691795\") " pod="openshift-ingress/router-default-5444994796-x5b9n" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.102843 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/55cac199-8203-40b3-8a4f-c571017d998b-audit-dir\") pod \"oauth-openshift-558db77b4-wnfzj\" (UID: \"55cac199-8203-40b3-8a4f-c571017d998b\") " pod="openshift-authentication/oauth-openshift-558db77b4-wnfzj" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.102930 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/94294737-067b-4a4f-b59b-f9f987fa9127-audit\") pod \"apiserver-76f77b778f-vh974\" (UID: \"94294737-067b-4a4f-b59b-f9f987fa9127\") " pod="openshift-apiserver/apiserver-76f77b778f-vh974" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.102968 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/55cac199-8203-40b3-8a4f-c571017d998b-audit-dir\") pod \"oauth-openshift-558db77b4-wnfzj\" (UID: \"55cac199-8203-40b3-8a4f-c571017d998b\") " pod="openshift-authentication/oauth-openshift-558db77b4-wnfzj" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.103088 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/d58d000b-3f2c-44cb-90c5-8023e9f4f5db-plugins-dir\") pod \"csi-hostpathplugin-n27w4\" (UID: \"d58d000b-3f2c-44cb-90c5-8023e9f4f5db\") " pod="hostpath-provisioner/csi-hostpathplugin-n27w4" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.103213 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/dfa022a0-f86e-4f4b-99f5-22b84e9d6558-profile-collector-cert\") pod \"olm-operator-6b444d44fb-9n9mc\" (UID: \"dfa022a0-f86e-4f4b-99f5-22b84e9d6558\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9n9mc" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.103345 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/55cac199-8203-40b3-8a4f-c571017d998b-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-wnfzj\" (UID: \"55cac199-8203-40b3-8a4f-c571017d998b\") " pod="openshift-authentication/oauth-openshift-558db77b4-wnfzj" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.103456 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6r7x6\" (UniqueName: \"kubernetes.io/projected/75fa6283-acd8-41bc-9803-2f55119f6828-kube-api-access-6r7x6\") pod \"cni-sysctl-allowlist-ds-gcnts\" (UID: \"75fa6283-acd8-41bc-9803-2f55119f6828\") " pod="openshift-multus/cni-sysctl-allowlist-ds-gcnts" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.103568 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/cb9c37ae-6151-452f-b467-6b7d14bbb2f8-profile-collector-cert\") pod \"catalog-operator-68c6474976-h6f5v\" (UID: \"cb9c37ae-6151-452f-b467-6b7d14bbb2f8\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-h6f5v" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.103674 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1d0ebe8e-6cea-43d3-9f15-b67c4cbc29e7-cert\") pod \"ingress-canary-qc925\" (UID: \"1d0ebe8e-6cea-43d3-9f15-b67c4cbc29e7\") " pod="openshift-ingress-canary/ingress-canary-qc925" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.103808 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d32dc4fd-3e62-402c-a1d9-8f2c83f2e334-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-tnmrh\" (UID: \"d32dc4fd-3e62-402c-a1d9-8f2c83f2e334\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-tnmrh" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.103922 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/94294737-067b-4a4f-b59b-f9f987fa9127-serving-cert\") pod \"apiserver-76f77b778f-vh974\" (UID: \"94294737-067b-4a4f-b59b-f9f987fa9127\") " pod="openshift-apiserver/apiserver-76f77b778f-vh974" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.104043 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7dc5fba4-2ba8-4fa2-a76f-0ad8d1196838-audit-policies\") pod \"apiserver-7bbb656c7d-frls8\" (UID: \"7dc5fba4-2ba8-4fa2-a76f-0ad8d1196838\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-frls8" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.104158 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rcg4j\" (UniqueName: \"kubernetes.io/projected/0776ae85-c7b7-4d22-b1fb-40b56785205a-kube-api-access-rcg4j\") pod \"etcd-operator-b45778765-k5c2k\" (UID: \"0776ae85-c7b7-4d22-b1fb-40b56785205a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-k5c2k" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.104264 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1c517779-505a-4371-9cd7-9537c2fd8763-webhook-cert\") pod \"packageserver-d55dfcdfc-2dxsq\" (UID: \"1c517779-505a-4371-9cd7-9537c2fd8763\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2dxsq" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.104064 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/55cac199-8203-40b3-8a4f-c571017d998b-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-wnfzj\" (UID: \"55cac199-8203-40b3-8a4f-c571017d998b\") " pod="openshift-authentication/oauth-openshift-558db77b4-wnfzj" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.104643 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/75fa6283-acd8-41bc-9803-2f55119f6828-cni-sysctl-allowlist\") pod \"cni-sysctl-allowlist-ds-gcnts\" (UID: \"75fa6283-acd8-41bc-9803-2f55119f6828\") " pod="openshift-multus/cni-sysctl-allowlist-ds-gcnts" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.104760 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/55cac199-8203-40b3-8a4f-c571017d998b-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-wnfzj\" (UID: \"55cac199-8203-40b3-8a4f-c571017d998b\") " pod="openshift-authentication/oauth-openshift-558db77b4-wnfzj" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.104869 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qwgcd\" (UniqueName: \"kubernetes.io/projected/f8d25be9-c80c-4057-b84f-7e78598d2e75-kube-api-access-qwgcd\") pod \"package-server-manager-789f6589d5-sszl6\" (UID: \"f8d25be9-c80c-4057-b84f-7e78598d2e75\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-sszl6" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.105004 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xvst4\" (UniqueName: \"kubernetes.io/projected/9eb8d5d8-3ad7-41d1-9bb2-2d6865619967-kube-api-access-xvst4\") pod \"marketplace-operator-79b997595-77svl\" (UID: \"9eb8d5d8-3ad7-41d1-9bb2-2d6865619967\") " pod="openshift-marketplace/marketplace-operator-79b997595-77svl" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.104657 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7dc5fba4-2ba8-4fa2-a76f-0ad8d1196838-audit-policies\") pod \"apiserver-7bbb656c7d-frls8\" (UID: \"7dc5fba4-2ba8-4fa2-a76f-0ad8d1196838\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-frls8" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.104813 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/7dc5fba4-2ba8-4fa2-a76f-0ad8d1196838-etcd-client\") pod \"apiserver-7bbb656c7d-frls8\" (UID: \"7dc5fba4-2ba8-4fa2-a76f-0ad8d1196838\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-frls8" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.105264 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7dc5fba4-2ba8-4fa2-a76f-0ad8d1196838-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-frls8\" (UID: \"7dc5fba4-2ba8-4fa2-a76f-0ad8d1196838\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-frls8" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.105347 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/55cac199-8203-40b3-8a4f-c571017d998b-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-wnfzj\" (UID: \"55cac199-8203-40b3-8a4f-c571017d998b\") " pod="openshift-authentication/oauth-openshift-558db77b4-wnfzj" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.105361 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/94294737-067b-4a4f-b59b-f9f987fa9127-etcd-serving-ca\") pod \"apiserver-76f77b778f-vh974\" (UID: \"94294737-067b-4a4f-b59b-f9f987fa9127\") " pod="openshift-apiserver/apiserver-76f77b778f-vh974" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.105456 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/f8d25be9-c80c-4057-b84f-7e78598d2e75-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-sszl6\" (UID: \"f8d25be9-c80c-4057-b84f-7e78598d2e75\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-sszl6" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.105490 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/10112213-cdd9-4bab-a7d5-bc7a3fab5a48-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-fr4zx\" (UID: \"10112213-cdd9-4bab-a7d5-bc7a3fab5a48\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-fr4zx" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.105532 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-smhnq\" (UniqueName: \"kubernetes.io/projected/78d69d0e-bfac-4582-b339-b2e585de5eff-kube-api-access-smhnq\") pod \"machine-config-operator-74547568cd-4fbj7\" (UID: \"78d69d0e-bfac-4582-b339-b2e585de5eff\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4fbj7" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.105565 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/31bd2041-6cb8-4eb6-bb22-5efe6c6009ee-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-2m4cb\" (UID: \"31bd2041-6cb8-4eb6-bb22-5efe6c6009ee\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2m4cb" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.105609 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/c562523e-6eb7-4c72-8ba2-f968f171d387-node-bootstrap-token\") pod \"machine-config-server-qh9np\" (UID: \"c562523e-6eb7-4c72-8ba2-f968f171d387\") " pod="openshift-machine-config-operator/machine-config-server-qh9np" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.105652 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kznfz\" (UniqueName: \"kubernetes.io/projected/ac3d1daa-cb6d-41fc-9798-e687d9b2a21d-kube-api-access-kznfz\") pod \"collect-profiles-29396625-nrhcv\" (UID: \"ac3d1daa-cb6d-41fc-9798-e687d9b2a21d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396625-nrhcv" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.105686 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cb90f2d2-3321-4118-a449-fdc9343b23e1-trusted-ca\") pod \"ingress-operator-5b745b69d9-bsrcz\" (UID: \"cb90f2d2-3321-4118-a449-fdc9343b23e1\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-bsrcz" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.105825 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/31bd2041-6cb8-4eb6-bb22-5efe6c6009ee-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-2m4cb\" (UID: \"31bd2041-6cb8-4eb6-bb22-5efe6c6009ee\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2m4cb" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.105682 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7dc5fba4-2ba8-4fa2-a76f-0ad8d1196838-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-frls8\" (UID: \"7dc5fba4-2ba8-4fa2-a76f-0ad8d1196838\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-frls8" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.105869 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/352693a0-531b-4017-9582-5424c4691795-stats-auth\") pod \"router-default-5444994796-x5b9n\" (UID: \"352693a0-531b-4017-9582-5424c4691795\") " pod="openshift-ingress/router-default-5444994796-x5b9n" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.105896 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/55cac199-8203-40b3-8a4f-c571017d998b-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-wnfzj\" (UID: \"55cac199-8203-40b3-8a4f-c571017d998b\") " pod="openshift-authentication/oauth-openshift-558db77b4-wnfzj" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.105923 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/0776ae85-c7b7-4d22-b1fb-40b56785205a-etcd-client\") pod \"etcd-operator-b45778765-k5c2k\" (UID: \"0776ae85-c7b7-4d22-b1fb-40b56785205a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-k5c2k" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.106000 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/d58d000b-3f2c-44cb-90c5-8023e9f4f5db-csi-data-dir\") pod \"csi-hostpathplugin-n27w4\" (UID: \"d58d000b-3f2c-44cb-90c5-8023e9f4f5db\") " pod="hostpath-provisioner/csi-hostpathplugin-n27w4" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.106022 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f2c80a41-3ceb-4420-a46e-566ff4a28dcb-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-zl87n\" (UID: \"f2c80a41-3ceb-4420-a46e-566ff4a28dcb\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zl87n" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.106058 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9eb8d5d8-3ad7-41d1-9bb2-2d6865619967-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-77svl\" (UID: \"9eb8d5d8-3ad7-41d1-9bb2-2d6865619967\") " pod="openshift-marketplace/marketplace-operator-79b997595-77svl" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.106091 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-42kgf\" (UniqueName: \"kubernetes.io/projected/cb90f2d2-3321-4118-a449-fdc9343b23e1-kube-api-access-42kgf\") pod \"ingress-operator-5b745b69d9-bsrcz\" (UID: \"cb90f2d2-3321-4118-a449-fdc9343b23e1\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-bsrcz" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.106116 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/0776ae85-c7b7-4d22-b1fb-40b56785205a-etcd-service-ca\") pod \"etcd-operator-b45778765-k5c2k\" (UID: \"0776ae85-c7b7-4d22-b1fb-40b56785205a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-k5c2k" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.106149 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1c517779-505a-4371-9cd7-9537c2fd8763-apiservice-cert\") pod \"packageserver-d55dfcdfc-2dxsq\" (UID: \"1c517779-505a-4371-9cd7-9537c2fd8763\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2dxsq" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.106198 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d32dc4fd-3e62-402c-a1d9-8f2c83f2e334-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-tnmrh\" (UID: \"d32dc4fd-3e62-402c-a1d9-8f2c83f2e334\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-tnmrh" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.106235 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/7dc5fba4-2ba8-4fa2-a76f-0ad8d1196838-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-frls8\" (UID: \"7dc5fba4-2ba8-4fa2-a76f-0ad8d1196838\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-frls8" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.106685 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cb90f2d2-3321-4118-a449-fdc9343b23e1-trusted-ca\") pod \"ingress-operator-5b745b69d9-bsrcz\" (UID: \"cb90f2d2-3321-4118-a449-fdc9343b23e1\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-bsrcz" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.106792 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/55cac199-8203-40b3-8a4f-c571017d998b-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-wnfzj\" (UID: \"55cac199-8203-40b3-8a4f-c571017d998b\") " pod="openshift-authentication/oauth-openshift-558db77b4-wnfzj" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.106889 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/7dc5fba4-2ba8-4fa2-a76f-0ad8d1196838-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-frls8\" (UID: \"7dc5fba4-2ba8-4fa2-a76f-0ad8d1196838\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-frls8" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.107023 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/cb9c37ae-6151-452f-b467-6b7d14bbb2f8-profile-collector-cert\") pod \"catalog-operator-68c6474976-h6f5v\" (UID: \"cb9c37ae-6151-452f-b467-6b7d14bbb2f8\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-h6f5v" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.107105 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bgr4n\" (UniqueName: \"kubernetes.io/projected/94294737-067b-4a4f-b59b-f9f987fa9127-kube-api-access-bgr4n\") pod \"apiserver-76f77b778f-vh974\" (UID: \"94294737-067b-4a4f-b59b-f9f987fa9127\") " pod="openshift-apiserver/apiserver-76f77b778f-vh974" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.107179 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/78d69d0e-bfac-4582-b339-b2e585de5eff-images\") pod \"machine-config-operator-74547568cd-4fbj7\" (UID: \"78d69d0e-bfac-4582-b339-b2e585de5eff\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4fbj7" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.107214 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/d58d000b-3f2c-44cb-90c5-8023e9f4f5db-mountpoint-dir\") pod \"csi-hostpathplugin-n27w4\" (UID: \"d58d000b-3f2c-44cb-90c5-8023e9f4f5db\") " pod="hostpath-provisioner/csi-hostpathplugin-n27w4" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.107245 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/cb9c37ae-6151-452f-b467-6b7d14bbb2f8-srv-cert\") pod \"catalog-operator-68c6474976-h6f5v\" (UID: \"cb9c37ae-6151-452f-b467-6b7d14bbb2f8\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-h6f5v" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.107265 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/065af0aa-e76c-4eab-9a85-d43956522aca-signing-key\") pod \"service-ca-9c57cc56f-whfpb\" (UID: \"065af0aa-e76c-4eab-9a85-d43956522aca\") " pod="openshift-service-ca/service-ca-9c57cc56f-whfpb" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.107284 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dvrrf\" (UniqueName: \"kubernetes.io/projected/7f8f3dfa-5fc0-4642-96c0-22b6f3081676-kube-api-access-dvrrf\") pod \"kube-storage-version-migrator-operator-b67b599dd-d4l6q\" (UID: \"7f8f3dfa-5fc0-4642-96c0-22b6f3081676\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-d4l6q" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.107305 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1507f6ee-6a14-4d67-98b3-1e151b7949cc-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-kdt8l\" (UID: \"1507f6ee-6a14-4d67-98b3-1e151b7949cc\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-kdt8l" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.107321 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n6hh2\" (UniqueName: \"kubernetes.io/projected/352693a0-531b-4017-9582-5424c4691795-kube-api-access-n6hh2\") pod \"router-default-5444994796-x5b9n\" (UID: \"352693a0-531b-4017-9582-5424c4691795\") " pod="openshift-ingress/router-default-5444994796-x5b9n" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.107353 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/cb90f2d2-3321-4118-a449-fdc9343b23e1-metrics-tls\") pod \"ingress-operator-5b745b69d9-bsrcz\" (UID: \"cb90f2d2-3321-4118-a449-fdc9343b23e1\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-bsrcz" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.107371 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/94294737-067b-4a4f-b59b-f9f987fa9127-trusted-ca-bundle\") pod \"apiserver-76f77b778f-vh974\" (UID: \"94294737-067b-4a4f-b59b-f9f987fa9127\") " pod="openshift-apiserver/apiserver-76f77b778f-vh974" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.107391 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1507f6ee-6a14-4d67-98b3-1e151b7949cc-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-kdt8l\" (UID: \"1507f6ee-6a14-4d67-98b3-1e151b7949cc\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-kdt8l" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.107408 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/9eb8d5d8-3ad7-41d1-9bb2-2d6865619967-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-77svl\" (UID: \"9eb8d5d8-3ad7-41d1-9bb2-2d6865619967\") " pod="openshift-marketplace/marketplace-operator-79b997595-77svl" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.107446 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/78d69d0e-bfac-4582-b339-b2e585de5eff-auth-proxy-config\") pod \"machine-config-operator-74547568cd-4fbj7\" (UID: \"78d69d0e-bfac-4582-b339-b2e585de5eff\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4fbj7" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.107464 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d32dc4fd-3e62-402c-a1d9-8f2c83f2e334-config\") pod \"kube-controller-manager-operator-78b949d7b-tnmrh\" (UID: \"d32dc4fd-3e62-402c-a1d9-8f2c83f2e334\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-tnmrh" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.107485 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/1c517779-505a-4371-9cd7-9537c2fd8763-tmpfs\") pod \"packageserver-d55dfcdfc-2dxsq\" (UID: \"1c517779-505a-4371-9cd7-9537c2fd8763\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2dxsq" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.107509 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7dc5fba4-2ba8-4fa2-a76f-0ad8d1196838-serving-cert\") pod \"apiserver-7bbb656c7d-frls8\" (UID: \"7dc5fba4-2ba8-4fa2-a76f-0ad8d1196838\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-frls8" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.107531 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/55cac199-8203-40b3-8a4f-c571017d998b-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-wnfzj\" (UID: \"55cac199-8203-40b3-8a4f-c571017d998b\") " pod="openshift-authentication/oauth-openshift-558db77b4-wnfzj" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.107557 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ac3d1daa-cb6d-41fc-9798-e687d9b2a21d-secret-volume\") pod \"collect-profiles-29396625-nrhcv\" (UID: \"ac3d1daa-cb6d-41fc-9798-e687d9b2a21d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396625-nrhcv" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.107580 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/55cac199-8203-40b3-8a4f-c571017d998b-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-wnfzj\" (UID: \"55cac199-8203-40b3-8a4f-c571017d998b\") " pod="openshift-authentication/oauth-openshift-558db77b4-wnfzj" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.107622 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vj4w5\" (UniqueName: \"kubernetes.io/projected/dd618778-dc82-4797-b737-14231eefea15-kube-api-access-vj4w5\") pod \"service-ca-operator-777779d784-fg7wj\" (UID: \"dd618778-dc82-4797-b737-14231eefea15\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-fg7wj" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.107645 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/dfa022a0-f86e-4f4b-99f5-22b84e9d6558-srv-cert\") pod \"olm-operator-6b444d44fb-9n9mc\" (UID: \"dfa022a0-f86e-4f4b-99f5-22b84e9d6558\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9n9mc" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.107685 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8b4tq\" (UniqueName: \"kubernetes.io/projected/10112213-cdd9-4bab-a7d5-bc7a3fab5a48-kube-api-access-8b4tq\") pod \"control-plane-machine-set-operator-78cbb6b69f-fr4zx\" (UID: \"10112213-cdd9-4bab-a7d5-bc7a3fab5a48\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-fr4zx" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.107709 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/7dc5fba4-2ba8-4fa2-a76f-0ad8d1196838-encryption-config\") pod \"apiserver-7bbb656c7d-frls8\" (UID: \"7dc5fba4-2ba8-4fa2-a76f-0ad8d1196838\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-frls8" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.107731 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/55cac199-8203-40b3-8a4f-c571017d998b-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-wnfzj\" (UID: \"55cac199-8203-40b3-8a4f-c571017d998b\") " pod="openshift-authentication/oauth-openshift-558db77b4-wnfzj" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.107782 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/f2c80a41-3ceb-4420-a46e-566ff4a28dcb-proxy-tls\") pod \"machine-config-controller-84d6567774-zl87n\" (UID: \"f2c80a41-3ceb-4420-a46e-566ff4a28dcb\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zl87n" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.107810 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/55cac199-8203-40b3-8a4f-c571017d998b-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-wnfzj\" (UID: \"55cac199-8203-40b3-8a4f-c571017d998b\") " pod="openshift-authentication/oauth-openshift-558db77b4-wnfzj" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.107851 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/94294737-067b-4a4f-b59b-f9f987fa9127-config\") pod \"apiserver-76f77b778f-vh974\" (UID: \"94294737-067b-4a4f-b59b-f9f987fa9127\") " pod="openshift-apiserver/apiserver-76f77b778f-vh974" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.107870 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/94294737-067b-4a4f-b59b-f9f987fa9127-node-pullsecrets\") pod \"apiserver-76f77b778f-vh974\" (UID: \"94294737-067b-4a4f-b59b-f9f987fa9127\") " pod="openshift-apiserver/apiserver-76f77b778f-vh974" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.107888 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9vb7l\" (UniqueName: \"kubernetes.io/projected/f2c80a41-3ceb-4420-a46e-566ff4a28dcb-kube-api-access-9vb7l\") pod \"machine-config-controller-84d6567774-zl87n\" (UID: \"f2c80a41-3ceb-4420-a46e-566ff4a28dcb\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zl87n" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.107906 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7f8f3dfa-5fc0-4642-96c0-22b6f3081676-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-d4l6q\" (UID: \"7f8f3dfa-5fc0-4642-96c0-22b6f3081676\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-d4l6q" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.107925 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ac3d1daa-cb6d-41fc-9798-e687d9b2a21d-config-volume\") pod \"collect-profiles-29396625-nrhcv\" (UID: \"ac3d1daa-cb6d-41fc-9798-e687d9b2a21d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396625-nrhcv" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.107994 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/d58d000b-3f2c-44cb-90c5-8023e9f4f5db-socket-dir\") pod \"csi-hostpathplugin-n27w4\" (UID: \"d58d000b-3f2c-44cb-90c5-8023e9f4f5db\") " pod="hostpath-provisioner/csi-hostpathplugin-n27w4" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.108549 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/dfa022a0-f86e-4f4b-99f5-22b84e9d6558-profile-collector-cert\") pod \"olm-operator-6b444d44fb-9n9mc\" (UID: \"dfa022a0-f86e-4f4b-99f5-22b84e9d6558\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9n9mc" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.108656 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/78d69d0e-bfac-4582-b339-b2e585de5eff-auth-proxy-config\") pod \"machine-config-operator-74547568cd-4fbj7\" (UID: \"78d69d0e-bfac-4582-b339-b2e585de5eff\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4fbj7" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.108694 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1507f6ee-6a14-4d67-98b3-1e151b7949cc-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-kdt8l\" (UID: \"1507f6ee-6a14-4d67-98b3-1e151b7949cc\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-kdt8l" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.108795 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/78d69d0e-bfac-4582-b339-b2e585de5eff-images\") pod \"machine-config-operator-74547568cd-4fbj7\" (UID: \"78d69d0e-bfac-4582-b339-b2e585de5eff\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4fbj7" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.108827 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d086acbe-d982-4880-9538-04a184cb4148-config\") pod \"machine-api-operator-5694c8668f-spnzw\" (UID: \"d086acbe-d982-4880-9538-04a184cb4148\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-spnzw" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.108849 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/55cac199-8203-40b3-8a4f-c571017d998b-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-wnfzj\" (UID: \"55cac199-8203-40b3-8a4f-c571017d998b\") " pod="openshift-authentication/oauth-openshift-558db77b4-wnfzj" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.108867 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/75fa6283-acd8-41bc-9803-2f55119f6828-ready\") pod \"cni-sysctl-allowlist-ds-gcnts\" (UID: \"75fa6283-acd8-41bc-9803-2f55119f6828\") " pod="openshift-multus/cni-sysctl-allowlist-ds-gcnts" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.108887 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/94294737-067b-4a4f-b59b-f9f987fa9127-node-pullsecrets\") pod \"apiserver-76f77b778f-vh974\" (UID: \"94294737-067b-4a4f-b59b-f9f987fa9127\") " pod="openshift-apiserver/apiserver-76f77b778f-vh974" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.108900 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/94294737-067b-4a4f-b59b-f9f987fa9127-image-import-ca\") pod \"apiserver-76f77b778f-vh974\" (UID: \"94294737-067b-4a4f-b59b-f9f987fa9127\") " pod="openshift-apiserver/apiserver-76f77b778f-vh974" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.109308 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/55cac199-8203-40b3-8a4f-c571017d998b-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-wnfzj\" (UID: \"55cac199-8203-40b3-8a4f-c571017d998b\") " pod="openshift-authentication/oauth-openshift-558db77b4-wnfzj" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.109432 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/55cac199-8203-40b3-8a4f-c571017d998b-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-wnfzj\" (UID: \"55cac199-8203-40b3-8a4f-c571017d998b\") " pod="openshift-authentication/oauth-openshift-558db77b4-wnfzj" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.109464 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/55cac199-8203-40b3-8a4f-c571017d998b-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-wnfzj\" (UID: \"55cac199-8203-40b3-8a4f-c571017d998b\") " pod="openshift-authentication/oauth-openshift-558db77b4-wnfzj" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.109874 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/94294737-067b-4a4f-b59b-f9f987fa9127-etcd-serving-ca\") pod \"apiserver-76f77b778f-vh974\" (UID: \"94294737-067b-4a4f-b59b-f9f987fa9127\") " pod="openshift-apiserver/apiserver-76f77b778f-vh974" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.110247 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ac3d1daa-cb6d-41fc-9798-e687d9b2a21d-config-volume\") pod \"collect-profiles-29396625-nrhcv\" (UID: \"ac3d1daa-cb6d-41fc-9798-e687d9b2a21d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396625-nrhcv" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.110463 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g7qk9\" (UniqueName: \"kubernetes.io/projected/d58d000b-3f2c-44cb-90c5-8023e9f4f5db-kube-api-access-g7qk9\") pod \"csi-hostpathplugin-n27w4\" (UID: \"d58d000b-3f2c-44cb-90c5-8023e9f4f5db\") " pod="hostpath-provisioner/csi-hostpathplugin-n27w4" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.110513 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zgzpt\" (UniqueName: \"kubernetes.io/projected/1d0ebe8e-6cea-43d3-9f15-b67c4cbc29e7-kube-api-access-zgzpt\") pod \"ingress-canary-qc925\" (UID: \"1d0ebe8e-6cea-43d3-9f15-b67c4cbc29e7\") " pod="openshift-ingress-canary/ingress-canary-qc925" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.110539 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/55cac199-8203-40b3-8a4f-c571017d998b-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-wnfzj\" (UID: \"55cac199-8203-40b3-8a4f-c571017d998b\") " pod="openshift-authentication/oauth-openshift-558db77b4-wnfzj" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.110638 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/94294737-067b-4a4f-b59b-f9f987fa9127-audit-dir\") pod \"apiserver-76f77b778f-vh974\" (UID: \"94294737-067b-4a4f-b59b-f9f987fa9127\") " pod="openshift-apiserver/apiserver-76f77b778f-vh974" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.110672 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/94294737-067b-4a4f-b59b-f9f987fa9127-audit-dir\") pod \"apiserver-76f77b778f-vh974\" (UID: \"94294737-067b-4a4f-b59b-f9f987fa9127\") " pod="openshift-apiserver/apiserver-76f77b778f-vh974" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.110687 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cgptp\" (UniqueName: \"kubernetes.io/projected/065af0aa-e76c-4eab-9a85-d43956522aca-kube-api-access-cgptp\") pod \"service-ca-9c57cc56f-whfpb\" (UID: \"065af0aa-e76c-4eab-9a85-d43956522aca\") " pod="openshift-service-ca/service-ca-9c57cc56f-whfpb" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.110722 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wdcz2\" (UniqueName: \"kubernetes.io/projected/55cac199-8203-40b3-8a4f-c571017d998b-kube-api-access-wdcz2\") pod \"oauth-openshift-558db77b4-wnfzj\" (UID: \"55cac199-8203-40b3-8a4f-c571017d998b\") " pod="openshift-authentication/oauth-openshift-558db77b4-wnfzj" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.110694 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/94294737-067b-4a4f-b59b-f9f987fa9127-trusted-ca-bundle\") pod \"apiserver-76f77b778f-vh974\" (UID: \"94294737-067b-4a4f-b59b-f9f987fa9127\") " pod="openshift-apiserver/apiserver-76f77b778f-vh974" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.110765 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dkw4c\" (UniqueName: \"kubernetes.io/projected/1c517779-505a-4371-9cd7-9537c2fd8763-kube-api-access-dkw4c\") pod \"packageserver-d55dfcdfc-2dxsq\" (UID: \"1c517779-505a-4371-9cd7-9537c2fd8763\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2dxsq" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.110828 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kppbq\" (UniqueName: \"kubernetes.io/projected/d086acbe-d982-4880-9538-04a184cb4148-kube-api-access-kppbq\") pod \"machine-api-operator-5694c8668f-spnzw\" (UID: \"d086acbe-d982-4880-9538-04a184cb4148\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-spnzw" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.110858 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s86jn\" (UniqueName: \"kubernetes.io/projected/3cfa3a64-3e7a-4336-9ae0-86822095a1b4-kube-api-access-s86jn\") pod \"multus-admission-controller-857f4d67dd-fb49f\" (UID: \"3cfa3a64-3e7a-4336-9ae0-86822095a1b4\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-fb49f" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.110884 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/d58d000b-3f2c-44cb-90c5-8023e9f4f5db-registration-dir\") pod \"csi-hostpathplugin-n27w4\" (UID: \"d58d000b-3f2c-44cb-90c5-8023e9f4f5db\") " pod="hostpath-provisioner/csi-hostpathplugin-n27w4" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.110912 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/cb90f2d2-3321-4118-a449-fdc9343b23e1-bound-sa-token\") pod \"ingress-operator-5b745b69d9-bsrcz\" (UID: \"cb90f2d2-3321-4118-a449-fdc9343b23e1\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-bsrcz" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.110936 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/31bd2041-6cb8-4eb6-bb22-5efe6c6009ee-config\") pod \"kube-apiserver-operator-766d6c64bb-2m4cb\" (UID: \"31bd2041-6cb8-4eb6-bb22-5efe6c6009ee\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2m4cb" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.110959 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0776ae85-c7b7-4d22-b1fb-40b56785205a-serving-cert\") pod \"etcd-operator-b45778765-k5c2k\" (UID: \"0776ae85-c7b7-4d22-b1fb-40b56785205a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-k5c2k" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.110983 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2mgz5\" (UniqueName: \"kubernetes.io/projected/7dc5fba4-2ba8-4fa2-a76f-0ad8d1196838-kube-api-access-2mgz5\") pod \"apiserver-7bbb656c7d-frls8\" (UID: \"7dc5fba4-2ba8-4fa2-a76f-0ad8d1196838\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-frls8" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.111033 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dd618778-dc82-4797-b737-14231eefea15-serving-cert\") pod \"service-ca-operator-777779d784-fg7wj\" (UID: \"dd618778-dc82-4797-b737-14231eefea15\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-fg7wj" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.111143 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/94294737-067b-4a4f-b59b-f9f987fa9127-serving-cert\") pod \"apiserver-76f77b778f-vh974\" (UID: \"94294737-067b-4a4f-b59b-f9f987fa9127\") " pod="openshift-apiserver/apiserver-76f77b778f-vh974" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.111153 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/94294737-067b-4a4f-b59b-f9f987fa9127-encryption-config\") pod \"apiserver-76f77b778f-vh974\" (UID: \"94294737-067b-4a4f-b59b-f9f987fa9127\") " pod="openshift-apiserver/apiserver-76f77b778f-vh974" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.111189 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/94294737-067b-4a4f-b59b-f9f987fa9127-etcd-client\") pod \"apiserver-76f77b778f-vh974\" (UID: \"94294737-067b-4a4f-b59b-f9f987fa9127\") " pod="openshift-apiserver/apiserver-76f77b778f-vh974" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.111217 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vzd9t\" (UniqueName: \"kubernetes.io/projected/cb9c37ae-6151-452f-b467-6b7d14bbb2f8-kube-api-access-vzd9t\") pod \"catalog-operator-68c6474976-h6f5v\" (UID: \"cb9c37ae-6151-452f-b467-6b7d14bbb2f8\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-h6f5v" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.111246 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/065af0aa-e76c-4eab-9a85-d43956522aca-signing-cabundle\") pod \"service-ca-9c57cc56f-whfpb\" (UID: \"065af0aa-e76c-4eab-9a85-d43956522aca\") " pod="openshift-service-ca/service-ca-9c57cc56f-whfpb" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.111280 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0776ae85-c7b7-4d22-b1fb-40b56785205a-config\") pod \"etcd-operator-b45778765-k5c2k\" (UID: \"0776ae85-c7b7-4d22-b1fb-40b56785205a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-k5c2k" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.111306 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/0776ae85-c7b7-4d22-b1fb-40b56785205a-etcd-ca\") pod \"etcd-operator-b45778765-k5c2k\" (UID: \"0776ae85-c7b7-4d22-b1fb-40b56785205a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-k5c2k" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.111337 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/78d69d0e-bfac-4582-b339-b2e585de5eff-proxy-tls\") pod \"machine-config-operator-74547568cd-4fbj7\" (UID: \"78d69d0e-bfac-4582-b339-b2e585de5eff\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4fbj7" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.111398 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1507f6ee-6a14-4d67-98b3-1e151b7949cc-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-kdt8l\" (UID: \"1507f6ee-6a14-4d67-98b3-1e151b7949cc\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-kdt8l" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.111445 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/352693a0-531b-4017-9582-5424c4691795-default-certificate\") pod \"router-default-5444994796-x5b9n\" (UID: \"352693a0-531b-4017-9582-5424c4691795\") " pod="openshift-ingress/router-default-5444994796-x5b9n" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.111848 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/55cac199-8203-40b3-8a4f-c571017d998b-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-wnfzj\" (UID: \"55cac199-8203-40b3-8a4f-c571017d998b\") " pod="openshift-authentication/oauth-openshift-558db77b4-wnfzj" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.112013 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/94294737-067b-4a4f-b59b-f9f987fa9127-image-import-ca\") pod \"apiserver-76f77b778f-vh974\" (UID: \"94294737-067b-4a4f-b59b-f9f987fa9127\") " pod="openshift-apiserver/apiserver-76f77b778f-vh974" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.112154 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ac3d1daa-cb6d-41fc-9798-e687d9b2a21d-secret-volume\") pod \"collect-profiles-29396625-nrhcv\" (UID: \"ac3d1daa-cb6d-41fc-9798-e687d9b2a21d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396625-nrhcv" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.112304 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/94294737-067b-4a4f-b59b-f9f987fa9127-config\") pod \"apiserver-76f77b778f-vh974\" (UID: \"94294737-067b-4a4f-b59b-f9f987fa9127\") " pod="openshift-apiserver/apiserver-76f77b778f-vh974" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.113551 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7dc5fba4-2ba8-4fa2-a76f-0ad8d1196838-serving-cert\") pod \"apiserver-7bbb656c7d-frls8\" (UID: \"7dc5fba4-2ba8-4fa2-a76f-0ad8d1196838\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-frls8" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.113728 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/55cac199-8203-40b3-8a4f-c571017d998b-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-wnfzj\" (UID: \"55cac199-8203-40b3-8a4f-c571017d998b\") " pod="openshift-authentication/oauth-openshift-558db77b4-wnfzj" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.115336 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/7dc5fba4-2ba8-4fa2-a76f-0ad8d1196838-encryption-config\") pod \"apiserver-7bbb656c7d-frls8\" (UID: \"7dc5fba4-2ba8-4fa2-a76f-0ad8d1196838\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-frls8" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.114157 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/94294737-067b-4a4f-b59b-f9f987fa9127-encryption-config\") pod \"apiserver-76f77b778f-vh974\" (UID: \"94294737-067b-4a4f-b59b-f9f987fa9127\") " pod="openshift-apiserver/apiserver-76f77b778f-vh974" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.114407 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/55cac199-8203-40b3-8a4f-c571017d998b-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-wnfzj\" (UID: \"55cac199-8203-40b3-8a4f-c571017d998b\") " pod="openshift-authentication/oauth-openshift-558db77b4-wnfzj" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.114325 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/55cac199-8203-40b3-8a4f-c571017d998b-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-wnfzj\" (UID: \"55cac199-8203-40b3-8a4f-c571017d998b\") " pod="openshift-authentication/oauth-openshift-558db77b4-wnfzj" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.115154 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1507f6ee-6a14-4d67-98b3-1e151b7949cc-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-kdt8l\" (UID: \"1507f6ee-6a14-4d67-98b3-1e151b7949cc\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-kdt8l" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.115161 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/94294737-067b-4a4f-b59b-f9f987fa9127-etcd-client\") pod \"apiserver-76f77b778f-vh974\" (UID: \"94294737-067b-4a4f-b59b-f9f987fa9127\") " pod="openshift-apiserver/apiserver-76f77b778f-vh974" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.115573 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/cb90f2d2-3321-4118-a449-fdc9343b23e1-metrics-tls\") pod \"ingress-operator-5b745b69d9-bsrcz\" (UID: \"cb90f2d2-3321-4118-a449-fdc9343b23e1\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-bsrcz" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.115964 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/78d69d0e-bfac-4582-b339-b2e585de5eff-proxy-tls\") pod \"machine-config-operator-74547568cd-4fbj7\" (UID: \"78d69d0e-bfac-4582-b339-b2e585de5eff\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4fbj7" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.121643 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.142271 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.161232 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.181452 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.186412 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/d086acbe-d982-4880-9538-04a184cb4148-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-spnzw\" (UID: \"d086acbe-d982-4880-9538-04a184cb4148\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-spnzw" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.201331 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.211227 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d086acbe-d982-4880-9538-04a184cb4148-config\") pod \"machine-api-operator-5694c8668f-spnzw\" (UID: \"d086acbe-d982-4880-9538-04a184cb4148\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-spnzw" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.212105 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/d58d000b-3f2c-44cb-90c5-8023e9f4f5db-socket-dir\") pod \"csi-hostpathplugin-n27w4\" (UID: \"d58d000b-3f2c-44cb-90c5-8023e9f4f5db\") " pod="hostpath-provisioner/csi-hostpathplugin-n27w4" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.212140 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9vb7l\" (UniqueName: \"kubernetes.io/projected/f2c80a41-3ceb-4420-a46e-566ff4a28dcb-kube-api-access-9vb7l\") pod \"machine-config-controller-84d6567774-zl87n\" (UID: \"f2c80a41-3ceb-4420-a46e-566ff4a28dcb\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zl87n" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.212178 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7f8f3dfa-5fc0-4642-96c0-22b6f3081676-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-d4l6q\" (UID: \"7f8f3dfa-5fc0-4642-96c0-22b6f3081676\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-d4l6q" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.212198 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/75fa6283-acd8-41bc-9803-2f55119f6828-ready\") pod \"cni-sysctl-allowlist-ds-gcnts\" (UID: \"75fa6283-acd8-41bc-9803-2f55119f6828\") " pod="openshift-multus/cni-sysctl-allowlist-ds-gcnts" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.212232 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g7qk9\" (UniqueName: \"kubernetes.io/projected/d58d000b-3f2c-44cb-90c5-8023e9f4f5db-kube-api-access-g7qk9\") pod \"csi-hostpathplugin-n27w4\" (UID: \"d58d000b-3f2c-44cb-90c5-8023e9f4f5db\") " pod="hostpath-provisioner/csi-hostpathplugin-n27w4" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.212270 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zgzpt\" (UniqueName: \"kubernetes.io/projected/1d0ebe8e-6cea-43d3-9f15-b67c4cbc29e7-kube-api-access-zgzpt\") pod \"ingress-canary-qc925\" (UID: \"1d0ebe8e-6cea-43d3-9f15-b67c4cbc29e7\") " pod="openshift-ingress-canary/ingress-canary-qc925" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.212905 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cgptp\" (UniqueName: \"kubernetes.io/projected/065af0aa-e76c-4eab-9a85-d43956522aca-kube-api-access-cgptp\") pod \"service-ca-9c57cc56f-whfpb\" (UID: \"065af0aa-e76c-4eab-9a85-d43956522aca\") " pod="openshift-service-ca/service-ca-9c57cc56f-whfpb" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.213033 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dkw4c\" (UniqueName: \"kubernetes.io/projected/1c517779-505a-4371-9cd7-9537c2fd8763-kube-api-access-dkw4c\") pod \"packageserver-d55dfcdfc-2dxsq\" (UID: \"1c517779-505a-4371-9cd7-9537c2fd8763\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2dxsq" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.213323 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/75fa6283-acd8-41bc-9803-2f55119f6828-ready\") pod \"cni-sysctl-allowlist-ds-gcnts\" (UID: \"75fa6283-acd8-41bc-9803-2f55119f6828\") " pod="openshift-multus/cni-sysctl-allowlist-ds-gcnts" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.213381 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/d58d000b-3f2c-44cb-90c5-8023e9f4f5db-registration-dir\") pod \"csi-hostpathplugin-n27w4\" (UID: \"d58d000b-3f2c-44cb-90c5-8023e9f4f5db\") " pod="hostpath-provisioner/csi-hostpathplugin-n27w4" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.213444 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0776ae85-c7b7-4d22-b1fb-40b56785205a-serving-cert\") pod \"etcd-operator-b45778765-k5c2k\" (UID: \"0776ae85-c7b7-4d22-b1fb-40b56785205a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-k5c2k" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.213479 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/d58d000b-3f2c-44cb-90c5-8023e9f4f5db-socket-dir\") pod \"csi-hostpathplugin-n27w4\" (UID: \"d58d000b-3f2c-44cb-90c5-8023e9f4f5db\") " pod="hostpath-provisioner/csi-hostpathplugin-n27w4" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.213504 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/065af0aa-e76c-4eab-9a85-d43956522aca-signing-cabundle\") pod \"service-ca-9c57cc56f-whfpb\" (UID: \"065af0aa-e76c-4eab-9a85-d43956522aca\") " pod="openshift-service-ca/service-ca-9c57cc56f-whfpb" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.213582 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0776ae85-c7b7-4d22-b1fb-40b56785205a-config\") pod \"etcd-operator-b45778765-k5c2k\" (UID: \"0776ae85-c7b7-4d22-b1fb-40b56785205a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-k5c2k" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.213644 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/0776ae85-c7b7-4d22-b1fb-40b56785205a-etcd-ca\") pod \"etcd-operator-b45778765-k5c2k\" (UID: \"0776ae85-c7b7-4d22-b1fb-40b56785205a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-k5c2k" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.213589 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/d58d000b-3f2c-44cb-90c5-8023e9f4f5db-registration-dir\") pod \"csi-hostpathplugin-n27w4\" (UID: \"d58d000b-3f2c-44cb-90c5-8023e9f4f5db\") " pod="hostpath-provisioner/csi-hostpathplugin-n27w4" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.213848 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/75fa6283-acd8-41bc-9803-2f55119f6828-tuning-conf-dir\") pod \"cni-sysctl-allowlist-ds-gcnts\" (UID: \"75fa6283-acd8-41bc-9803-2f55119f6828\") " pod="openshift-multus/cni-sysctl-allowlist-ds-gcnts" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.213895 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/c562523e-6eb7-4c72-8ba2-f968f171d387-certs\") pod \"machine-config-server-qh9np\" (UID: \"c562523e-6eb7-4c72-8ba2-f968f171d387\") " pod="openshift-machine-config-operator/machine-config-server-qh9np" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.213924 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xk2ns\" (UniqueName: \"kubernetes.io/projected/c562523e-6eb7-4c72-8ba2-f968f171d387-kube-api-access-xk2ns\") pod \"machine-config-server-qh9np\" (UID: \"c562523e-6eb7-4c72-8ba2-f968f171d387\") " pod="openshift-machine-config-operator/machine-config-server-qh9np" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.213979 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f8f3dfa-5fc0-4642-96c0-22b6f3081676-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-d4l6q\" (UID: \"7f8f3dfa-5fc0-4642-96c0-22b6f3081676\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-d4l6q" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.214057 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/d58d000b-3f2c-44cb-90c5-8023e9f4f5db-plugins-dir\") pod \"csi-hostpathplugin-n27w4\" (UID: \"d58d000b-3f2c-44cb-90c5-8023e9f4f5db\") " pod="hostpath-provisioner/csi-hostpathplugin-n27w4" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.214096 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6r7x6\" (UniqueName: \"kubernetes.io/projected/75fa6283-acd8-41bc-9803-2f55119f6828-kube-api-access-6r7x6\") pod \"cni-sysctl-allowlist-ds-gcnts\" (UID: \"75fa6283-acd8-41bc-9803-2f55119f6828\") " pod="openshift-multus/cni-sysctl-allowlist-ds-gcnts" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.214130 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1d0ebe8e-6cea-43d3-9f15-b67c4cbc29e7-cert\") pod \"ingress-canary-qc925\" (UID: \"1d0ebe8e-6cea-43d3-9f15-b67c4cbc29e7\") " pod="openshift-ingress-canary/ingress-canary-qc925" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.214171 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rcg4j\" (UniqueName: \"kubernetes.io/projected/0776ae85-c7b7-4d22-b1fb-40b56785205a-kube-api-access-rcg4j\") pod \"etcd-operator-b45778765-k5c2k\" (UID: \"0776ae85-c7b7-4d22-b1fb-40b56785205a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-k5c2k" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.214207 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1c517779-505a-4371-9cd7-9537c2fd8763-webhook-cert\") pod \"packageserver-d55dfcdfc-2dxsq\" (UID: \"1c517779-505a-4371-9cd7-9537c2fd8763\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2dxsq" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.214251 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/75fa6283-acd8-41bc-9803-2f55119f6828-cni-sysctl-allowlist\") pod \"cni-sysctl-allowlist-ds-gcnts\" (UID: \"75fa6283-acd8-41bc-9803-2f55119f6828\") " pod="openshift-multus/cni-sysctl-allowlist-ds-gcnts" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.214287 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qwgcd\" (UniqueName: \"kubernetes.io/projected/f8d25be9-c80c-4057-b84f-7e78598d2e75-kube-api-access-qwgcd\") pod \"package-server-manager-789f6589d5-sszl6\" (UID: \"f8d25be9-c80c-4057-b84f-7e78598d2e75\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-sszl6" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.214322 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xvst4\" (UniqueName: \"kubernetes.io/projected/9eb8d5d8-3ad7-41d1-9bb2-2d6865619967-kube-api-access-xvst4\") pod \"marketplace-operator-79b997595-77svl\" (UID: \"9eb8d5d8-3ad7-41d1-9bb2-2d6865619967\") " pod="openshift-marketplace/marketplace-operator-79b997595-77svl" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.214369 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/f8d25be9-c80c-4057-b84f-7e78598d2e75-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-sszl6\" (UID: \"f8d25be9-c80c-4057-b84f-7e78598d2e75\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-sszl6" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.214407 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/10112213-cdd9-4bab-a7d5-bc7a3fab5a48-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-fr4zx\" (UID: \"10112213-cdd9-4bab-a7d5-bc7a3fab5a48\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-fr4zx" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.214448 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/c562523e-6eb7-4c72-8ba2-f968f171d387-node-bootstrap-token\") pod \"machine-config-server-qh9np\" (UID: \"c562523e-6eb7-4c72-8ba2-f968f171d387\") " pod="openshift-machine-config-operator/machine-config-server-qh9np" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.214540 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/d58d000b-3f2c-44cb-90c5-8023e9f4f5db-csi-data-dir\") pod \"csi-hostpathplugin-n27w4\" (UID: \"d58d000b-3f2c-44cb-90c5-8023e9f4f5db\") " pod="hostpath-provisioner/csi-hostpathplugin-n27w4" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.214569 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/0776ae85-c7b7-4d22-b1fb-40b56785205a-etcd-client\") pod \"etcd-operator-b45778765-k5c2k\" (UID: \"0776ae85-c7b7-4d22-b1fb-40b56785205a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-k5c2k" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.214607 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/0776ae85-c7b7-4d22-b1fb-40b56785205a-etcd-service-ca\") pod \"etcd-operator-b45778765-k5c2k\" (UID: \"0776ae85-c7b7-4d22-b1fb-40b56785205a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-k5c2k" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.214641 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f2c80a41-3ceb-4420-a46e-566ff4a28dcb-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-zl87n\" (UID: \"f2c80a41-3ceb-4420-a46e-566ff4a28dcb\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zl87n" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.214668 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9eb8d5d8-3ad7-41d1-9bb2-2d6865619967-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-77svl\" (UID: \"9eb8d5d8-3ad7-41d1-9bb2-2d6865619967\") " pod="openshift-marketplace/marketplace-operator-79b997595-77svl" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.214722 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/75fa6283-acd8-41bc-9803-2f55119f6828-tuning-conf-dir\") pod \"cni-sysctl-allowlist-ds-gcnts\" (UID: \"75fa6283-acd8-41bc-9803-2f55119f6828\") " pod="openshift-multus/cni-sysctl-allowlist-ds-gcnts" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.214804 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1c517779-505a-4371-9cd7-9537c2fd8763-apiservice-cert\") pod \"packageserver-d55dfcdfc-2dxsq\" (UID: \"1c517779-505a-4371-9cd7-9537c2fd8763\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2dxsq" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.214963 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/d58d000b-3f2c-44cb-90c5-8023e9f4f5db-mountpoint-dir\") pod \"csi-hostpathplugin-n27w4\" (UID: \"d58d000b-3f2c-44cb-90c5-8023e9f4f5db\") " pod="hostpath-provisioner/csi-hostpathplugin-n27w4" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.214999 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/d58d000b-3f2c-44cb-90c5-8023e9f4f5db-csi-data-dir\") pod \"csi-hostpathplugin-n27w4\" (UID: \"d58d000b-3f2c-44cb-90c5-8023e9f4f5db\") " pod="hostpath-provisioner/csi-hostpathplugin-n27w4" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.215024 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/065af0aa-e76c-4eab-9a85-d43956522aca-signing-key\") pod \"service-ca-9c57cc56f-whfpb\" (UID: \"065af0aa-e76c-4eab-9a85-d43956522aca\") " pod="openshift-service-ca/service-ca-9c57cc56f-whfpb" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.215064 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dvrrf\" (UniqueName: \"kubernetes.io/projected/7f8f3dfa-5fc0-4642-96c0-22b6f3081676-kube-api-access-dvrrf\") pod \"kube-storage-version-migrator-operator-b67b599dd-d4l6q\" (UID: \"7f8f3dfa-5fc0-4642-96c0-22b6f3081676\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-d4l6q" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.215092 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/d58d000b-3f2c-44cb-90c5-8023e9f4f5db-plugins-dir\") pod \"csi-hostpathplugin-n27w4\" (UID: \"d58d000b-3f2c-44cb-90c5-8023e9f4f5db\") " pod="hostpath-provisioner/csi-hostpathplugin-n27w4" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.215137 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/9eb8d5d8-3ad7-41d1-9bb2-2d6865619967-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-77svl\" (UID: \"9eb8d5d8-3ad7-41d1-9bb2-2d6865619967\") " pod="openshift-marketplace/marketplace-operator-79b997595-77svl" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.215194 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/1c517779-505a-4371-9cd7-9537c2fd8763-tmpfs\") pod \"packageserver-d55dfcdfc-2dxsq\" (UID: \"1c517779-505a-4371-9cd7-9537c2fd8763\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2dxsq" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.215478 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8b4tq\" (UniqueName: \"kubernetes.io/projected/10112213-cdd9-4bab-a7d5-bc7a3fab5a48-kube-api-access-8b4tq\") pod \"control-plane-machine-set-operator-78cbb6b69f-fr4zx\" (UID: \"10112213-cdd9-4bab-a7d5-bc7a3fab5a48\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-fr4zx" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.215638 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/f2c80a41-3ceb-4420-a46e-566ff4a28dcb-proxy-tls\") pod \"machine-config-controller-84d6567774-zl87n\" (UID: \"f2c80a41-3ceb-4420-a46e-566ff4a28dcb\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zl87n" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.216045 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f2c80a41-3ceb-4420-a46e-566ff4a28dcb-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-zl87n\" (UID: \"f2c80a41-3ceb-4420-a46e-566ff4a28dcb\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zl87n" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.216104 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/1c517779-505a-4371-9cd7-9537c2fd8763-tmpfs\") pod \"packageserver-d55dfcdfc-2dxsq\" (UID: \"1c517779-505a-4371-9cd7-9537c2fd8763\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2dxsq" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.216104 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/d58d000b-3f2c-44cb-90c5-8023e9f4f5db-mountpoint-dir\") pod \"csi-hostpathplugin-n27w4\" (UID: \"d58d000b-3f2c-44cb-90c5-8023e9f4f5db\") " pod="hostpath-provisioner/csi-hostpathplugin-n27w4" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.221591 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.231376 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/d086acbe-d982-4880-9538-04a184cb4148-images\") pod \"machine-api-operator-5694c8668f-spnzw\" (UID: \"d086acbe-d982-4880-9538-04a184cb4148\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-spnzw" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.242247 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.261698 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.282520 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.294957 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/3cfa3a64-3e7a-4336-9ae0-86822095a1b4-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-fb49f\" (UID: \"3cfa3a64-3e7a-4336-9ae0-86822095a1b4\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-fb49f" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.302098 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.322111 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.327579 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d32dc4fd-3e62-402c-a1d9-8f2c83f2e334-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-tnmrh\" (UID: \"d32dc4fd-3e62-402c-a1d9-8f2c83f2e334\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-tnmrh" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.341647 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.350283 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d32dc4fd-3e62-402c-a1d9-8f2c83f2e334-config\") pod \"kube-controller-manager-operator-78b949d7b-tnmrh\" (UID: \"d32dc4fd-3e62-402c-a1d9-8f2c83f2e334\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-tnmrh" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.362143 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.381568 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.402633 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.421887 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.435146 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dd618778-dc82-4797-b737-14231eefea15-serving-cert\") pod \"service-ca-operator-777779d784-fg7wj\" (UID: \"dd618778-dc82-4797-b737-14231eefea15\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-fg7wj" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.441829 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.452540 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd618778-dc82-4797-b737-14231eefea15-config\") pod \"service-ca-operator-777779d784-fg7wj\" (UID: \"dd618778-dc82-4797-b737-14231eefea15\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-fg7wj" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.461665 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.483150 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.501296 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.523491 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.528727 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/31bd2041-6cb8-4eb6-bb22-5efe6c6009ee-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-2m4cb\" (UID: \"31bd2041-6cb8-4eb6-bb22-5efe6c6009ee\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2m4cb" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.540967 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.541651 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/31bd2041-6cb8-4eb6-bb22-5efe6c6009ee-config\") pod \"kube-apiserver-operator-766d6c64bb-2m4cb\" (UID: \"31bd2041-6cb8-4eb6-bb22-5efe6c6009ee\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2m4cb" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.561255 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.582214 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.594908 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/dfa022a0-f86e-4f4b-99f5-22b84e9d6558-srv-cert\") pod \"olm-operator-6b444d44fb-9n9mc\" (UID: \"dfa022a0-f86e-4f4b-99f5-22b84e9d6558\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9n9mc" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.601387 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.611083 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/cb9c37ae-6151-452f-b467-6b7d14bbb2f8-srv-cert\") pod \"catalog-operator-68c6474976-h6f5v\" (UID: \"cb9c37ae-6151-452f-b467-6b7d14bbb2f8\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-h6f5v" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.621983 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.635842 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/352693a0-531b-4017-9582-5424c4691795-default-certificate\") pod \"router-default-5444994796-x5b9n\" (UID: \"352693a0-531b-4017-9582-5424c4691795\") " pod="openshift-ingress/router-default-5444994796-x5b9n" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.641939 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.650171 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/352693a0-531b-4017-9582-5424c4691795-stats-auth\") pod \"router-default-5444994796-x5b9n\" (UID: \"352693a0-531b-4017-9582-5424c4691795\") " pod="openshift-ingress/router-default-5444994796-x5b9n" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.661688 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.674665 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/352693a0-531b-4017-9582-5424c4691795-metrics-certs\") pod \"router-default-5444994796-x5b9n\" (UID: \"352693a0-531b-4017-9582-5424c4691795\") " pod="openshift-ingress/router-default-5444994796-x5b9n" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.681086 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.701048 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.704097 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/352693a0-531b-4017-9582-5424c4691795-service-ca-bundle\") pod \"router-default-5444994796-x5b9n\" (UID: \"352693a0-531b-4017-9582-5424c4691795\") " pod="openshift-ingress/router-default-5444994796-x5b9n" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.721495 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.741598 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.761703 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.782013 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.801966 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.821383 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.830141 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/f2c80a41-3ceb-4420-a46e-566ff4a28dcb-proxy-tls\") pod \"machine-config-controller-84d6567774-zl87n\" (UID: \"f2c80a41-3ceb-4420-a46e-566ff4a28dcb\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zl87n" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.841418 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.862259 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.882079 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.901943 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.921632 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.940075 4789 request.go:700] Waited for 1.007045152s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-storage-version-migrator-operator/secrets?fieldSelector=metadata.name%3Dkube-storage-version-migrator-operator-dockercfg-2bh8d&limit=500&resourceVersion=0 Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.941918 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.961575 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.965102 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-q7rc4" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.965116 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.966208 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f8f3dfa-5fc0-4642-96c0-22b6f3081676-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-d4l6q\" (UID: \"7f8f3dfa-5fc0-4642-96c0-22b6f3081676\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-d4l6q" Nov 22 07:57:19 crc kubenswrapper[4789]: I1122 07:57:19.981249 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Nov 22 07:57:19 crc kubenswrapper[4789]: E1122 07:57:19.994841 4789 secret.go:188] Couldn't get secret openshift-controller-manager-operator/openshift-controller-manager-operator-serving-cert: failed to sync secret cache: timed out waiting for the condition Nov 22 07:57:19 crc kubenswrapper[4789]: E1122 07:57:19.994930 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ffef360f-8678-442d-9b27-b8de7646f1a3-serving-cert podName:ffef360f-8678-442d-9b27-b8de7646f1a3 nodeName:}" failed. No retries permitted until 2025-11-22 07:57:20.494909142 +0000 UTC m=+54.729309415 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/ffef360f-8678-442d-9b27-b8de7646f1a3-serving-cert") pod "openshift-controller-manager-operator-756b6f6bc6-t5d22" (UID: "ffef360f-8678-442d-9b27-b8de7646f1a3") : failed to sync secret cache: timed out waiting for the condition Nov 22 07:57:20 crc kubenswrapper[4789]: I1122 07:57:20.001512 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Nov 22 07:57:20 crc kubenswrapper[4789]: I1122 07:57:20.021576 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Nov 22 07:57:20 crc kubenswrapper[4789]: I1122 07:57:20.026532 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7f8f3dfa-5fc0-4642-96c0-22b6f3081676-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-d4l6q\" (UID: \"7f8f3dfa-5fc0-4642-96c0-22b6f3081676\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-d4l6q" Nov 22 07:57:20 crc kubenswrapper[4789]: I1122 07:57:20.042184 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Nov 22 07:57:20 crc kubenswrapper[4789]: I1122 07:57:20.049645 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/9eb8d5d8-3ad7-41d1-9bb2-2d6865619967-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-77svl\" (UID: \"9eb8d5d8-3ad7-41d1-9bb2-2d6865619967\") " pod="openshift-marketplace/marketplace-operator-79b997595-77svl" Nov 22 07:57:20 crc kubenswrapper[4789]: I1122 07:57:20.062300 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Nov 22 07:57:20 crc kubenswrapper[4789]: I1122 07:57:20.082076 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Nov 22 07:57:20 crc kubenswrapper[4789]: I1122 07:57:20.114937 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Nov 22 07:57:20 crc kubenswrapper[4789]: I1122 07:57:20.116480 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9eb8d5d8-3ad7-41d1-9bb2-2d6865619967-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-77svl\" (UID: \"9eb8d5d8-3ad7-41d1-9bb2-2d6865619967\") " pod="openshift-marketplace/marketplace-operator-79b997595-77svl" Nov 22 07:57:20 crc kubenswrapper[4789]: I1122 07:57:20.121184 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Nov 22 07:57:20 crc kubenswrapper[4789]: I1122 07:57:20.141329 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Nov 22 07:57:20 crc kubenswrapper[4789]: I1122 07:57:20.149235 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/10112213-cdd9-4bab-a7d5-bc7a3fab5a48-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-fr4zx\" (UID: \"10112213-cdd9-4bab-a7d5-bc7a3fab5a48\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-fr4zx" Nov 22 07:57:20 crc kubenswrapper[4789]: I1122 07:57:20.161172 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Nov 22 07:57:20 crc kubenswrapper[4789]: I1122 07:57:20.181119 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Nov 22 07:57:20 crc kubenswrapper[4789]: I1122 07:57:20.201677 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Nov 22 07:57:20 crc kubenswrapper[4789]: E1122 07:57:20.213856 4789 configmap.go:193] Couldn't get configMap openshift-etcd-operator/etcd-ca-bundle: failed to sync configmap cache: timed out waiting for the condition Nov 22 07:57:20 crc kubenswrapper[4789]: E1122 07:57:20.213878 4789 configmap.go:193] Couldn't get configMap openshift-etcd-operator/etcd-operator-config: failed to sync configmap cache: timed out waiting for the condition Nov 22 07:57:20 crc kubenswrapper[4789]: E1122 07:57:20.213943 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/0776ae85-c7b7-4d22-b1fb-40b56785205a-etcd-ca podName:0776ae85-c7b7-4d22-b1fb-40b56785205a nodeName:}" failed. No retries permitted until 2025-11-22 07:57:20.713922949 +0000 UTC m=+54.948323213 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etcd-ca" (UniqueName: "kubernetes.io/configmap/0776ae85-c7b7-4d22-b1fb-40b56785205a-etcd-ca") pod "etcd-operator-b45778765-k5c2k" (UID: "0776ae85-c7b7-4d22-b1fb-40b56785205a") : failed to sync configmap cache: timed out waiting for the condition Nov 22 07:57:20 crc kubenswrapper[4789]: E1122 07:57:20.213963 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/0776ae85-c7b7-4d22-b1fb-40b56785205a-config podName:0776ae85-c7b7-4d22-b1fb-40b56785205a nodeName:}" failed. No retries permitted until 2025-11-22 07:57:20.71395646 +0000 UTC m=+54.948356733 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/0776ae85-c7b7-4d22-b1fb-40b56785205a-config") pod "etcd-operator-b45778765-k5c2k" (UID: "0776ae85-c7b7-4d22-b1fb-40b56785205a") : failed to sync configmap cache: timed out waiting for the condition Nov 22 07:57:20 crc kubenswrapper[4789]: E1122 07:57:20.213971 4789 secret.go:188] Couldn't get secret openshift-etcd-operator/etcd-operator-serving-cert: failed to sync secret cache: timed out waiting for the condition Nov 22 07:57:20 crc kubenswrapper[4789]: E1122 07:57:20.214010 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0776ae85-c7b7-4d22-b1fb-40b56785205a-serving-cert podName:0776ae85-c7b7-4d22-b1fb-40b56785205a nodeName:}" failed. No retries permitted until 2025-11-22 07:57:20.713990871 +0000 UTC m=+54.948391144 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/0776ae85-c7b7-4d22-b1fb-40b56785205a-serving-cert") pod "etcd-operator-b45778765-k5c2k" (UID: "0776ae85-c7b7-4d22-b1fb-40b56785205a") : failed to sync secret cache: timed out waiting for the condition Nov 22 07:57:20 crc kubenswrapper[4789]: E1122 07:57:20.213946 4789 configmap.go:193] Couldn't get configMap openshift-service-ca/signing-cabundle: failed to sync configmap cache: timed out waiting for the condition Nov 22 07:57:20 crc kubenswrapper[4789]: E1122 07:57:20.214045 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/065af0aa-e76c-4eab-9a85-d43956522aca-signing-cabundle podName:065af0aa-e76c-4eab-9a85-d43956522aca nodeName:}" failed. No retries permitted until 2025-11-22 07:57:20.714037332 +0000 UTC m=+54.948437726 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "signing-cabundle" (UniqueName: "kubernetes.io/configmap/065af0aa-e76c-4eab-9a85-d43956522aca-signing-cabundle") pod "service-ca-9c57cc56f-whfpb" (UID: "065af0aa-e76c-4eab-9a85-d43956522aca") : failed to sync configmap cache: timed out waiting for the condition Nov 22 07:57:20 crc kubenswrapper[4789]: E1122 07:57:20.215174 4789 secret.go:188] Couldn't get secret openshift-machine-config-operator/node-bootstrapper-token: failed to sync secret cache: timed out waiting for the condition Nov 22 07:57:20 crc kubenswrapper[4789]: E1122 07:57:20.215205 4789 secret.go:188] Couldn't get secret openshift-etcd-operator/etcd-client: failed to sync secret cache: timed out waiting for the condition Nov 22 07:57:20 crc kubenswrapper[4789]: E1122 07:57:20.215231 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c562523e-6eb7-4c72-8ba2-f968f171d387-node-bootstrap-token podName:c562523e-6eb7-4c72-8ba2-f968f171d387 nodeName:}" failed. No retries permitted until 2025-11-22 07:57:20.715218884 +0000 UTC m=+54.949619247 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "node-bootstrap-token" (UniqueName: "kubernetes.io/secret/c562523e-6eb7-4c72-8ba2-f968f171d387-node-bootstrap-token") pod "machine-config-server-qh9np" (UID: "c562523e-6eb7-4c72-8ba2-f968f171d387") : failed to sync secret cache: timed out waiting for the condition Nov 22 07:57:20 crc kubenswrapper[4789]: E1122 07:57:20.215238 4789 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/package-server-manager-serving-cert: failed to sync secret cache: timed out waiting for the condition Nov 22 07:57:20 crc kubenswrapper[4789]: E1122 07:57:20.215251 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0776ae85-c7b7-4d22-b1fb-40b56785205a-etcd-client podName:0776ae85-c7b7-4d22-b1fb-40b56785205a nodeName:}" failed. No retries permitted until 2025-11-22 07:57:20.715241835 +0000 UTC m=+54.949642208 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etcd-client" (UniqueName: "kubernetes.io/secret/0776ae85-c7b7-4d22-b1fb-40b56785205a-etcd-client") pod "etcd-operator-b45778765-k5c2k" (UID: "0776ae85-c7b7-4d22-b1fb-40b56785205a") : failed to sync secret cache: timed out waiting for the condition Nov 22 07:57:20 crc kubenswrapper[4789]: E1122 07:57:20.215268 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f8d25be9-c80c-4057-b84f-7e78598d2e75-package-server-manager-serving-cert podName:f8d25be9-c80c-4057-b84f-7e78598d2e75 nodeName:}" failed. No retries permitted until 2025-11-22 07:57:20.715260256 +0000 UTC m=+54.949660629 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "package-server-manager-serving-cert" (UniqueName: "kubernetes.io/secret/f8d25be9-c80c-4057-b84f-7e78598d2e75-package-server-manager-serving-cert") pod "package-server-manager-789f6589d5-sszl6" (UID: "f8d25be9-c80c-4057-b84f-7e78598d2e75") : failed to sync secret cache: timed out waiting for the condition Nov 22 07:57:20 crc kubenswrapper[4789]: E1122 07:57:20.215273 4789 configmap.go:193] Couldn't get configMap openshift-etcd-operator/etcd-service-ca-bundle: failed to sync configmap cache: timed out waiting for the condition Nov 22 07:57:20 crc kubenswrapper[4789]: E1122 07:57:20.215297 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/0776ae85-c7b7-4d22-b1fb-40b56785205a-etcd-service-ca podName:0776ae85-c7b7-4d22-b1fb-40b56785205a nodeName:}" failed. No retries permitted until 2025-11-22 07:57:20.715289556 +0000 UTC m=+54.949689829 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etcd-service-ca" (UniqueName: "kubernetes.io/configmap/0776ae85-c7b7-4d22-b1fb-40b56785205a-etcd-service-ca") pod "etcd-operator-b45778765-k5c2k" (UID: "0776ae85-c7b7-4d22-b1fb-40b56785205a") : failed to sync configmap cache: timed out waiting for the condition Nov 22 07:57:20 crc kubenswrapper[4789]: E1122 07:57:20.215321 4789 secret.go:188] Couldn't get secret openshift-ingress-canary/canary-serving-cert: failed to sync secret cache: timed out waiting for the condition Nov 22 07:57:20 crc kubenswrapper[4789]: E1122 07:57:20.215331 4789 configmap.go:193] Couldn't get configMap openshift-multus/cni-sysctl-allowlist: failed to sync configmap cache: timed out waiting for the condition Nov 22 07:57:20 crc kubenswrapper[4789]: E1122 07:57:20.215346 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1d0ebe8e-6cea-43d3-9f15-b67c4cbc29e7-cert podName:1d0ebe8e-6cea-43d3-9f15-b67c4cbc29e7 nodeName:}" failed. No retries permitted until 2025-11-22 07:57:20.715337918 +0000 UTC m=+54.949738301 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/1d0ebe8e-6cea-43d3-9f15-b67c4cbc29e7-cert") pod "ingress-canary-qc925" (UID: "1d0ebe8e-6cea-43d3-9f15-b67c4cbc29e7") : failed to sync secret cache: timed out waiting for the condition Nov 22 07:57:20 crc kubenswrapper[4789]: E1122 07:57:20.215376 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/75fa6283-acd8-41bc-9803-2f55119f6828-cni-sysctl-allowlist podName:75fa6283-acd8-41bc-9803-2f55119f6828 nodeName:}" failed. No retries permitted until 2025-11-22 07:57:20.715366578 +0000 UTC m=+54.949766951 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cni-sysctl-allowlist" (UniqueName: "kubernetes.io/configmap/75fa6283-acd8-41bc-9803-2f55119f6828-cni-sysctl-allowlist") pod "cni-sysctl-allowlist-ds-gcnts" (UID: "75fa6283-acd8-41bc-9803-2f55119f6828") : failed to sync configmap cache: timed out waiting for the condition Nov 22 07:57:20 crc kubenswrapper[4789]: E1122 07:57:20.216537 4789 secret.go:188] Couldn't get secret openshift-machine-config-operator/machine-config-server-tls: failed to sync secret cache: timed out waiting for the condition Nov 22 07:57:20 crc kubenswrapper[4789]: E1122 07:57:20.216540 4789 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/packageserver-service-cert: failed to sync secret cache: timed out waiting for the condition Nov 22 07:57:20 crc kubenswrapper[4789]: E1122 07:57:20.216597 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c562523e-6eb7-4c72-8ba2-f968f171d387-certs podName:c562523e-6eb7-4c72-8ba2-f968f171d387 nodeName:}" failed. No retries permitted until 2025-11-22 07:57:20.716585181 +0000 UTC m=+54.950985534 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "certs" (UniqueName: "kubernetes.io/secret/c562523e-6eb7-4c72-8ba2-f968f171d387-certs") pod "machine-config-server-qh9np" (UID: "c562523e-6eb7-4c72-8ba2-f968f171d387") : failed to sync secret cache: timed out waiting for the condition Nov 22 07:57:20 crc kubenswrapper[4789]: E1122 07:57:20.216625 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1c517779-505a-4371-9cd7-9537c2fd8763-webhook-cert podName:1c517779-505a-4371-9cd7-9537c2fd8763 nodeName:}" failed. No retries permitted until 2025-11-22 07:57:20.716607642 +0000 UTC m=+54.951007915 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-cert" (UniqueName: "kubernetes.io/secret/1c517779-505a-4371-9cd7-9537c2fd8763-webhook-cert") pod "packageserver-d55dfcdfc-2dxsq" (UID: "1c517779-505a-4371-9cd7-9537c2fd8763") : failed to sync secret cache: timed out waiting for the condition Nov 22 07:57:20 crc kubenswrapper[4789]: E1122 07:57:20.216624 4789 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/packageserver-service-cert: failed to sync secret cache: timed out waiting for the condition Nov 22 07:57:20 crc kubenswrapper[4789]: E1122 07:57:20.216646 4789 secret.go:188] Couldn't get secret openshift-service-ca/signing-key: failed to sync secret cache: timed out waiting for the condition Nov 22 07:57:20 crc kubenswrapper[4789]: E1122 07:57:20.216660 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1c517779-505a-4371-9cd7-9537c2fd8763-apiservice-cert podName:1c517779-505a-4371-9cd7-9537c2fd8763 nodeName:}" failed. No retries permitted until 2025-11-22 07:57:20.716651913 +0000 UTC m=+54.951052296 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "apiservice-cert" (UniqueName: "kubernetes.io/secret/1c517779-505a-4371-9cd7-9537c2fd8763-apiservice-cert") pod "packageserver-d55dfcdfc-2dxsq" (UID: "1c517779-505a-4371-9cd7-9537c2fd8763") : failed to sync secret cache: timed out waiting for the condition Nov 22 07:57:20 crc kubenswrapper[4789]: E1122 07:57:20.216674 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/065af0aa-e76c-4eab-9a85-d43956522aca-signing-key podName:065af0aa-e76c-4eab-9a85-d43956522aca nodeName:}" failed. No retries permitted until 2025-11-22 07:57:20.716668124 +0000 UTC m=+54.951068397 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "signing-key" (UniqueName: "kubernetes.io/secret/065af0aa-e76c-4eab-9a85-d43956522aca-signing-key") pod "service-ca-9c57cc56f-whfpb" (UID: "065af0aa-e76c-4eab-9a85-d43956522aca") : failed to sync secret cache: timed out waiting for the condition Nov 22 07:57:20 crc kubenswrapper[4789]: I1122 07:57:20.221724 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Nov 22 07:57:20 crc kubenswrapper[4789]: I1122 07:57:20.240694 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Nov 22 07:57:20 crc kubenswrapper[4789]: I1122 07:57:20.260891 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Nov 22 07:57:20 crc kubenswrapper[4789]: I1122 07:57:20.281380 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Nov 22 07:57:20 crc kubenswrapper[4789]: I1122 07:57:20.301813 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Nov 22 07:57:20 crc kubenswrapper[4789]: I1122 07:57:20.322034 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Nov 22 07:57:20 crc kubenswrapper[4789]: I1122 07:57:20.340938 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Nov 22 07:57:20 crc kubenswrapper[4789]: I1122 07:57:20.361415 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Nov 22 07:57:20 crc kubenswrapper[4789]: I1122 07:57:20.385375 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Nov 22 07:57:20 crc kubenswrapper[4789]: I1122 07:57:20.401733 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Nov 22 07:57:20 crc kubenswrapper[4789]: I1122 07:57:20.420913 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Nov 22 07:57:20 crc kubenswrapper[4789]: I1122 07:57:20.461989 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Nov 22 07:57:20 crc kubenswrapper[4789]: I1122 07:57:20.481738 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Nov 22 07:57:20 crc kubenswrapper[4789]: I1122 07:57:20.502111 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Nov 22 07:57:20 crc kubenswrapper[4789]: I1122 07:57:20.521586 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Nov 22 07:57:20 crc kubenswrapper[4789]: I1122 07:57:20.536552 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ffef360f-8678-442d-9b27-b8de7646f1a3-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-t5d22\" (UID: \"ffef360f-8678-442d-9b27-b8de7646f1a3\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-t5d22" Nov 22 07:57:20 crc kubenswrapper[4789]: I1122 07:57:20.541360 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Nov 22 07:57:20 crc kubenswrapper[4789]: I1122 07:57:20.560870 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Nov 22 07:57:20 crc kubenswrapper[4789]: I1122 07:57:20.580969 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Nov 22 07:57:20 crc kubenswrapper[4789]: I1122 07:57:20.601554 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-sysctl-allowlist" Nov 22 07:57:20 crc kubenswrapper[4789]: I1122 07:57:20.622035 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Nov 22 07:57:20 crc kubenswrapper[4789]: I1122 07:57:20.641340 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Nov 22 07:57:20 crc kubenswrapper[4789]: I1122 07:57:20.661013 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Nov 22 07:57:20 crc kubenswrapper[4789]: I1122 07:57:20.681500 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Nov 22 07:57:20 crc kubenswrapper[4789]: I1122 07:57:20.701876 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Nov 22 07:57:20 crc kubenswrapper[4789]: I1122 07:57:20.721358 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Nov 22 07:57:20 crc kubenswrapper[4789]: I1122 07:57:20.739135 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0776ae85-c7b7-4d22-b1fb-40b56785205a-serving-cert\") pod \"etcd-operator-b45778765-k5c2k\" (UID: \"0776ae85-c7b7-4d22-b1fb-40b56785205a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-k5c2k" Nov 22 07:57:20 crc kubenswrapper[4789]: I1122 07:57:20.739182 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/065af0aa-e76c-4eab-9a85-d43956522aca-signing-cabundle\") pod \"service-ca-9c57cc56f-whfpb\" (UID: \"065af0aa-e76c-4eab-9a85-d43956522aca\") " pod="openshift-service-ca/service-ca-9c57cc56f-whfpb" Nov 22 07:57:20 crc kubenswrapper[4789]: I1122 07:57:20.739204 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0776ae85-c7b7-4d22-b1fb-40b56785205a-config\") pod \"etcd-operator-b45778765-k5c2k\" (UID: \"0776ae85-c7b7-4d22-b1fb-40b56785205a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-k5c2k" Nov 22 07:57:20 crc kubenswrapper[4789]: I1122 07:57:20.739219 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/0776ae85-c7b7-4d22-b1fb-40b56785205a-etcd-ca\") pod \"etcd-operator-b45778765-k5c2k\" (UID: \"0776ae85-c7b7-4d22-b1fb-40b56785205a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-k5c2k" Nov 22 07:57:20 crc kubenswrapper[4789]: I1122 07:57:20.739253 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/c562523e-6eb7-4c72-8ba2-f968f171d387-certs\") pod \"machine-config-server-qh9np\" (UID: \"c562523e-6eb7-4c72-8ba2-f968f171d387\") " pod="openshift-machine-config-operator/machine-config-server-qh9np" Nov 22 07:57:20 crc kubenswrapper[4789]: I1122 07:57:20.739294 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1d0ebe8e-6cea-43d3-9f15-b67c4cbc29e7-cert\") pod \"ingress-canary-qc925\" (UID: \"1d0ebe8e-6cea-43d3-9f15-b67c4cbc29e7\") " pod="openshift-ingress-canary/ingress-canary-qc925" Nov 22 07:57:20 crc kubenswrapper[4789]: I1122 07:57:20.739320 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1c517779-505a-4371-9cd7-9537c2fd8763-webhook-cert\") pod \"packageserver-d55dfcdfc-2dxsq\" (UID: \"1c517779-505a-4371-9cd7-9537c2fd8763\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2dxsq" Nov 22 07:57:20 crc kubenswrapper[4789]: I1122 07:57:20.739354 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/75fa6283-acd8-41bc-9803-2f55119f6828-cni-sysctl-allowlist\") pod \"cni-sysctl-allowlist-ds-gcnts\" (UID: \"75fa6283-acd8-41bc-9803-2f55119f6828\") " pod="openshift-multus/cni-sysctl-allowlist-ds-gcnts" Nov 22 07:57:20 crc kubenswrapper[4789]: I1122 07:57:20.739385 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/f8d25be9-c80c-4057-b84f-7e78598d2e75-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-sszl6\" (UID: \"f8d25be9-c80c-4057-b84f-7e78598d2e75\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-sszl6" Nov 22 07:57:20 crc kubenswrapper[4789]: I1122 07:57:20.739421 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/c562523e-6eb7-4c72-8ba2-f968f171d387-node-bootstrap-token\") pod \"machine-config-server-qh9np\" (UID: \"c562523e-6eb7-4c72-8ba2-f968f171d387\") " pod="openshift-machine-config-operator/machine-config-server-qh9np" Nov 22 07:57:20 crc kubenswrapper[4789]: I1122 07:57:20.739458 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/0776ae85-c7b7-4d22-b1fb-40b56785205a-etcd-client\") pod \"etcd-operator-b45778765-k5c2k\" (UID: \"0776ae85-c7b7-4d22-b1fb-40b56785205a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-k5c2k" Nov 22 07:57:20 crc kubenswrapper[4789]: I1122 07:57:20.739479 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/0776ae85-c7b7-4d22-b1fb-40b56785205a-etcd-service-ca\") pod \"etcd-operator-b45778765-k5c2k\" (UID: \"0776ae85-c7b7-4d22-b1fb-40b56785205a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-k5c2k" Nov 22 07:57:20 crc kubenswrapper[4789]: I1122 07:57:20.739499 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1c517779-505a-4371-9cd7-9537c2fd8763-apiservice-cert\") pod \"packageserver-d55dfcdfc-2dxsq\" (UID: \"1c517779-505a-4371-9cd7-9537c2fd8763\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2dxsq" Nov 22 07:57:20 crc kubenswrapper[4789]: I1122 07:57:20.739544 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/065af0aa-e76c-4eab-9a85-d43956522aca-signing-key\") pod \"service-ca-9c57cc56f-whfpb\" (UID: \"065af0aa-e76c-4eab-9a85-d43956522aca\") " pod="openshift-service-ca/service-ca-9c57cc56f-whfpb" Nov 22 07:57:20 crc kubenswrapper[4789]: I1122 07:57:20.739989 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0776ae85-c7b7-4d22-b1fb-40b56785205a-config\") pod \"etcd-operator-b45778765-k5c2k\" (UID: \"0776ae85-c7b7-4d22-b1fb-40b56785205a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-k5c2k" Nov 22 07:57:20 crc kubenswrapper[4789]: I1122 07:57:20.740680 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/0776ae85-c7b7-4d22-b1fb-40b56785205a-etcd-ca\") pod \"etcd-operator-b45778765-k5c2k\" (UID: \"0776ae85-c7b7-4d22-b1fb-40b56785205a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-k5c2k" Nov 22 07:57:20 crc kubenswrapper[4789]: I1122 07:57:20.741225 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/0776ae85-c7b7-4d22-b1fb-40b56785205a-etcd-service-ca\") pod \"etcd-operator-b45778765-k5c2k\" (UID: \"0776ae85-c7b7-4d22-b1fb-40b56785205a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-k5c2k" Nov 22 07:57:20 crc kubenswrapper[4789]: I1122 07:57:20.741331 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/75fa6283-acd8-41bc-9803-2f55119f6828-cni-sysctl-allowlist\") pod \"cni-sysctl-allowlist-ds-gcnts\" (UID: \"75fa6283-acd8-41bc-9803-2f55119f6828\") " pod="openshift-multus/cni-sysctl-allowlist-ds-gcnts" Nov 22 07:57:20 crc kubenswrapper[4789]: I1122 07:57:20.741492 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Nov 22 07:57:20 crc kubenswrapper[4789]: I1122 07:57:20.741816 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/065af0aa-e76c-4eab-9a85-d43956522aca-signing-cabundle\") pod \"service-ca-9c57cc56f-whfpb\" (UID: \"065af0aa-e76c-4eab-9a85-d43956522aca\") " pod="openshift-service-ca/service-ca-9c57cc56f-whfpb" Nov 22 07:57:20 crc kubenswrapper[4789]: I1122 07:57:20.743283 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0776ae85-c7b7-4d22-b1fb-40b56785205a-serving-cert\") pod \"etcd-operator-b45778765-k5c2k\" (UID: \"0776ae85-c7b7-4d22-b1fb-40b56785205a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-k5c2k" Nov 22 07:57:20 crc kubenswrapper[4789]: I1122 07:57:20.743967 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/0776ae85-c7b7-4d22-b1fb-40b56785205a-etcd-client\") pod \"etcd-operator-b45778765-k5c2k\" (UID: \"0776ae85-c7b7-4d22-b1fb-40b56785205a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-k5c2k" Nov 22 07:57:20 crc kubenswrapper[4789]: I1122 07:57:20.744124 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1c517779-505a-4371-9cd7-9537c2fd8763-webhook-cert\") pod \"packageserver-d55dfcdfc-2dxsq\" (UID: \"1c517779-505a-4371-9cd7-9537c2fd8763\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2dxsq" Nov 22 07:57:20 crc kubenswrapper[4789]: I1122 07:57:20.744679 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/065af0aa-e76c-4eab-9a85-d43956522aca-signing-key\") pod \"service-ca-9c57cc56f-whfpb\" (UID: \"065af0aa-e76c-4eab-9a85-d43956522aca\") " pod="openshift-service-ca/service-ca-9c57cc56f-whfpb" Nov 22 07:57:20 crc kubenswrapper[4789]: I1122 07:57:20.744813 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/f8d25be9-c80c-4057-b84f-7e78598d2e75-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-sszl6\" (UID: \"f8d25be9-c80c-4057-b84f-7e78598d2e75\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-sszl6" Nov 22 07:57:20 crc kubenswrapper[4789]: I1122 07:57:20.745063 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/c562523e-6eb7-4c72-8ba2-f968f171d387-certs\") pod \"machine-config-server-qh9np\" (UID: \"c562523e-6eb7-4c72-8ba2-f968f171d387\") " pod="openshift-machine-config-operator/machine-config-server-qh9np" Nov 22 07:57:20 crc kubenswrapper[4789]: I1122 07:57:20.745885 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1c517779-505a-4371-9cd7-9537c2fd8763-apiservice-cert\") pod \"packageserver-d55dfcdfc-2dxsq\" (UID: \"1c517779-505a-4371-9cd7-9537c2fd8763\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2dxsq" Nov 22 07:57:20 crc kubenswrapper[4789]: I1122 07:57:20.748102 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/c562523e-6eb7-4c72-8ba2-f968f171d387-node-bootstrap-token\") pod \"machine-config-server-qh9np\" (UID: \"c562523e-6eb7-4c72-8ba2-f968f171d387\") " pod="openshift-machine-config-operator/machine-config-server-qh9np" Nov 22 07:57:20 crc kubenswrapper[4789]: I1122 07:57:20.748631 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1d0ebe8e-6cea-43d3-9f15-b67c4cbc29e7-cert\") pod \"ingress-canary-qc925\" (UID: \"1d0ebe8e-6cea-43d3-9f15-b67c4cbc29e7\") " pod="openshift-ingress-canary/ingress-canary-qc925" Nov 22 07:57:20 crc kubenswrapper[4789]: I1122 07:57:20.776202 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bpjnq\" (UniqueName: \"kubernetes.io/projected/d4d9ab07-53b6-4aef-bcac-c0140d870370-kube-api-access-bpjnq\") pod \"route-controller-manager-6576b87f9c-fpw88\" (UID: \"d4d9ab07-53b6-4aef-bcac-c0140d870370\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fpw88" Nov 22 07:57:20 crc kubenswrapper[4789]: I1122 07:57:20.798859 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bb5dq\" (UniqueName: \"kubernetes.io/projected/23967110-3d55-4767-9fa9-230138c92e42-kube-api-access-bb5dq\") pod \"controller-manager-879f6c89f-4s4sh\" (UID: \"23967110-3d55-4767-9fa9-230138c92e42\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4s4sh" Nov 22 07:57:20 crc kubenswrapper[4789]: I1122 07:57:20.815997 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tsm8d\" (UniqueName: \"kubernetes.io/projected/c0beaf3d-a9eb-47cc-bb53-392e75959282-kube-api-access-tsm8d\") pod \"cluster-image-registry-operator-dc59b4c8b-ckmwl\" (UID: \"c0beaf3d-a9eb-47cc-bb53-392e75959282\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ckmwl" Nov 22 07:57:20 crc kubenswrapper[4789]: I1122 07:57:20.841200 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rv7fq\" (UniqueName: \"kubernetes.io/projected/ffef360f-8678-442d-9b27-b8de7646f1a3-kube-api-access-rv7fq\") pod \"openshift-controller-manager-operator-756b6f6bc6-t5d22\" (UID: \"ffef360f-8678-442d-9b27-b8de7646f1a3\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-t5d22" Nov 22 07:57:20 crc kubenswrapper[4789]: I1122 07:57:20.858218 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c0beaf3d-a9eb-47cc-bb53-392e75959282-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-ckmwl\" (UID: \"c0beaf3d-a9eb-47cc-bb53-392e75959282\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ckmwl" Nov 22 07:57:20 crc kubenswrapper[4789]: I1122 07:57:20.880235 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7dqck\" (UniqueName: \"kubernetes.io/projected/83ac1d78-cc4c-48b7-b975-2f88802e2d17-kube-api-access-7dqck\") pod \"downloads-7954f5f757-tdpht\" (UID: \"83ac1d78-cc4c-48b7-b975-2f88802e2d17\") " pod="openshift-console/downloads-7954f5f757-tdpht" Nov 22 07:57:20 crc kubenswrapper[4789]: I1122 07:57:20.897715 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hmtsb\" (UniqueName: \"kubernetes.io/projected/1ca00755-c6ae-42d2-833e-fc31b6a480df-kube-api-access-hmtsb\") pod \"authentication-operator-69f744f599-ksdn9\" (UID: \"1ca00755-c6ae-42d2-833e-fc31b6a480df\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ksdn9" Nov 22 07:57:20 crc kubenswrapper[4789]: I1122 07:57:20.901680 4789 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Nov 22 07:57:20 crc kubenswrapper[4789]: I1122 07:57:20.922338 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Nov 22 07:57:20 crc kubenswrapper[4789]: I1122 07:57:20.942170 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Nov 22 07:57:20 crc kubenswrapper[4789]: I1122 07:57:20.960121 4789 request.go:700] Waited for 1.966362307s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-console-operator/serviceaccounts/console-operator/token Nov 22 07:57:20 crc kubenswrapper[4789]: I1122 07:57:20.976371 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qsdgk\" (UniqueName: \"kubernetes.io/projected/79e3d4d9-ec8e-4578-871c-a336ad22a2ac-kube-api-access-qsdgk\") pod \"console-operator-58897d9998-9lzll\" (UID: \"79e3d4d9-ec8e-4578-871c-a336ad22a2ac\") " pod="openshift-console-operator/console-operator-58897d9998-9lzll" Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:20.997363 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xh6c5\" (UniqueName: \"kubernetes.io/projected/9b861750-2517-4cf3-9c78-930480711af0-kube-api-access-xh6c5\") pod \"console-f9d7485db-pb98x\" (UID: \"9b861750-2517-4cf3-9c78-930480711af0\") " pod="openshift-console/console-f9d7485db-pb98x" Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.003400 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fpw88" Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.016332 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sp9jx\" (UniqueName: \"kubernetes.io/projected/be26df1a-a3b3-411b-9e84-df356f3a7af3-kube-api-access-sp9jx\") pod \"machine-approver-56656f9798-v4wls\" (UID: \"be26df1a-a3b3-411b-9e84-df356f3a7af3\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-v4wls" Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.016937 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-tdpht" Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.031703 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-9lzll" Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.039242 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5btzg\" (UniqueName: \"kubernetes.io/projected/ff518a19-6241-4c65-b2a0-bee6ab939a6c-kube-api-access-5btzg\") pod \"cluster-samples-operator-665b6dd947-6jw6b\" (UID: \"ff518a19-6241-4c65-b2a0-bee6ab939a6c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6jw6b" Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.054842 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kb79q\" (UniqueName: \"kubernetes.io/projected/ce67cd3e-c2d2-4562-bc26-0a68bce1a1f0-kube-api-access-kb79q\") pod \"openshift-config-operator-7777fb866f-qwhkx\" (UID: \"ce67cd3e-c2d2-4562-bc26-0a68bce1a1f0\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-qwhkx" Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.060422 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-ksdn9" Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.071213 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-pb98x" Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.077307 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h8d7k\" (UniqueName: \"kubernetes.io/projected/0fb11c40-1e46-42a7-b384-626786bc4e3e-kube-api-access-h8d7k\") pod \"openshift-apiserver-operator-796bbdcf4f-wjmrv\" (UID: \"0fb11c40-1e46-42a7-b384-626786bc4e3e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wjmrv" Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.078519 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ckmwl" Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.082527 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.089243 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-4s4sh" Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.101326 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.122295 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.180791 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jf27d\" (UniqueName: \"kubernetes.io/projected/dfa022a0-f86e-4f4b-99f5-22b84e9d6558-kube-api-access-jf27d\") pod \"olm-operator-6b444d44fb-9n9mc\" (UID: \"dfa022a0-f86e-4f4b-99f5-22b84e9d6558\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9n9mc" Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.181083 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6jw6b" Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.203094 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-smhnq\" (UniqueName: \"kubernetes.io/projected/78d69d0e-bfac-4582-b339-b2e585de5eff-kube-api-access-smhnq\") pod \"machine-config-operator-74547568cd-4fbj7\" (UID: \"78d69d0e-bfac-4582-b339-b2e585de5eff\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4fbj7" Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.218322 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/31bd2041-6cb8-4eb6-bb22-5efe6c6009ee-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-2m4cb\" (UID: \"31bd2041-6cb8-4eb6-bb22-5efe6c6009ee\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2m4cb" Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.219133 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-v4wls" Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.242103 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kznfz\" (UniqueName: \"kubernetes.io/projected/ac3d1daa-cb6d-41fc-9798-e687d9b2a21d-kube-api-access-kznfz\") pod \"collect-profiles-29396625-nrhcv\" (UID: \"ac3d1daa-cb6d-41fc-9798-e687d9b2a21d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396625-nrhcv" Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.244788 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wjmrv" Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.250613 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8adfba92-180a-4f0a-adc0-d1e2bdd62dba-metrics-certs\") pod \"network-metrics-daemon-q7rc4\" (UID: \"8adfba92-180a-4f0a-adc0-d1e2bdd62dba\") " pod="openshift-multus/network-metrics-daemon-q7rc4" Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.258764 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-42kgf\" (UniqueName: \"kubernetes.io/projected/cb90f2d2-3321-4118-a449-fdc9343b23e1-kube-api-access-42kgf\") pod \"ingress-operator-5b745b69d9-bsrcz\" (UID: \"cb90f2d2-3321-4118-a449-fdc9343b23e1\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-bsrcz" Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.264571 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2m4cb" Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.272984 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9n9mc" Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.279290 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d32dc4fd-3e62-402c-a1d9-8f2c83f2e334-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-tnmrh\" (UID: \"d32dc4fd-3e62-402c-a1d9-8f2c83f2e334\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-tnmrh" Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.281496 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-qwhkx" Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.288915 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-v4wls" event={"ID":"be26df1a-a3b3-411b-9e84-df356f3a7af3","Type":"ContainerStarted","Data":"64b68f901e7c8a1bd95b2edb30ed1f51a2eb435b8af1fd1905454875b2289e1f"} Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.299247 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bgr4n\" (UniqueName: \"kubernetes.io/projected/94294737-067b-4a4f-b59b-f9f987fa9127-kube-api-access-bgr4n\") pod \"apiserver-76f77b778f-vh974\" (UID: \"94294737-067b-4a4f-b59b-f9f987fa9127\") " pod="openshift-apiserver/apiserver-76f77b778f-vh974" Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.317374 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n6hh2\" (UniqueName: \"kubernetes.io/projected/352693a0-531b-4017-9582-5424c4691795-kube-api-access-n6hh2\") pod \"router-default-5444994796-x5b9n\" (UID: \"352693a0-531b-4017-9582-5424c4691795\") " pod="openshift-ingress/router-default-5444994796-x5b9n" Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.334994 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1507f6ee-6a14-4d67-98b3-1e151b7949cc-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-kdt8l\" (UID: \"1507f6ee-6a14-4d67-98b3-1e151b7949cc\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-kdt8l" Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.346258 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-x5b9n" Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.361767 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vj4w5\" (UniqueName: \"kubernetes.io/projected/dd618778-dc82-4797-b737-14231eefea15-kube-api-access-vj4w5\") pod \"service-ca-operator-777779d784-fg7wj\" (UID: \"dd618778-dc82-4797-b737-14231eefea15\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-fg7wj" Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.364294 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-9lzll"] Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.377133 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wdcz2\" (UniqueName: \"kubernetes.io/projected/55cac199-8203-40b3-8a4f-c571017d998b-kube-api-access-wdcz2\") pod \"oauth-openshift-558db77b4-wnfzj\" (UID: \"55cac199-8203-40b3-8a4f-c571017d998b\") " pod="openshift-authentication/oauth-openshift-558db77b4-wnfzj" Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.397959 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-wnfzj" Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.410532 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kppbq\" (UniqueName: \"kubernetes.io/projected/d086acbe-d982-4880-9538-04a184cb4148-kube-api-access-kppbq\") pod \"machine-api-operator-5694c8668f-spnzw\" (UID: \"d086acbe-d982-4880-9538-04a184cb4148\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-spnzw" Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.428917 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s86jn\" (UniqueName: \"kubernetes.io/projected/3cfa3a64-3e7a-4336-9ae0-86822095a1b4-kube-api-access-s86jn\") pod \"multus-admission-controller-857f4d67dd-fb49f\" (UID: \"3cfa3a64-3e7a-4336-9ae0-86822095a1b4\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-fb49f" Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.438737 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/cb90f2d2-3321-4118-a449-fdc9343b23e1-bound-sa-token\") pod \"ingress-operator-5b745b69d9-bsrcz\" (UID: \"cb90f2d2-3321-4118-a449-fdc9343b23e1\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-bsrcz" Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.444190 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-bsrcz" Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.458845 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2mgz5\" (UniqueName: \"kubernetes.io/projected/7dc5fba4-2ba8-4fa2-a76f-0ad8d1196838-kube-api-access-2mgz5\") pod \"apiserver-7bbb656c7d-frls8\" (UID: \"7dc5fba4-2ba8-4fa2-a76f-0ad8d1196838\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-frls8" Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.465889 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-vh974" Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.466947 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-kdt8l" Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.476167 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396625-nrhcv" Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.483563 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4fbj7" Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.492821 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vzd9t\" (UniqueName: \"kubernetes.io/projected/cb9c37ae-6151-452f-b467-6b7d14bbb2f8-kube-api-access-vzd9t\") pod \"catalog-operator-68c6474976-h6f5v\" (UID: \"cb9c37ae-6151-452f-b467-6b7d14bbb2f8\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-h6f5v" Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.492915 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-spnzw" Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.500462 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-tnmrh" Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.502420 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9vb7l\" (UniqueName: \"kubernetes.io/projected/f2c80a41-3ceb-4420-a46e-566ff4a28dcb-kube-api-access-9vb7l\") pod \"machine-config-controller-84d6567774-zl87n\" (UID: \"f2c80a41-3ceb-4420-a46e-566ff4a28dcb\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zl87n" Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.515202 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g7qk9\" (UniqueName: \"kubernetes.io/projected/d58d000b-3f2c-44cb-90c5-8023e9f4f5db-kube-api-access-g7qk9\") pod \"csi-hostpathplugin-n27w4\" (UID: \"d58d000b-3f2c-44cb-90c5-8023e9f4f5db\") " pod="hostpath-provisioner/csi-hostpathplugin-n27w4" Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.531067 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wjmrv"] Nov 22 07:57:21 crc kubenswrapper[4789]: E1122 07:57:21.538775 4789 secret.go:188] Couldn't get secret openshift-controller-manager-operator/openshift-controller-manager-operator-serving-cert: failed to sync secret cache: timed out waiting for the condition Nov 22 07:57:21 crc kubenswrapper[4789]: E1122 07:57:21.538842 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ffef360f-8678-442d-9b27-b8de7646f1a3-serving-cert podName:ffef360f-8678-442d-9b27-b8de7646f1a3 nodeName:}" failed. No retries permitted until 2025-11-22 07:57:22.538813584 +0000 UTC m=+56.773213857 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/ffef360f-8678-442d-9b27-b8de7646f1a3-serving-cert") pod "openshift-controller-manager-operator-756b6f6bc6-t5d22" (UID: "ffef360f-8678-442d-9b27-b8de7646f1a3") : failed to sync secret cache: timed out waiting for the condition Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.543521 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zgzpt\" (UniqueName: \"kubernetes.io/projected/1d0ebe8e-6cea-43d3-9f15-b67c4cbc29e7-kube-api-access-zgzpt\") pod \"ingress-canary-qc925\" (UID: \"1d0ebe8e-6cea-43d3-9f15-b67c4cbc29e7\") " pod="openshift-ingress-canary/ingress-canary-qc925" Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.551904 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-fb49f" Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.557168 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-fg7wj" Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.563262 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-ksdn9"] Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.567775 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-tdpht"] Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.576336 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cgptp\" (UniqueName: \"kubernetes.io/projected/065af0aa-e76c-4eab-9a85-d43956522aca-kube-api-access-cgptp\") pod \"service-ca-9c57cc56f-whfpb\" (UID: \"065af0aa-e76c-4eab-9a85-d43956522aca\") " pod="openshift-service-ca/service-ca-9c57cc56f-whfpb" Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.581424 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dkw4c\" (UniqueName: \"kubernetes.io/projected/1c517779-505a-4371-9cd7-9537c2fd8763-kube-api-access-dkw4c\") pod \"packageserver-d55dfcdfc-2dxsq\" (UID: \"1c517779-505a-4371-9cd7-9537c2fd8763\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2dxsq" Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.581998 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-h6f5v" Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.591613 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-fpw88"] Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.599014 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zl87n" Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.606318 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rcg4j\" (UniqueName: \"kubernetes.io/projected/0776ae85-c7b7-4d22-b1fb-40b56785205a-kube-api-access-rcg4j\") pod \"etcd-operator-b45778765-k5c2k\" (UID: \"0776ae85-c7b7-4d22-b1fb-40b56785205a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-k5c2k" Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.616680 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-pb98x"] Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.621812 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qwgcd\" (UniqueName: \"kubernetes.io/projected/f8d25be9-c80c-4057-b84f-7e78598d2e75-kube-api-access-qwgcd\") pod \"package-server-manager-789f6589d5-sszl6\" (UID: \"f8d25be9-c80c-4057-b84f-7e78598d2e75\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-sszl6" Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.639434 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xvst4\" (UniqueName: \"kubernetes.io/projected/9eb8d5d8-3ad7-41d1-9bb2-2d6865619967-kube-api-access-xvst4\") pod \"marketplace-operator-79b997595-77svl\" (UID: \"9eb8d5d8-3ad7-41d1-9bb2-2d6865619967\") " pod="openshift-marketplace/marketplace-operator-79b997595-77svl" Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.642491 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2dxsq" Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.648932 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-sszl6" Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.658607 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xk2ns\" (UniqueName: \"kubernetes.io/projected/c562523e-6eb7-4c72-8ba2-f968f171d387-kube-api-access-xk2ns\") pod \"machine-config-server-qh9np\" (UID: \"c562523e-6eb7-4c72-8ba2-f968f171d387\") " pod="openshift-machine-config-operator/machine-config-server-qh9np" Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.661045 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-k5c2k" Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.678918 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6jw6b"] Nov 22 07:57:21 crc kubenswrapper[4789]: W1122 07:57:21.680900 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd4d9ab07_53b6_4aef_bcac_c0140d870370.slice/crio-c9c0e86307f5a170b6e76d3806c5edd7f987998f93cbaa73abaafef7adf10ad8 WatchSource:0}: Error finding container c9c0e86307f5a170b6e76d3806c5edd7f987998f93cbaa73abaafef7adf10ad8: Status 404 returned error can't find the container with id c9c0e86307f5a170b6e76d3806c5edd7f987998f93cbaa73abaafef7adf10ad8 Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.683522 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ckmwl"] Nov 22 07:57:21 crc kubenswrapper[4789]: W1122 07:57:21.684886 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9b861750_2517_4cf3_9c78_930480711af0.slice/crio-ec47c547dee63090f313aec803697265dcce9767c0aacf0c48dddcb8f18c8b4c WatchSource:0}: Error finding container ec47c547dee63090f313aec803697265dcce9767c0aacf0c48dddcb8f18c8b4c: Status 404 returned error can't find the container with id ec47c547dee63090f313aec803697265dcce9767c0aacf0c48dddcb8f18c8b4c Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.690913 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dvrrf\" (UniqueName: \"kubernetes.io/projected/7f8f3dfa-5fc0-4642-96c0-22b6f3081676-kube-api-access-dvrrf\") pod \"kube-storage-version-migrator-operator-b67b599dd-d4l6q\" (UID: \"7f8f3dfa-5fc0-4642-96c0-22b6f3081676\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-d4l6q" Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.693214 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-whfpb" Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.698263 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-4s4sh"] Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.701109 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6r7x6\" (UniqueName: \"kubernetes.io/projected/75fa6283-acd8-41bc-9803-2f55119f6828-kube-api-access-6r7x6\") pod \"cni-sysctl-allowlist-ds-gcnts\" (UID: \"75fa6283-acd8-41bc-9803-2f55119f6828\") " pod="openshift-multus/cni-sysctl-allowlist-ds-gcnts" Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.703046 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-qh9np" Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.708435 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-qc925" Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.716091 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8b4tq\" (UniqueName: \"kubernetes.io/projected/10112213-cdd9-4bab-a7d5-bc7a3fab5a48-kube-api-access-8b4tq\") pod \"control-plane-machine-set-operator-78cbb6b69f-fr4zx\" (UID: \"10112213-cdd9-4bab-a7d5-bc7a3fab5a48\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-fr4zx" Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.722065 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.735732 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-wnfzj"] Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.742306 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Nov 22 07:57:21 crc kubenswrapper[4789]: W1122 07:57:21.745474 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc0beaf3d_a9eb_47cc_bb53_392e75959282.slice/crio-f4a4324ffb2c9ccec790a1dc16eee2d8a1b34067f505b7ef476fe2a1cbf2a9d7 WatchSource:0}: Error finding container f4a4324ffb2c9ccec790a1dc16eee2d8a1b34067f505b7ef476fe2a1cbf2a9d7: Status 404 returned error can't find the container with id f4a4324ffb2c9ccec790a1dc16eee2d8a1b34067f505b7ef476fe2a1cbf2a9d7 Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.750810 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-frls8" Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.751674 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-n27w4" Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.753559 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-bsrcz"] Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.762371 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.781716 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.794681 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8adfba92-180a-4f0a-adc0-d1e2bdd62dba-metrics-certs\") pod \"network-metrics-daemon-q7rc4\" (UID: \"8adfba92-180a-4f0a-adc0-d1e2bdd62dba\") " pod="openshift-multus/network-metrics-daemon-q7rc4" Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.797411 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9n9mc"] Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.801682 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.829034 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.829515 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2m4cb"] Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.857563 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-qwhkx"] Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.879619 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-vh974"] Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.912790 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-d4l6q" Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.918056 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-77svl" Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.927278 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-fr4zx" Nov 22 07:57:21 crc kubenswrapper[4789]: W1122 07:57:21.941160 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod94294737_067b_4a4f_b59b_f9f987fa9127.slice/crio-b5d43ff5e588f7405f90f8f46ea168642d7c0ae269c193efeff8ba0483a68b09 WatchSource:0}: Error finding container b5d43ff5e588f7405f90f8f46ea168642d7c0ae269c193efeff8ba0483a68b09: Status 404 returned error can't find the container with id b5d43ff5e588f7405f90f8f46ea168642d7c0ae269c193efeff8ba0483a68b09 Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.962493 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b80556f6-5b32-486d-8459-f17bf90c4e24-trusted-ca\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.962527 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6vwfl\" (UniqueName: \"kubernetes.io/projected/85468946-5921-40d2-bf72-14e0a18ebe6f-kube-api-access-6vwfl\") pod \"dns-default-szb9d\" (UID: \"85468946-5921-40d2-bf72-14e0a18ebe6f\") " pod="openshift-dns/dns-default-szb9d" Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.962556 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e2a09474-4a22-4f1f-8906-95779f62e570-metrics-tls\") pod \"dns-operator-744455d44c-xlf6m\" (UID: \"e2a09474-4a22-4f1f-8906-95779f62e570\") " pod="openshift-dns-operator/dns-operator-744455d44c-xlf6m" Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.962584 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/b80556f6-5b32-486d-8459-f17bf90c4e24-ca-trust-extracted\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.962615 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.962648 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fc9n7\" (UniqueName: \"kubernetes.io/projected/b80556f6-5b32-486d-8459-f17bf90c4e24-kube-api-access-fc9n7\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.962665 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/85468946-5921-40d2-bf72-14e0a18ebe6f-config-volume\") pod \"dns-default-szb9d\" (UID: \"85468946-5921-40d2-bf72-14e0a18ebe6f\") " pod="openshift-dns/dns-default-szb9d" Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.962680 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/b80556f6-5b32-486d-8459-f17bf90c4e24-registry-tls\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.962698 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/85468946-5921-40d2-bf72-14e0a18ebe6f-metrics-tls\") pod \"dns-default-szb9d\" (UID: \"85468946-5921-40d2-bf72-14e0a18ebe6f\") " pod="openshift-dns/dns-default-szb9d" Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.962827 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rz8zk\" (UniqueName: \"kubernetes.io/projected/bf1ce57a-0abc-425b-a274-ca4681cd487d-kube-api-access-rz8zk\") pod \"migrator-59844c95c7-hszd6\" (UID: \"bf1ce57a-0abc-425b-a274-ca4681cd487d\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-hszd6" Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.962952 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b80556f6-5b32-486d-8459-f17bf90c4e24-bound-sa-token\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.962979 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/b80556f6-5b32-486d-8459-f17bf90c4e24-installation-pull-secrets\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.963030 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vf6cv\" (UniqueName: \"kubernetes.io/projected/e2a09474-4a22-4f1f-8906-95779f62e570-kube-api-access-vf6cv\") pod \"dns-operator-744455d44c-xlf6m\" (UID: \"e2a09474-4a22-4f1f-8906-95779f62e570\") " pod="openshift-dns-operator/dns-operator-744455d44c-xlf6m" Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.963045 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/b80556f6-5b32-486d-8459-f17bf90c4e24-registry-certificates\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:21 crc kubenswrapper[4789]: E1122 07:57:21.964894 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 07:57:22.464876805 +0000 UTC m=+56.699277088 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l4q7j" (UID: "b80556f6-5b32-486d-8459-f17bf90c4e24") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:21 crc kubenswrapper[4789]: I1122 07:57:21.984127 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/cni-sysctl-allowlist-ds-gcnts" Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.002343 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-4fbj7"] Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.004874 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396625-nrhcv"] Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.019856 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-kdt8l"] Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.029036 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.064913 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-fg7wj"] Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.068042 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-fb49f"] Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.068527 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.068809 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b80556f6-5b32-486d-8459-f17bf90c4e24-trusted-ca\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.068847 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6vwfl\" (UniqueName: \"kubernetes.io/projected/85468946-5921-40d2-bf72-14e0a18ebe6f-kube-api-access-6vwfl\") pod \"dns-default-szb9d\" (UID: \"85468946-5921-40d2-bf72-14e0a18ebe6f\") " pod="openshift-dns/dns-default-szb9d" Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.068891 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e2a09474-4a22-4f1f-8906-95779f62e570-metrics-tls\") pod \"dns-operator-744455d44c-xlf6m\" (UID: \"e2a09474-4a22-4f1f-8906-95779f62e570\") " pod="openshift-dns-operator/dns-operator-744455d44c-xlf6m" Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.068924 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/b80556f6-5b32-486d-8459-f17bf90c4e24-ca-trust-extracted\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.069083 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fc9n7\" (UniqueName: \"kubernetes.io/projected/b80556f6-5b32-486d-8459-f17bf90c4e24-kube-api-access-fc9n7\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.069115 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/85468946-5921-40d2-bf72-14e0a18ebe6f-config-volume\") pod \"dns-default-szb9d\" (UID: \"85468946-5921-40d2-bf72-14e0a18ebe6f\") " pod="openshift-dns/dns-default-szb9d" Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.069145 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/b80556f6-5b32-486d-8459-f17bf90c4e24-registry-tls\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.069194 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/85468946-5921-40d2-bf72-14e0a18ebe6f-metrics-tls\") pod \"dns-default-szb9d\" (UID: \"85468946-5921-40d2-bf72-14e0a18ebe6f\") " pod="openshift-dns/dns-default-szb9d" Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.069380 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rz8zk\" (UniqueName: \"kubernetes.io/projected/bf1ce57a-0abc-425b-a274-ca4681cd487d-kube-api-access-rz8zk\") pod \"migrator-59844c95c7-hszd6\" (UID: \"bf1ce57a-0abc-425b-a274-ca4681cd487d\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-hszd6" Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.069646 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b80556f6-5b32-486d-8459-f17bf90c4e24-bound-sa-token\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.069709 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/b80556f6-5b32-486d-8459-f17bf90c4e24-installation-pull-secrets\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.069875 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vf6cv\" (UniqueName: \"kubernetes.io/projected/e2a09474-4a22-4f1f-8906-95779f62e570-kube-api-access-vf6cv\") pod \"dns-operator-744455d44c-xlf6m\" (UID: \"e2a09474-4a22-4f1f-8906-95779f62e570\") " pod="openshift-dns-operator/dns-operator-744455d44c-xlf6m" Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.069902 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/b80556f6-5b32-486d-8459-f17bf90c4e24-registry-certificates\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:22 crc kubenswrapper[4789]: E1122 07:57:22.070893 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 07:57:22.570875574 +0000 UTC m=+56.805275847 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.076742 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b80556f6-5b32-486d-8459-f17bf90c4e24-trusted-ca\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.078255 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.079805 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/b80556f6-5b32-486d-8459-f17bf90c4e24-ca-trust-extracted\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.081195 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/85468946-5921-40d2-bf72-14e0a18ebe6f-config-volume\") pod \"dns-default-szb9d\" (UID: \"85468946-5921-40d2-bf72-14e0a18ebe6f\") " pod="openshift-dns/dns-default-szb9d" Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.082279 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/b80556f6-5b32-486d-8459-f17bf90c4e24-registry-certificates\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.084460 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-q7rc4" Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.098199 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/b80556f6-5b32-486d-8459-f17bf90c4e24-installation-pull-secrets\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.098870 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-spnzw"] Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.104299 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/85468946-5921-40d2-bf72-14e0a18ebe6f-metrics-tls\") pod \"dns-default-szb9d\" (UID: \"85468946-5921-40d2-bf72-14e0a18ebe6f\") " pod="openshift-dns/dns-default-szb9d" Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.105361 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/b80556f6-5b32-486d-8459-f17bf90c4e24-registry-tls\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.105473 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e2a09474-4a22-4f1f-8906-95779f62e570-metrics-tls\") pod \"dns-operator-744455d44c-xlf6m\" (UID: \"e2a09474-4a22-4f1f-8906-95779f62e570\") " pod="openshift-dns-operator/dns-operator-744455d44c-xlf6m" Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.152472 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6vwfl\" (UniqueName: \"kubernetes.io/projected/85468946-5921-40d2-bf72-14e0a18ebe6f-kube-api-access-6vwfl\") pod \"dns-default-szb9d\" (UID: \"85468946-5921-40d2-bf72-14e0a18ebe6f\") " pod="openshift-dns/dns-default-szb9d" Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.160463 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b80556f6-5b32-486d-8459-f17bf90c4e24-bound-sa-token\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.167581 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rz8zk\" (UniqueName: \"kubernetes.io/projected/bf1ce57a-0abc-425b-a274-ca4681cd487d-kube-api-access-rz8zk\") pod \"migrator-59844c95c7-hszd6\" (UID: \"bf1ce57a-0abc-425b-a274-ca4681cd487d\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-hszd6" Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.171410 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:22 crc kubenswrapper[4789]: E1122 07:57:22.172239 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 07:57:22.672223058 +0000 UTC m=+56.906623331 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l4q7j" (UID: "b80556f6-5b32-486d-8459-f17bf90c4e24") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.179536 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vf6cv\" (UniqueName: \"kubernetes.io/projected/e2a09474-4a22-4f1f-8906-95779f62e570-kube-api-access-vf6cv\") pod \"dns-operator-744455d44c-xlf6m\" (UID: \"e2a09474-4a22-4f1f-8906-95779f62e570\") " pod="openshift-dns-operator/dns-operator-744455d44c-xlf6m" Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.191934 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-xlf6m" Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.199479 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fc9n7\" (UniqueName: \"kubernetes.io/projected/b80556f6-5b32-486d-8459-f17bf90c4e24-kube-api-access-fc9n7\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.205100 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-hszd6" Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.272415 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 07:57:22 crc kubenswrapper[4789]: E1122 07:57:22.272555 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 07:57:22.772531853 +0000 UTC m=+57.006932126 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.272787 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:22 crc kubenswrapper[4789]: E1122 07:57:22.273406 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 07:57:22.773394257 +0000 UTC m=+57.007794530 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l4q7j" (UID: "b80556f6-5b32-486d-8459-f17bf90c4e24") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.310859 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-x5b9n" event={"ID":"352693a0-531b-4017-9582-5424c4691795","Type":"ContainerStarted","Data":"ddeb62b6c028a792f86fbb9e7bc9b5284b1e0219b21cc70cee140d0851487d41"} Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.310902 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-x5b9n" event={"ID":"352693a0-531b-4017-9582-5424c4691795","Type":"ContainerStarted","Data":"bd45e4585dac807cc4529c55cd899086677937abe640af0efaf4b44cc2b2fac9"} Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.321360 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-spnzw" event={"ID":"d086acbe-d982-4880-9538-04a184cb4148","Type":"ContainerStarted","Data":"bbad60ca047e9ae86d97618a917812e88cdb079f8da68cf1dfdafbee3b599efe"} Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.326328 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-9lzll" event={"ID":"79e3d4d9-ec8e-4578-871c-a336ad22a2ac","Type":"ContainerStarted","Data":"faddaffaf01f7aae3cb0aa293c3fa3c46b78b7466f94ca0d3b9c95dec171629f"} Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.326372 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-9lzll" event={"ID":"79e3d4d9-ec8e-4578-871c-a336ad22a2ac","Type":"ContainerStarted","Data":"872ed17168027593421d5a8e4bed0013a664591840eaee492be97ec0f30db643"} Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.327004 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-9lzll" Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.345672 4789 patch_prober.go:28] interesting pod/console-operator-58897d9998-9lzll container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.11:8443/readyz\": dial tcp 10.217.0.11:8443: connect: connection refused" start-of-body= Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.345716 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-9lzll" podUID="79e3d4d9-ec8e-4578-871c-a336ad22a2ac" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.11:8443/readyz\": dial tcp 10.217.0.11:8443: connect: connection refused" Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.347228 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-x5b9n" Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.356426 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-v4wls" event={"ID":"be26df1a-a3b3-411b-9e84-df356f3a7af3","Type":"ContainerStarted","Data":"b5cba3dea465264c4087eb9ed6fcefbb9c156315811ddc2307ab2ff788737c25"} Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.357438 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-qh9np" event={"ID":"c562523e-6eb7-4c72-8ba2-f968f171d387","Type":"ContainerStarted","Data":"46428aa8c1359f9fd70dcc85676a7a32514ce9ec8e791aa6e5b536b858ae41fd"} Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.357909 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-szb9d" Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.358605 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wjmrv" event={"ID":"0fb11c40-1e46-42a7-b384-626786bc4e3e","Type":"ContainerStarted","Data":"13f90d488d54e901687f471c3bed0858a43f60230e5785a7ce8ad3f1da84c365"} Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.358628 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wjmrv" event={"ID":"0fb11c40-1e46-42a7-b384-626786bc4e3e","Type":"ContainerStarted","Data":"6f6282e028a77af6520072fa88b5b9313065b070c483885501990230640e0d21"} Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.363297 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-fb49f" event={"ID":"3cfa3a64-3e7a-4336-9ae0-86822095a1b4","Type":"ContainerStarted","Data":"4ba5715e8bbcc567a2231ba1462946fa46c24165b26c5bd7aa97305175fd5c86"} Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.371667 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-pb98x" event={"ID":"9b861750-2517-4cf3-9c78-930480711af0","Type":"ContainerStarted","Data":"ec47c547dee63090f313aec803697265dcce9767c0aacf0c48dddcb8f18c8b4c"} Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.375589 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 07:57:22 crc kubenswrapper[4789]: E1122 07:57:22.375742 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 07:57:22.875713917 +0000 UTC m=+57.110114200 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.375994 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:22 crc kubenswrapper[4789]: E1122 07:57:22.376347 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 07:57:22.876340083 +0000 UTC m=+57.110740346 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l4q7j" (UID: "b80556f6-5b32-486d-8459-f17bf90c4e24") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.377800 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29396625-nrhcv" event={"ID":"ac3d1daa-cb6d-41fc-9798-e687d9b2a21d","Type":"ContainerStarted","Data":"12999710c702d658eae70903ce4222dc88be544709994e1063ed779ff6d938e1"} Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.383379 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2m4cb" event={"ID":"31bd2041-6cb8-4eb6-bb22-5efe6c6009ee","Type":"ContainerStarted","Data":"784420d89b42bdd0cae79424d3dcccac70d965986e6033e34146de32595db2bc"} Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.383703 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6jw6b" event={"ID":"ff518a19-6241-4c65-b2a0-bee6ab939a6c","Type":"ContainerStarted","Data":"4840d00aaf721f649f0682630c272b3758419fd90b4459010baebba5cf976010"} Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.393039 4789 patch_prober.go:28] interesting pod/router-default-5444994796-x5b9n container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 22 07:57:22 crc kubenswrapper[4789]: [-]has-synced failed: reason withheld Nov 22 07:57:22 crc kubenswrapper[4789]: [+]process-running ok Nov 22 07:57:22 crc kubenswrapper[4789]: healthz check failed Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.393094 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-x5b9n" podUID="352693a0-531b-4017-9582-5424c4691795" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.395877 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9n9mc" event={"ID":"dfa022a0-f86e-4f4b-99f5-22b84e9d6558","Type":"ContainerStarted","Data":"b5ef9c6e3892dae4b5d789d4fc23d4a2e5ebdf42b23c7a0d44164fa807d39b66"} Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.402402 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ckmwl" event={"ID":"c0beaf3d-a9eb-47cc-bb53-392e75959282","Type":"ContainerStarted","Data":"f4a4324ffb2c9ccec790a1dc16eee2d8a1b34067f505b7ef476fe2a1cbf2a9d7"} Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.409263 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-tdpht" event={"ID":"83ac1d78-cc4c-48b7-b975-2f88802e2d17","Type":"ContainerStarted","Data":"084102c15dfc9903deae674f52758549f636eed65f1b31bf902c3187bf03cd6a"} Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.428777 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-bsrcz" event={"ID":"cb90f2d2-3321-4118-a449-fdc9343b23e1","Type":"ContainerStarted","Data":"f0afa782888a08744cedb926c014adf2c3f6c330e0911afba87539bef7ab16e5"} Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.433434 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fpw88" event={"ID":"d4d9ab07-53b6-4aef-bcac-c0140d870370","Type":"ContainerStarted","Data":"c9c0e86307f5a170b6e76d3806c5edd7f987998f93cbaa73abaafef7adf10ad8"} Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.435169 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-qwhkx" event={"ID":"ce67cd3e-c2d2-4562-bc26-0a68bce1a1f0","Type":"ContainerStarted","Data":"4db5ea1934fc2e9dcfbcea9a2b44eac4fc72cdbbc65bfd6f9a39c08adf6d0d9c"} Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.437443 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-vh974" event={"ID":"94294737-067b-4a4f-b59b-f9f987fa9127","Type":"ContainerStarted","Data":"b5d43ff5e588f7405f90f8f46ea168642d7c0ae269c193efeff8ba0483a68b09"} Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.440109 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-wnfzj" event={"ID":"55cac199-8203-40b3-8a4f-c571017d998b","Type":"ContainerStarted","Data":"e7832ac8d5e80153fd0aacde4e65f0d420c572e913313f1ea089032b5310a7ea"} Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.445681 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-ksdn9" event={"ID":"1ca00755-c6ae-42d2-833e-fc31b6a480df","Type":"ContainerStarted","Data":"f48c4df0b830ed17509e43354c3d95db42f783af384a0e37ae85b45589edf9df"} Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.449005 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-4s4sh" event={"ID":"23967110-3d55-4767-9fa9-230138c92e42","Type":"ContainerStarted","Data":"26d3983f74d131481af9bae3a85a2475bc0cb6288922a712da2ae1e6904adf38"} Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.452356 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4fbj7" event={"ID":"78d69d0e-bfac-4582-b339-b2e585de5eff","Type":"ContainerStarted","Data":"47ba240916d915ee231e94d7dd8f44d7f61e03dfb64c958247d47f74b3de35d1"} Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.458538 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-kdt8l" event={"ID":"1507f6ee-6a14-4d67-98b3-1e151b7949cc","Type":"ContainerStarted","Data":"f6b9862ab479fe83ecad2e8d8a3b87f5ae40c84396ad27995c21e1df01822c50"} Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.477596 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 07:57:22 crc kubenswrapper[4789]: E1122 07:57:22.479264 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 07:57:22.979242469 +0000 UTC m=+57.213642742 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.582968 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.583379 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.583442 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.583485 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ffef360f-8678-442d-9b27-b8de7646f1a3-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-t5d22\" (UID: \"ffef360f-8678-442d-9b27-b8de7646f1a3\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-t5d22" Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.583548 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.583694 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 07:57:22 crc kubenswrapper[4789]: E1122 07:57:22.585155 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 07:57:23.085136535 +0000 UTC m=+57.319536878 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l4q7j" (UID: "b80556f6-5b32-486d-8459-f17bf90c4e24") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.592355 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.592964 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.593739 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.594649 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ffef360f-8678-442d-9b27-b8de7646f1a3-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-t5d22\" (UID: \"ffef360f-8678-442d-9b27-b8de7646f1a3\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-t5d22" Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.627493 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-zl87n"] Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.629519 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.629764 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-sszl6"] Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.632578 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-tnmrh"] Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.679108 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.686576 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.725644 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-qc925"] Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.726883 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-t5d22" Nov 22 07:57:22 crc kubenswrapper[4789]: E1122 07:57:22.746695 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 07:57:23.246639671 +0000 UTC m=+57.481039944 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.803234 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-h6f5v"] Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.812718 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:22 crc kubenswrapper[4789]: E1122 07:57:22.813655 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 07:57:23.313641098 +0000 UTC m=+57.548041371 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l4q7j" (UID: "b80556f6-5b32-486d-8459-f17bf90c4e24") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.829121 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2dxsq"] Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.843042 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-whfpb"] Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.847319 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-fr4zx"] Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.853168 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-k5c2k"] Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.867053 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.875088 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.887843 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-q7rc4"] Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.899991 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-x5b9n" podStartSLOduration=31.899970237 podStartE2EDuration="31.899970237s" podCreationTimestamp="2025-11-22 07:56:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 07:57:22.899203166 +0000 UTC m=+57.133603439" watchObservedRunningTime="2025-11-22 07:57:22.899970237 +0000 UTC m=+57.134370500" Nov 22 07:57:22 crc kubenswrapper[4789]: W1122 07:57:22.905275 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1d0ebe8e_6cea_43d3_9f15_b67c4cbc29e7.slice/crio-ad94701575b64c551ac4c13d438f78d258c582a42d7398786c2d5327d6e91daa WatchSource:0}: Error finding container ad94701575b64c551ac4c13d438f78d258c582a42d7398786c2d5327d6e91daa: Status 404 returned error can't find the container with id ad94701575b64c551ac4c13d438f78d258c582a42d7398786c2d5327d6e91daa Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.918339 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 07:57:22 crc kubenswrapper[4789]: E1122 07:57:22.920817 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 07:57:23.420790958 +0000 UTC m=+57.655191231 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.920963 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:22 crc kubenswrapper[4789]: E1122 07:57:22.921304 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 07:57:23.421297431 +0000 UTC m=+57.655697704 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l4q7j" (UID: "b80556f6-5b32-486d-8459-f17bf90c4e24") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.961832 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-frls8"] Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.965194 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-77svl"] Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.978873 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-n27w4"] Nov 22 07:57:22 crc kubenswrapper[4789]: I1122 07:57:22.987030 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-d4l6q"] Nov 22 07:57:23 crc kubenswrapper[4789]: I1122 07:57:23.021785 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 07:57:23 crc kubenswrapper[4789]: E1122 07:57:23.022350 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 07:57:23.522333026 +0000 UTC m=+57.756733299 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:23 crc kubenswrapper[4789]: I1122 07:57:23.055178 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=1.055160922 podStartE2EDuration="1.055160922s" podCreationTimestamp="2025-11-22 07:57:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 07:57:23.052021188 +0000 UTC m=+57.286421461" watchObservedRunningTime="2025-11-22 07:57:23.055160922 +0000 UTC m=+57.289561195" Nov 22 07:57:23 crc kubenswrapper[4789]: W1122 07:57:23.056040 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1c517779_505a_4371_9cd7_9537c2fd8763.slice/crio-80452c5a83decb5a121ed06981b2a3f9ace82ff03fb6bc8605c7a82dacc7fcfe WatchSource:0}: Error finding container 80452c5a83decb5a121ed06981b2a3f9ace82ff03fb6bc8605c7a82dacc7fcfe: Status 404 returned error can't find the container with id 80452c5a83decb5a121ed06981b2a3f9ace82ff03fb6bc8605c7a82dacc7fcfe Nov 22 07:57:23 crc kubenswrapper[4789]: W1122 07:57:23.106270 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8adfba92_180a_4f0a_adc0_d1e2bdd62dba.slice/crio-8a90391a1a085047af3a114e07be8da41230aa2d3b36622bb18377d6deea2a32 WatchSource:0}: Error finding container 8a90391a1a085047af3a114e07be8da41230aa2d3b36622bb18377d6deea2a32: Status 404 returned error can't find the container with id 8a90391a1a085047af3a114e07be8da41230aa2d3b36622bb18377d6deea2a32 Nov 22 07:57:23 crc kubenswrapper[4789]: I1122 07:57:23.124847 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:23 crc kubenswrapper[4789]: E1122 07:57:23.125309 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 07:57:23.625299274 +0000 UTC m=+57.859699547 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l4q7j" (UID: "b80556f6-5b32-486d-8459-f17bf90c4e24") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:23 crc kubenswrapper[4789]: W1122 07:57:23.207281 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7f8f3dfa_5fc0_4642_96c0_22b6f3081676.slice/crio-17f20ba7a810ea9af6e9307ea8346b90daffcee6ce10c69de3735ccbba159476 WatchSource:0}: Error finding container 17f20ba7a810ea9af6e9307ea8346b90daffcee6ce10c69de3735ccbba159476: Status 404 returned error can't find the container with id 17f20ba7a810ea9af6e9307ea8346b90daffcee6ce10c69de3735ccbba159476 Nov 22 07:57:23 crc kubenswrapper[4789]: I1122 07:57:23.226951 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 07:57:23 crc kubenswrapper[4789]: E1122 07:57:23.227264 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 07:57:23.727243514 +0000 UTC m=+57.961643787 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:23 crc kubenswrapper[4789]: I1122 07:57:23.275802 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-9lzll" podStartSLOduration=33.275779512 podStartE2EDuration="33.275779512s" podCreationTimestamp="2025-11-22 07:56:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 07:57:23.226409461 +0000 UTC m=+57.460809734" watchObservedRunningTime="2025-11-22 07:57:23.275779512 +0000 UTC m=+57.510179805" Nov 22 07:57:23 crc kubenswrapper[4789]: W1122 07:57:23.276277 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd58d000b_3f2c_44cb_90c5_8023e9f4f5db.slice/crio-682179aebfb60d5ea6e7d9489ecc4cc0d45d515f23ead6884acfcc4589b7d6dd WatchSource:0}: Error finding container 682179aebfb60d5ea6e7d9489ecc4cc0d45d515f23ead6884acfcc4589b7d6dd: Status 404 returned error can't find the container with id 682179aebfb60d5ea6e7d9489ecc4cc0d45d515f23ead6884acfcc4589b7d6dd Nov 22 07:57:23 crc kubenswrapper[4789]: I1122 07:57:23.332890 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:23 crc kubenswrapper[4789]: I1122 07:57:23.333157 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-hszd6"] Nov 22 07:57:23 crc kubenswrapper[4789]: E1122 07:57:23.334187 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 07:57:23.834137017 +0000 UTC m=+58.068537290 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l4q7j" (UID: "b80556f6-5b32-486d-8459-f17bf90c4e24") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:23 crc kubenswrapper[4789]: I1122 07:57:23.353653 4789 patch_prober.go:28] interesting pod/router-default-5444994796-x5b9n container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 22 07:57:23 crc kubenswrapper[4789]: [-]has-synced failed: reason withheld Nov 22 07:57:23 crc kubenswrapper[4789]: [+]process-running ok Nov 22 07:57:23 crc kubenswrapper[4789]: healthz check failed Nov 22 07:57:23 crc kubenswrapper[4789]: I1122 07:57:23.353741 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-x5b9n" podUID="352693a0-531b-4017-9582-5424c4691795" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 22 07:57:23 crc kubenswrapper[4789]: I1122 07:57:23.389365 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-xlf6m"] Nov 22 07:57:23 crc kubenswrapper[4789]: I1122 07:57:23.437355 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 07:57:23 crc kubenswrapper[4789]: E1122 07:57:23.437727 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 07:57:23.93770868 +0000 UTC m=+58.172108953 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:23 crc kubenswrapper[4789]: I1122 07:57:23.480189 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wjmrv" podStartSLOduration=33.480169265 podStartE2EDuration="33.480169265s" podCreationTimestamp="2025-11-22 07:56:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 07:57:23.459521599 +0000 UTC m=+57.693921862" watchObservedRunningTime="2025-11-22 07:57:23.480169265 +0000 UTC m=+57.714569538" Nov 22 07:57:23 crc kubenswrapper[4789]: W1122 07:57:23.483944 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode2a09474_4a22_4f1f_8906_95779f62e570.slice/crio-eab5b6c00541ab77d953bbd19c37427490219652c4237542ebb25fd35f55908b WatchSource:0}: Error finding container eab5b6c00541ab77d953bbd19c37427490219652c4237542ebb25fd35f55908b: Status 404 returned error can't find the container with id eab5b6c00541ab77d953bbd19c37427490219652c4237542ebb25fd35f55908b Nov 22 07:57:23 crc kubenswrapper[4789]: I1122 07:57:23.495713 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-4s4sh" event={"ID":"23967110-3d55-4767-9fa9-230138c92e42","Type":"ContainerStarted","Data":"a635c33cabfc04dd7ad49a8cbc51985f284a781bc5913fb9856dae4473104bce"} Nov 22 07:57:23 crc kubenswrapper[4789]: I1122 07:57:23.496381 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-4s4sh" Nov 22 07:57:23 crc kubenswrapper[4789]: I1122 07:57:23.505509 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-fr4zx" event={"ID":"10112213-cdd9-4bab-a7d5-bc7a3fab5a48","Type":"ContainerStarted","Data":"fe5ed62265c6823794eb5dbcd95720b53d927f11ec44c99bbfae9226b46d2528"} Nov 22 07:57:23 crc kubenswrapper[4789]: I1122 07:57:23.505670 4789 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-4s4sh container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.21:8443/healthz\": dial tcp 10.217.0.21:8443: connect: connection refused" start-of-body= Nov 22 07:57:23 crc kubenswrapper[4789]: I1122 07:57:23.505699 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-4s4sh" podUID="23967110-3d55-4767-9fa9-230138c92e42" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.21:8443/healthz\": dial tcp 10.217.0.21:8443: connect: connection refused" Nov 22 07:57:23 crc kubenswrapper[4789]: I1122 07:57:23.540240 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-pb98x" event={"ID":"9b861750-2517-4cf3-9c78-930480711af0","Type":"ContainerStarted","Data":"ddcb04daed5507caf65f440b3db89e0e66a92775018aca69eac59b75a82ace6f"} Nov 22 07:57:23 crc kubenswrapper[4789]: I1122 07:57:23.541698 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:23 crc kubenswrapper[4789]: E1122 07:57:23.543357 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 07:57:24.043332459 +0000 UTC m=+58.277732752 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l4q7j" (UID: "b80556f6-5b32-486d-8459-f17bf90c4e24") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:23 crc kubenswrapper[4789]: I1122 07:57:23.549039 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-qh9np" event={"ID":"c562523e-6eb7-4c72-8ba2-f968f171d387","Type":"ContainerStarted","Data":"e5e97af5fbdd808bbca0c78ab45b0abf6be70e65a510a217885dc2c04acf7c32"} Nov 22 07:57:23 crc kubenswrapper[4789]: I1122 07:57:23.551675 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-sszl6" event={"ID":"f8d25be9-c80c-4057-b84f-7e78598d2e75","Type":"ContainerStarted","Data":"99c0bd7301dcbc26431a481ed4f0393d12993be1d6890a3b202ce3779eab7401"} Nov 22 07:57:23 crc kubenswrapper[4789]: I1122 07:57:23.574054 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9n9mc" event={"ID":"dfa022a0-f86e-4f4b-99f5-22b84e9d6558","Type":"ContainerStarted","Data":"f5ca2f796bfec09e5a8a5c8cafdd4b517c7c9b5edce1ab85b66cae844b5b6c1d"} Nov 22 07:57:23 crc kubenswrapper[4789]: I1122 07:57:23.577135 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9n9mc" Nov 22 07:57:23 crc kubenswrapper[4789]: I1122 07:57:23.591437 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-q7rc4" event={"ID":"8adfba92-180a-4f0a-adc0-d1e2bdd62dba","Type":"ContainerStarted","Data":"8a90391a1a085047af3a114e07be8da41230aa2d3b36622bb18377d6deea2a32"} Nov 22 07:57:23 crc kubenswrapper[4789]: I1122 07:57:23.615890 4789 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-9n9mc container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.34:8443/healthz\": dial tcp 10.217.0.34:8443: connect: connection refused" start-of-body= Nov 22 07:57:23 crc kubenswrapper[4789]: I1122 07:57:23.616248 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9n9mc" podUID="dfa022a0-f86e-4f4b-99f5-22b84e9d6558" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.34:8443/healthz\": dial tcp 10.217.0.34:8443: connect: connection refused" Nov 22 07:57:23 crc kubenswrapper[4789]: I1122 07:57:23.636139 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6jw6b" event={"ID":"ff518a19-6241-4c65-b2a0-bee6ab939a6c","Type":"ContainerStarted","Data":"84b5c4656b9498b2d4cc958610716dc94438fe159771e1822ade821581cd58c4"} Nov 22 07:57:23 crc kubenswrapper[4789]: I1122 07:57:23.649012 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 07:57:23 crc kubenswrapper[4789]: E1122 07:57:23.649086 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 07:57:24.149067071 +0000 UTC m=+58.383467354 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:23 crc kubenswrapper[4789]: I1122 07:57:23.649245 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:23 crc kubenswrapper[4789]: E1122 07:57:23.650632 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 07:57:24.150621683 +0000 UTC m=+58.385021956 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l4q7j" (UID: "b80556f6-5b32-486d-8459-f17bf90c4e24") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:23 crc kubenswrapper[4789]: I1122 07:57:23.652074 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-whfpb" event={"ID":"065af0aa-e76c-4eab-9a85-d43956522aca","Type":"ContainerStarted","Data":"e3ca3b9377040baa1f80fa553e02afac728c11159d59d8f8ae895903fed6050b"} Nov 22 07:57:23 crc kubenswrapper[4789]: I1122 07:57:23.675739 4789 generic.go:334] "Generic (PLEG): container finished" podID="94294737-067b-4a4f-b59b-f9f987fa9127" containerID="2d053332a6f9bfe82ae6d1f8644cc6fc0547f932368a337896ec78b2d5912859" exitCode=0 Nov 22 07:57:23 crc kubenswrapper[4789]: I1122 07:57:23.675860 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-vh974" event={"ID":"94294737-067b-4a4f-b59b-f9f987fa9127","Type":"ContainerDied","Data":"2d053332a6f9bfe82ae6d1f8644cc6fc0547f932368a337896ec78b2d5912859"} Nov 22 07:57:23 crc kubenswrapper[4789]: I1122 07:57:23.681670 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-frls8" event={"ID":"7dc5fba4-2ba8-4fa2-a76f-0ad8d1196838","Type":"ContainerStarted","Data":"9f8272d1bd1683da299e84ce4964e481ce7f0f9e746778fc4938b494f8e45f15"} Nov 22 07:57:23 crc kubenswrapper[4789]: I1122 07:57:23.687163 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-tdpht" event={"ID":"83ac1d78-cc4c-48b7-b975-2f88802e2d17","Type":"ContainerStarted","Data":"54620f1f3b277d33854c3b2d2582c694baad3604484132173d061c2656aa3c6e"} Nov 22 07:57:23 crc kubenswrapper[4789]: I1122 07:57:23.688060 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-tdpht" Nov 22 07:57:23 crc kubenswrapper[4789]: I1122 07:57:23.702663 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-bsrcz" event={"ID":"cb90f2d2-3321-4118-a449-fdc9343b23e1","Type":"ContainerStarted","Data":"62433bbd6689c220da724e82b5aacaf850eed5d9d90c9d0d898bb7c89c2e1f30"} Nov 22 07:57:23 crc kubenswrapper[4789]: I1122 07:57:23.704076 4789 patch_prober.go:28] interesting pod/downloads-7954f5f757-tdpht container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Nov 22 07:57:23 crc kubenswrapper[4789]: I1122 07:57:23.704128 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-tdpht" podUID="83ac1d78-cc4c-48b7-b975-2f88802e2d17" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Nov 22 07:57:23 crc kubenswrapper[4789]: I1122 07:57:23.706034 4789 generic.go:334] "Generic (PLEG): container finished" podID="ce67cd3e-c2d2-4562-bc26-0a68bce1a1f0" containerID="3c71421d87a09e0996ec8b9c7a9e1cfd8f3c6999ea89abc266934d791c288997" exitCode=0 Nov 22 07:57:23 crc kubenswrapper[4789]: I1122 07:57:23.706080 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-qwhkx" event={"ID":"ce67cd3e-c2d2-4562-bc26-0a68bce1a1f0","Type":"ContainerDied","Data":"3c71421d87a09e0996ec8b9c7a9e1cfd8f3c6999ea89abc266934d791c288997"} Nov 22 07:57:23 crc kubenswrapper[4789]: I1122 07:57:23.708716 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-tnmrh" event={"ID":"d32dc4fd-3e62-402c-a1d9-8f2c83f2e334","Type":"ContainerStarted","Data":"fcc181c6cdb0218de58e9ca29f2bdfef1f21d019d8e144bad69f16b92b128fb6"} Nov 22 07:57:23 crc kubenswrapper[4789]: I1122 07:57:23.740511 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fpw88" event={"ID":"d4d9ab07-53b6-4aef-bcac-c0140d870370","Type":"ContainerStarted","Data":"855070c6843b68f8678ae5705eeb15a639521a84fa045cd6e5e761646a80d863"} Nov 22 07:57:23 crc kubenswrapper[4789]: I1122 07:57:23.740945 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fpw88" Nov 22 07:57:23 crc kubenswrapper[4789]: I1122 07:57:23.742808 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-fg7wj" event={"ID":"dd618778-dc82-4797-b737-14231eefea15","Type":"ContainerStarted","Data":"5bca88586b0f4c1cf9bfdc54f68f4ca86c5f72fdd93414ff51ec3eb774da0eeb"} Nov 22 07:57:23 crc kubenswrapper[4789]: I1122 07:57:23.744724 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-77svl" event={"ID":"9eb8d5d8-3ad7-41d1-9bb2-2d6865619967","Type":"ContainerStarted","Data":"1db435158b0366b3afe5909488c9c5c268a79c71daf279844c5c6598b7a1dc7c"} Nov 22 07:57:23 crc kubenswrapper[4789]: I1122 07:57:23.748127 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-t5d22"] Nov 22 07:57:23 crc kubenswrapper[4789]: I1122 07:57:23.751091 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 07:57:23 crc kubenswrapper[4789]: E1122 07:57:23.752055 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 07:57:24.252039939 +0000 UTC m=+58.486440212 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:23 crc kubenswrapper[4789]: I1122 07:57:23.757442 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zl87n" event={"ID":"f2c80a41-3ceb-4420-a46e-566ff4a28dcb","Type":"ContainerStarted","Data":"25bee47db6cf2b299b53cab2a37f87a55d6e2e6d8b93743791671696f49d007e"} Nov 22 07:57:23 crc kubenswrapper[4789]: I1122 07:57:23.761009 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ckmwl" event={"ID":"c0beaf3d-a9eb-47cc-bb53-392e75959282","Type":"ContainerStarted","Data":"5acb88396dbb0e3804cf969f104a08e80f7c176d7a38c40669334764c632366d"} Nov 22 07:57:23 crc kubenswrapper[4789]: I1122 07:57:23.765198 4789 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-fpw88 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.9:8443/healthz\": dial tcp 10.217.0.9:8443: connect: connection refused" start-of-body= Nov 22 07:57:23 crc kubenswrapper[4789]: I1122 07:57:23.765249 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fpw88" podUID="d4d9ab07-53b6-4aef-bcac-c0140d870370" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.9:8443/healthz\": dial tcp 10.217.0.9:8443: connect: connection refused" Nov 22 07:57:23 crc kubenswrapper[4789]: I1122 07:57:23.769269 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4fbj7" event={"ID":"78d69d0e-bfac-4582-b339-b2e585de5eff","Type":"ContainerStarted","Data":"d466778bebf6c1c883bf2a5d21cb61bbf0df7f6b20ddb423fb55566fe4885c45"} Nov 22 07:57:23 crc kubenswrapper[4789]: I1122 07:57:23.786422 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-n27w4" event={"ID":"d58d000b-3f2c-44cb-90c5-8023e9f4f5db","Type":"ContainerStarted","Data":"682179aebfb60d5ea6e7d9489ecc4cc0d45d515f23ead6884acfcc4589b7d6dd"} Nov 22 07:57:23 crc kubenswrapper[4789]: I1122 07:57:23.790790 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9n9mc" podStartSLOduration=32.790767143 podStartE2EDuration="32.790767143s" podCreationTimestamp="2025-11-22 07:56:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 07:57:23.775973714 +0000 UTC m=+58.010373987" watchObservedRunningTime="2025-11-22 07:57:23.790767143 +0000 UTC m=+58.025167416" Nov 22 07:57:23 crc kubenswrapper[4789]: I1122 07:57:23.802773 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-szb9d"] Nov 22 07:57:23 crc kubenswrapper[4789]: I1122 07:57:23.806659 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2m4cb" event={"ID":"31bd2041-6cb8-4eb6-bb22-5efe6c6009ee","Type":"ContainerStarted","Data":"f31793b2cf53402a008219043c0dc12f4a4be862a5e57101f336fb65a713bcdc"} Nov 22 07:57:23 crc kubenswrapper[4789]: I1122 07:57:23.808574 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-d4l6q" event={"ID":"7f8f3dfa-5fc0-4642-96c0-22b6f3081676","Type":"ContainerStarted","Data":"17f20ba7a810ea9af6e9307ea8346b90daffcee6ce10c69de3735ccbba159476"} Nov 22 07:57:23 crc kubenswrapper[4789]: I1122 07:57:23.811447 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-h6f5v" event={"ID":"cb9c37ae-6151-452f-b467-6b7d14bbb2f8","Type":"ContainerStarted","Data":"e7a638f23fa878044df3447199a33df591ebbcde4878ee533f2740faa99feadd"} Nov 22 07:57:23 crc kubenswrapper[4789]: I1122 07:57:23.823829 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-wnfzj" event={"ID":"55cac199-8203-40b3-8a4f-c571017d998b","Type":"ContainerStarted","Data":"a9caf00327bdcea8f671019a3e4ffeb6fb58a2fb6bc5f3641a2a64cdfadfb78d"} Nov 22 07:57:23 crc kubenswrapper[4789]: I1122 07:57:23.824579 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-wnfzj" Nov 22 07:57:23 crc kubenswrapper[4789]: I1122 07:57:23.834363 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-ksdn9" event={"ID":"1ca00755-c6ae-42d2-833e-fc31b6a480df","Type":"ContainerStarted","Data":"28e202c25ff322832895da613a0436eb2f2df4d318c222bbcd576d6d0e6ade05"} Nov 22 07:57:23 crc kubenswrapper[4789]: I1122 07:57:23.843520 4789 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-wnfzj container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.20:6443/healthz\": dial tcp 10.217.0.20:6443: connect: connection refused" start-of-body= Nov 22 07:57:23 crc kubenswrapper[4789]: I1122 07:57:23.843566 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-wnfzj" podUID="55cac199-8203-40b3-8a4f-c571017d998b" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.20:6443/healthz\": dial tcp 10.217.0.20:6443: connect: connection refused" Nov 22 07:57:23 crc kubenswrapper[4789]: I1122 07:57:23.853734 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:23 crc kubenswrapper[4789]: I1122 07:57:23.855581 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-qh9np" podStartSLOduration=5.855560221 podStartE2EDuration="5.855560221s" podCreationTimestamp="2025-11-22 07:57:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 07:57:23.819390065 +0000 UTC m=+58.053790338" watchObservedRunningTime="2025-11-22 07:57:23.855560221 +0000 UTC m=+58.089960494" Nov 22 07:57:23 crc kubenswrapper[4789]: I1122 07:57:23.856081 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-k5c2k" event={"ID":"0776ae85-c7b7-4d22-b1fb-40b56785205a","Type":"ContainerStarted","Data":"669048e5201aa1146ec8e37a624f27bd6ad8836a204e4e397048c25544231ae4"} Nov 22 07:57:23 crc kubenswrapper[4789]: E1122 07:57:23.856119 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 07:57:24.356106595 +0000 UTC m=+58.590506868 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l4q7j" (UID: "b80556f6-5b32-486d-8459-f17bf90c4e24") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:23 crc kubenswrapper[4789]: I1122 07:57:23.867745 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2dxsq" event={"ID":"1c517779-505a-4371-9cd7-9537c2fd8763","Type":"ContainerStarted","Data":"80452c5a83decb5a121ed06981b2a3f9ace82ff03fb6bc8605c7a82dacc7fcfe"} Nov 22 07:57:23 crc kubenswrapper[4789]: I1122 07:57:23.881522 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-tdpht" podStartSLOduration=33.88150046 podStartE2EDuration="33.88150046s" podCreationTimestamp="2025-11-22 07:56:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 07:57:23.856580968 +0000 UTC m=+58.090981241" watchObservedRunningTime="2025-11-22 07:57:23.88150046 +0000 UTC m=+58.115900733" Nov 22 07:57:23 crc kubenswrapper[4789]: I1122 07:57:23.886454 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-hszd6" event={"ID":"bf1ce57a-0abc-425b-a274-ca4681cd487d","Type":"ContainerStarted","Data":"3f292219c55f59324af5338215aacf928229d1a8efab6f236b8bf1f8b577982d"} Nov 22 07:57:23 crc kubenswrapper[4789]: I1122 07:57:23.893409 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/cni-sysctl-allowlist-ds-gcnts" event={"ID":"75fa6283-acd8-41bc-9803-2f55119f6828","Type":"ContainerStarted","Data":"7ce73911b860efe1f373ef3f3bed89467e181e04382094247fab65fcc1777c66"} Nov 22 07:57:23 crc kubenswrapper[4789]: I1122 07:57:23.896047 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-pb98x" podStartSLOduration=33.896034102 podStartE2EDuration="33.896034102s" podCreationTimestamp="2025-11-22 07:56:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 07:57:23.895652222 +0000 UTC m=+58.130052525" watchObservedRunningTime="2025-11-22 07:57:23.896034102 +0000 UTC m=+58.130434495" Nov 22 07:57:23 crc kubenswrapper[4789]: I1122 07:57:23.901586 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-v4wls" event={"ID":"be26df1a-a3b3-411b-9e84-df356f3a7af3","Type":"ContainerStarted","Data":"1393dedfa16a3c43f0aeb30dfd9b6930d81e999d539234b98254a16e989a484b"} Nov 22 07:57:23 crc kubenswrapper[4789]: W1122 07:57:23.904051 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podffef360f_8678_442d_9b27_b8de7646f1a3.slice/crio-1f8ad98affb59d63a038b512930e5efd9d76f2c2faee73ae5a9f26170d147ed4 WatchSource:0}: Error finding container 1f8ad98affb59d63a038b512930e5efd9d76f2c2faee73ae5a9f26170d147ed4: Status 404 returned error can't find the container with id 1f8ad98affb59d63a038b512930e5efd9d76f2c2faee73ae5a9f26170d147ed4 Nov 22 07:57:23 crc kubenswrapper[4789]: I1122 07:57:23.904264 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-qc925" event={"ID":"1d0ebe8e-6cea-43d3-9f15-b67c4cbc29e7","Type":"ContainerStarted","Data":"ad94701575b64c551ac4c13d438f78d258c582a42d7398786c2d5327d6e91daa"} Nov 22 07:57:23 crc kubenswrapper[4789]: I1122 07:57:23.908824 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-9lzll" Nov 22 07:57:23 crc kubenswrapper[4789]: W1122 07:57:23.911251 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b6479f0_333b_4a96_9adf_2099afdc2447.slice/crio-b45a8b4db5c794a73e6c05c4f66517dd39fb585d3531b0c2cf1cf9eefd82969e WatchSource:0}: Error finding container b45a8b4db5c794a73e6c05c4f66517dd39fb585d3531b0c2cf1cf9eefd82969e: Status 404 returned error can't find the container with id b45a8b4db5c794a73e6c05c4f66517dd39fb585d3531b0c2cf1cf9eefd82969e Nov 22 07:57:23 crc kubenswrapper[4789]: I1122 07:57:23.954554 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 07:57:23 crc kubenswrapper[4789]: E1122 07:57:23.962161 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 07:57:24.462137465 +0000 UTC m=+58.696537818 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:24 crc kubenswrapper[4789]: I1122 07:57:24.024895 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-4s4sh" podStartSLOduration=33.024871186 podStartE2EDuration="33.024871186s" podCreationTimestamp="2025-11-22 07:56:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 07:57:24.023278334 +0000 UTC m=+58.257678617" watchObservedRunningTime="2025-11-22 07:57:24.024871186 +0000 UTC m=+58.259271479" Nov 22 07:57:24 crc kubenswrapper[4789]: I1122 07:57:24.054450 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-v4wls" podStartSLOduration=34.054435165 podStartE2EDuration="34.054435165s" podCreationTimestamp="2025-11-22 07:56:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 07:57:24.052849841 +0000 UTC m=+58.287250114" watchObservedRunningTime="2025-11-22 07:57:24.054435165 +0000 UTC m=+58.288835438" Nov 22 07:57:24 crc kubenswrapper[4789]: I1122 07:57:24.056597 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:24 crc kubenswrapper[4789]: E1122 07:57:24.057741 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 07:57:24.557728063 +0000 UTC m=+58.792128336 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l4q7j" (UID: "b80556f6-5b32-486d-8459-f17bf90c4e24") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:24 crc kubenswrapper[4789]: I1122 07:57:24.110149 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ckmwl" podStartSLOduration=33.110129776 podStartE2EDuration="33.110129776s" podCreationTimestamp="2025-11-22 07:56:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 07:57:24.099096798 +0000 UTC m=+58.333497071" watchObservedRunningTime="2025-11-22 07:57:24.110129776 +0000 UTC m=+58.344530049" Nov 22 07:57:24 crc kubenswrapper[4789]: I1122 07:57:24.158196 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 07:57:24 crc kubenswrapper[4789]: E1122 07:57:24.158725 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 07:57:24.658709057 +0000 UTC m=+58.893109330 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:24 crc kubenswrapper[4789]: I1122 07:57:24.182924 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-qc925" podStartSLOduration=6.182903539 podStartE2EDuration="6.182903539s" podCreationTimestamp="2025-11-22 07:57:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 07:57:24.13843626 +0000 UTC m=+58.372836533" watchObservedRunningTime="2025-11-22 07:57:24.182903539 +0000 UTC m=+58.417303812" Nov 22 07:57:24 crc kubenswrapper[4789]: I1122 07:57:24.183263 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2m4cb" podStartSLOduration=33.183255489 podStartE2EDuration="33.183255489s" podCreationTimestamp="2025-11-22 07:56:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 07:57:24.174711638 +0000 UTC m=+58.409111961" watchObservedRunningTime="2025-11-22 07:57:24.183255489 +0000 UTC m=+58.417655762" Nov 22 07:57:24 crc kubenswrapper[4789]: I1122 07:57:24.223000 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-fg7wj" podStartSLOduration=33.22298069 podStartE2EDuration="33.22298069s" podCreationTimestamp="2025-11-22 07:56:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 07:57:24.220893084 +0000 UTC m=+58.455293367" watchObservedRunningTime="2025-11-22 07:57:24.22298069 +0000 UTC m=+58.457380963" Nov 22 07:57:24 crc kubenswrapper[4789]: I1122 07:57:24.259952 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:24 crc kubenswrapper[4789]: E1122 07:57:24.266384 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 07:57:24.76636261 +0000 UTC m=+59.000762873 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l4q7j" (UID: "b80556f6-5b32-486d-8459-f17bf90c4e24") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:24 crc kubenswrapper[4789]: I1122 07:57:24.271669 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-ksdn9" podStartSLOduration=34.271619663 podStartE2EDuration="34.271619663s" podCreationTimestamp="2025-11-22 07:56:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 07:57:24.260835861 +0000 UTC m=+58.495236144" watchObservedRunningTime="2025-11-22 07:57:24.271619663 +0000 UTC m=+58.506019936" Nov 22 07:57:24 crc kubenswrapper[4789]: I1122 07:57:24.327080 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fpw88" podStartSLOduration=33.327058677 podStartE2EDuration="33.327058677s" podCreationTimestamp="2025-11-22 07:56:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 07:57:24.30230345 +0000 UTC m=+58.536703723" watchObservedRunningTime="2025-11-22 07:57:24.327058677 +0000 UTC m=+58.561458950" Nov 22 07:57:24 crc kubenswrapper[4789]: I1122 07:57:24.351806 4789 patch_prober.go:28] interesting pod/router-default-5444994796-x5b9n container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 22 07:57:24 crc kubenswrapper[4789]: [-]has-synced failed: reason withheld Nov 22 07:57:24 crc kubenswrapper[4789]: [+]process-running ok Nov 22 07:57:24 crc kubenswrapper[4789]: healthz check failed Nov 22 07:57:24 crc kubenswrapper[4789]: I1122 07:57:24.351986 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-x5b9n" podUID="352693a0-531b-4017-9582-5424c4691795" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 22 07:57:24 crc kubenswrapper[4789]: I1122 07:57:24.362380 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 07:57:24 crc kubenswrapper[4789]: E1122 07:57:24.362616 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 07:57:24.862603437 +0000 UTC m=+59.097003710 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:24 crc kubenswrapper[4789]: I1122 07:57:24.377981 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-wnfzj" podStartSLOduration=34.377964991 podStartE2EDuration="34.377964991s" podCreationTimestamp="2025-11-22 07:56:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 07:57:24.376871621 +0000 UTC m=+58.611271914" watchObservedRunningTime="2025-11-22 07:57:24.377964991 +0000 UTC m=+58.612365264" Nov 22 07:57:24 crc kubenswrapper[4789]: I1122 07:57:24.464924 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:24 crc kubenswrapper[4789]: E1122 07:57:24.465266 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 07:57:24.965254935 +0000 UTC m=+59.199655208 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l4q7j" (UID: "b80556f6-5b32-486d-8459-f17bf90c4e24") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:24 crc kubenswrapper[4789]: I1122 07:57:24.565638 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 07:57:24 crc kubenswrapper[4789]: E1122 07:57:24.565851 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 07:57:25.065805487 +0000 UTC m=+59.300205760 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:24 crc kubenswrapper[4789]: I1122 07:57:24.565898 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:24 crc kubenswrapper[4789]: E1122 07:57:24.566232 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 07:57:25.066217818 +0000 UTC m=+59.300618101 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l4q7j" (UID: "b80556f6-5b32-486d-8459-f17bf90c4e24") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:24 crc kubenswrapper[4789]: I1122 07:57:24.666824 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 07:57:24 crc kubenswrapper[4789]: E1122 07:57:24.667019 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 07:57:25.166990066 +0000 UTC m=+59.401390339 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:24 crc kubenswrapper[4789]: I1122 07:57:24.667373 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:24 crc kubenswrapper[4789]: E1122 07:57:24.667667 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 07:57:25.167654695 +0000 UTC m=+59.402054968 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l4q7j" (UID: "b80556f6-5b32-486d-8459-f17bf90c4e24") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:24 crc kubenswrapper[4789]: I1122 07:57:24.768127 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 07:57:24 crc kubenswrapper[4789]: E1122 07:57:24.768569 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 07:57:25.268553416 +0000 UTC m=+59.502953689 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:24 crc kubenswrapper[4789]: I1122 07:57:24.870027 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:24 crc kubenswrapper[4789]: E1122 07:57:24.870451 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 07:57:25.370433253 +0000 UTC m=+59.604833526 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l4q7j" (UID: "b80556f6-5b32-486d-8459-f17bf90c4e24") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:24 crc kubenswrapper[4789]: I1122 07:57:24.937816 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-tnmrh" event={"ID":"d32dc4fd-3e62-402c-a1d9-8f2c83f2e334","Type":"ContainerStarted","Data":"50f21217e8e7540f860514d275afbbce83046aa4316a5f1e14cfe75d62c70059"} Nov 22 07:57:24 crc kubenswrapper[4789]: I1122 07:57:24.939254 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"9aa4beb93864c86c1e2b033f9406cb6b92dc409f9f6e0cd796c3557f1e315eb8"} Nov 22 07:57:24 crc kubenswrapper[4789]: I1122 07:57:24.947994 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-sszl6" event={"ID":"f8d25be9-c80c-4057-b84f-7e78598d2e75","Type":"ContainerStarted","Data":"0681dfffea2946f698448f7ac8856fb40c7be50d1fc6cf4bb9242ce54a174303"} Nov 22 07:57:24 crc kubenswrapper[4789]: I1122 07:57:24.959964 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-fg7wj" event={"ID":"dd618778-dc82-4797-b737-14231eefea15","Type":"ContainerStarted","Data":"d84d0bb8906a0177f5aea71f886c048d2f5497afdb069ea5b948ab5c54225014"} Nov 22 07:57:24 crc kubenswrapper[4789]: I1122 07:57:24.966441 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zl87n" event={"ID":"f2c80a41-3ceb-4420-a46e-566ff4a28dcb","Type":"ContainerStarted","Data":"fe3eb2d4ec10a6d70696fb83c9ce032da4fd1a0b4df296289ea6f3ace4f87f4e"} Nov 22 07:57:24 crc kubenswrapper[4789]: I1122 07:57:24.973144 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 07:57:24 crc kubenswrapper[4789]: E1122 07:57:24.973350 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 07:57:25.473327679 +0000 UTC m=+59.707727952 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:24 crc kubenswrapper[4789]: I1122 07:57:24.973559 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:24 crc kubenswrapper[4789]: E1122 07:57:24.974000 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 07:57:25.473989486 +0000 UTC m=+59.708389759 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l4q7j" (UID: "b80556f6-5b32-486d-8459-f17bf90c4e24") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:24 crc kubenswrapper[4789]: I1122 07:57:24.974878 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6jw6b" event={"ID":"ff518a19-6241-4c65-b2a0-bee6ab939a6c","Type":"ContainerStarted","Data":"1039db0fac521f30ef80d54098f6f38058f7dfcefc91b49cbdaf18850e8a18c3"} Nov 22 07:57:24 crc kubenswrapper[4789]: I1122 07:57:24.985738 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-tnmrh" podStartSLOduration=33.985722763 podStartE2EDuration="33.985722763s" podCreationTimestamp="2025-11-22 07:56:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 07:57:24.983648997 +0000 UTC m=+59.218049270" watchObservedRunningTime="2025-11-22 07:57:24.985722763 +0000 UTC m=+59.220123036" Nov 22 07:57:24 crc kubenswrapper[4789]: I1122 07:57:24.996390 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-bsrcz" event={"ID":"cb90f2d2-3321-4118-a449-fdc9343b23e1","Type":"ContainerStarted","Data":"29c290c7ebcee00e08f92bde804ea2ee22fb696640a878499b0c1b1eb47529a5"} Nov 22 07:57:25 crc kubenswrapper[4789]: I1122 07:57:25.034092 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-d4l6q" event={"ID":"7f8f3dfa-5fc0-4642-96c0-22b6f3081676","Type":"ContainerStarted","Data":"f1ecba34491905eafd61e256c0be15dcf4d995f8d031ab5f3a73e061adaf77ed"} Nov 22 07:57:25 crc kubenswrapper[4789]: I1122 07:57:25.038395 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6jw6b" podStartSLOduration=35.038369533 podStartE2EDuration="35.038369533s" podCreationTimestamp="2025-11-22 07:56:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 07:57:25.034219181 +0000 UTC m=+59.268619454" watchObservedRunningTime="2025-11-22 07:57:25.038369533 +0000 UTC m=+59.272769806" Nov 22 07:57:25 crc kubenswrapper[4789]: I1122 07:57:25.039736 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29396625-nrhcv" event={"ID":"ac3d1daa-cb6d-41fc-9798-e687d9b2a21d","Type":"ContainerStarted","Data":"5cc6b25550f1ec45cf8e7b0340383254d666cc9cb3aa291b0a9b7d2543f04a4e"} Nov 22 07:57:25 crc kubenswrapper[4789]: I1122 07:57:25.074362 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 07:57:25 crc kubenswrapper[4789]: E1122 07:57:25.076543 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 07:57:25.576515792 +0000 UTC m=+59.810916095 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:25 crc kubenswrapper[4789]: I1122 07:57:25.132732 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-bsrcz" podStartSLOduration=34.132715218 podStartE2EDuration="34.132715218s" podCreationTimestamp="2025-11-22 07:56:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 07:57:25.092963695 +0000 UTC m=+59.327363988" watchObservedRunningTime="2025-11-22 07:57:25.132715218 +0000 UTC m=+59.367115491" Nov 22 07:57:25 crc kubenswrapper[4789]: I1122 07:57:25.132949 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29396625-nrhcv" podStartSLOduration=34.132944074 podStartE2EDuration="34.132944074s" podCreationTimestamp="2025-11-22 07:56:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 07:57:25.130676022 +0000 UTC m=+59.365076295" watchObservedRunningTime="2025-11-22 07:57:25.132944074 +0000 UTC m=+59.367344347" Nov 22 07:57:25 crc kubenswrapper[4789]: I1122 07:57:25.140128 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-fb49f" event={"ID":"3cfa3a64-3e7a-4336-9ae0-86822095a1b4","Type":"ContainerStarted","Data":"0a17c93fb10bb027decef52db9505550ece31dfeb5b6ab254e4c42877463ff5f"} Nov 22 07:57:25 crc kubenswrapper[4789]: I1122 07:57:25.177895 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:25 crc kubenswrapper[4789]: E1122 07:57:25.178280 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 07:57:25.678265426 +0000 UTC m=+59.912665699 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l4q7j" (UID: "b80556f6-5b32-486d-8459-f17bf90c4e24") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:25 crc kubenswrapper[4789]: I1122 07:57:25.179975 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-szb9d" event={"ID":"85468946-5921-40d2-bf72-14e0a18ebe6f","Type":"ContainerStarted","Data":"ed40b02404adab8c8239ff77b104ff27d7029f13d78f3a8de07fdafeae7ce51b"} Nov 22 07:57:25 crc kubenswrapper[4789]: I1122 07:57:25.195583 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2dxsq" event={"ID":"1c517779-505a-4371-9cd7-9537c2fd8763","Type":"ContainerStarted","Data":"de31e92efe260108f525cf5c2095f431fd44a9b7ac9c506072b5dc80fa4a95d9"} Nov 22 07:57:25 crc kubenswrapper[4789]: I1122 07:57:25.196546 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2dxsq" Nov 22 07:57:25 crc kubenswrapper[4789]: I1122 07:57:25.204165 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4fbj7" event={"ID":"78d69d0e-bfac-4582-b339-b2e585de5eff","Type":"ContainerStarted","Data":"e197b2962a44acc2c7c7b873946462f9f25d9c43383a69e69d0067bfee3f7bd4"} Nov 22 07:57:25 crc kubenswrapper[4789]: I1122 07:57:25.212959 4789 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-2dxsq container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.36:5443/healthz\": dial tcp 10.217.0.36:5443: connect: connection refused" start-of-body= Nov 22 07:57:25 crc kubenswrapper[4789]: I1122 07:57:25.213006 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2dxsq" podUID="1c517779-505a-4371-9cd7-9537c2fd8763" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.36:5443/healthz\": dial tcp 10.217.0.36:5443: connect: connection refused" Nov 22 07:57:25 crc kubenswrapper[4789]: I1122 07:57:25.220675 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/cni-sysctl-allowlist-ds-gcnts" event={"ID":"75fa6283-acd8-41bc-9803-2f55119f6828","Type":"ContainerStarted","Data":"405ac72a4fd327b3c44cbc8cd4e15338078a9cfceba08604c5a9f00fd7e61592"} Nov 22 07:57:25 crc kubenswrapper[4789]: I1122 07:57:25.221234 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-multus/cni-sysctl-allowlist-ds-gcnts" Nov 22 07:57:25 crc kubenswrapper[4789]: I1122 07:57:25.227624 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2dxsq" podStartSLOduration=34.227607866 podStartE2EDuration="34.227607866s" podCreationTimestamp="2025-11-22 07:56:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 07:57:25.227082673 +0000 UTC m=+59.461482946" watchObservedRunningTime="2025-11-22 07:57:25.227607866 +0000 UTC m=+59.462008129" Nov 22 07:57:25 crc kubenswrapper[4789]: I1122 07:57:25.246044 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-qc925" event={"ID":"1d0ebe8e-6cea-43d3-9f15-b67c4cbc29e7","Type":"ContainerStarted","Data":"09174ed911fba30ff451f4019cab84e4eb0e91bd065f7bd96a73d3477801ad35"} Nov 22 07:57:25 crc kubenswrapper[4789]: I1122 07:57:25.259663 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4fbj7" podStartSLOduration=34.25963812 podStartE2EDuration="34.25963812s" podCreationTimestamp="2025-11-22 07:56:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 07:57:25.257474913 +0000 UTC m=+59.491875186" watchObservedRunningTime="2025-11-22 07:57:25.25963812 +0000 UTC m=+59.494038393" Nov 22 07:57:25 crc kubenswrapper[4789]: I1122 07:57:25.266897 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-h6f5v" Nov 22 07:57:25 crc kubenswrapper[4789]: I1122 07:57:25.269100 4789 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-h6f5v container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.30:8443/healthz\": dial tcp 10.217.0.30:8443: connect: connection refused" start-of-body= Nov 22 07:57:25 crc kubenswrapper[4789]: I1122 07:57:25.269149 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-h6f5v" podUID="cb9c37ae-6151-452f-b467-6b7d14bbb2f8" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.30:8443/healthz\": dial tcp 10.217.0.30:8443: connect: connection refused" Nov 22 07:57:25 crc kubenswrapper[4789]: I1122 07:57:25.279368 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 07:57:25 crc kubenswrapper[4789]: E1122 07:57:25.280623 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 07:57:25.780600546 +0000 UTC m=+60.015000839 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:25 crc kubenswrapper[4789]: I1122 07:57:25.285475 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-t5d22" event={"ID":"ffef360f-8678-442d-9b27-b8de7646f1a3","Type":"ContainerStarted","Data":"1f8ad98affb59d63a038b512930e5efd9d76f2c2faee73ae5a9f26170d147ed4"} Nov 22 07:57:25 crc kubenswrapper[4789]: I1122 07:57:25.328370 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/cni-sysctl-allowlist-ds-gcnts" podStartSLOduration=7.328345044 podStartE2EDuration="7.328345044s" podCreationTimestamp="2025-11-22 07:57:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 07:57:25.315703703 +0000 UTC m=+59.550103976" watchObservedRunningTime="2025-11-22 07:57:25.328345044 +0000 UTC m=+59.562745327" Nov 22 07:57:25 crc kubenswrapper[4789]: I1122 07:57:25.382525 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:25 crc kubenswrapper[4789]: E1122 07:57:25.384066 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 07:57:25.884053976 +0000 UTC m=+60.118454249 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l4q7j" (UID: "b80556f6-5b32-486d-8459-f17bf90c4e24") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:25 crc kubenswrapper[4789]: I1122 07:57:25.385059 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"e4b5b696a34e1599b95f313b2376b81dc928f3e1b097543f28e4d9337f2e87d1"} Nov 22 07:57:25 crc kubenswrapper[4789]: I1122 07:57:25.385587 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-multus/cni-sysctl-allowlist-ds-gcnts" Nov 22 07:57:25 crc kubenswrapper[4789]: I1122 07:57:25.385738 4789 patch_prober.go:28] interesting pod/router-default-5444994796-x5b9n container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 22 07:57:25 crc kubenswrapper[4789]: [-]has-synced failed: reason withheld Nov 22 07:57:25 crc kubenswrapper[4789]: [+]process-running ok Nov 22 07:57:25 crc kubenswrapper[4789]: healthz check failed Nov 22 07:57:25 crc kubenswrapper[4789]: I1122 07:57:25.385781 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-x5b9n" podUID="352693a0-531b-4017-9582-5424c4691795" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 22 07:57:25 crc kubenswrapper[4789]: I1122 07:57:25.402165 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-whfpb" event={"ID":"065af0aa-e76c-4eab-9a85-d43956522aca","Type":"ContainerStarted","Data":"d1f8dd497d56780cb8084db760563c3e69a2cd1c0caf1f5765abbde0c2f0bb49"} Nov 22 07:57:25 crc kubenswrapper[4789]: I1122 07:57:25.411914 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-kdt8l" event={"ID":"1507f6ee-6a14-4d67-98b3-1e151b7949cc","Type":"ContainerStarted","Data":"ce105f494a6013212103dcc1e371f7780f55ccbceb71d57fe9bb61752c4931db"} Nov 22 07:57:25 crc kubenswrapper[4789]: I1122 07:57:25.414104 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-xlf6m" event={"ID":"e2a09474-4a22-4f1f-8906-95779f62e570","Type":"ContainerStarted","Data":"eab5b6c00541ab77d953bbd19c37427490219652c4237542ebb25fd35f55908b"} Nov 22 07:57:25 crc kubenswrapper[4789]: I1122 07:57:25.421830 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-spnzw" event={"ID":"d086acbe-d982-4880-9538-04a184cb4148","Type":"ContainerStarted","Data":"a0ea698b8320d4db5f72d9d9269c23f9f8f9ee69927dca84511b2c4c5aa0a738"} Nov 22 07:57:25 crc kubenswrapper[4789]: I1122 07:57:25.443688 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-h6f5v" podStartSLOduration=34.443672565 podStartE2EDuration="34.443672565s" podCreationTimestamp="2025-11-22 07:56:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 07:57:25.443190961 +0000 UTC m=+59.677591234" watchObservedRunningTime="2025-11-22 07:57:25.443672565 +0000 UTC m=+59.678072838" Nov 22 07:57:25 crc kubenswrapper[4789]: I1122 07:57:25.459538 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"b45a8b4db5c794a73e6c05c4f66517dd39fb585d3531b0c2cf1cf9eefd82969e"} Nov 22 07:57:25 crc kubenswrapper[4789]: I1122 07:57:25.459924 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 07:57:25 crc kubenswrapper[4789]: I1122 07:57:25.482295 4789 patch_prober.go:28] interesting pod/downloads-7954f5f757-tdpht container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Nov 22 07:57:25 crc kubenswrapper[4789]: I1122 07:57:25.482362 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-tdpht" podUID="83ac1d78-cc4c-48b7-b975-2f88802e2d17" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Nov 22 07:57:25 crc kubenswrapper[4789]: I1122 07:57:25.484994 4789 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-4s4sh container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.21:8443/healthz\": dial tcp 10.217.0.21:8443: connect: connection refused" start-of-body= Nov 22 07:57:25 crc kubenswrapper[4789]: I1122 07:57:25.485051 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-4s4sh" podUID="23967110-3d55-4767-9fa9-230138c92e42" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.21:8443/healthz\": dial tcp 10.217.0.21:8443: connect: connection refused" Nov 22 07:57:25 crc kubenswrapper[4789]: I1122 07:57:25.485389 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-kdt8l" podStartSLOduration=34.48537264 podStartE2EDuration="34.48537264s" podCreationTimestamp="2025-11-22 07:56:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 07:57:25.47539028 +0000 UTC m=+59.709790553" watchObservedRunningTime="2025-11-22 07:57:25.48537264 +0000 UTC m=+59.719772913" Nov 22 07:57:25 crc kubenswrapper[4789]: I1122 07:57:25.486061 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 07:57:25 crc kubenswrapper[4789]: E1122 07:57:25.487510 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 07:57:25.987484426 +0000 UTC m=+60.221884699 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:25 crc kubenswrapper[4789]: I1122 07:57:25.496710 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fpw88" Nov 22 07:57:25 crc kubenswrapper[4789]: I1122 07:57:25.540963 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9n9mc" Nov 22 07:57:25 crc kubenswrapper[4789]: I1122 07:57:25.588685 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:25 crc kubenswrapper[4789]: E1122 07:57:25.598843 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 07:57:26.09882176 +0000 UTC m=+60.333222113 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l4q7j" (UID: "b80556f6-5b32-486d-8459-f17bf90c4e24") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:25 crc kubenswrapper[4789]: I1122 07:57:25.621153 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-spnzw" podStartSLOduration=34.621133382 podStartE2EDuration="34.621133382s" podCreationTimestamp="2025-11-22 07:56:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 07:57:25.58033943 +0000 UTC m=+59.814739703" watchObservedRunningTime="2025-11-22 07:57:25.621133382 +0000 UTC m=+59.855533655" Nov 22 07:57:25 crc kubenswrapper[4789]: I1122 07:57:25.621952 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-whfpb" podStartSLOduration=34.621931692 podStartE2EDuration="34.621931692s" podCreationTimestamp="2025-11-22 07:56:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 07:57:25.621435009 +0000 UTC m=+59.855835282" watchObservedRunningTime="2025-11-22 07:57:25.621931692 +0000 UTC m=+59.856331965" Nov 22 07:57:25 crc kubenswrapper[4789]: I1122 07:57:25.646344 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-multus/cni-sysctl-allowlist-ds-gcnts"] Nov 22 07:57:25 crc kubenswrapper[4789]: I1122 07:57:25.690386 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 07:57:25 crc kubenswrapper[4789]: E1122 07:57:25.690714 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 07:57:26.190691057 +0000 UTC m=+60.425091330 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:25 crc kubenswrapper[4789]: I1122 07:57:25.706505 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-wnfzj" Nov 22 07:57:25 crc kubenswrapper[4789]: I1122 07:57:25.795680 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:25 crc kubenswrapper[4789]: E1122 07:57:25.796097 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 07:57:26.296083719 +0000 UTC m=+60.530483992 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l4q7j" (UID: "b80556f6-5b32-486d-8459-f17bf90c4e24") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:25 crc kubenswrapper[4789]: I1122 07:57:25.896456 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 07:57:25 crc kubenswrapper[4789]: E1122 07:57:25.896724 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 07:57:26.396706223 +0000 UTC m=+60.631106496 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:25 crc kubenswrapper[4789]: I1122 07:57:25.997832 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:25 crc kubenswrapper[4789]: E1122 07:57:25.998555 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 07:57:26.49854031 +0000 UTC m=+60.732940583 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l4q7j" (UID: "b80556f6-5b32-486d-8459-f17bf90c4e24") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:26 crc kubenswrapper[4789]: I1122 07:57:26.099376 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 07:57:26 crc kubenswrapper[4789]: E1122 07:57:26.099512 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 07:57:26.599493333 +0000 UTC m=+60.833893606 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:26 crc kubenswrapper[4789]: I1122 07:57:26.099635 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:26 crc kubenswrapper[4789]: E1122 07:57:26.099965 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 07:57:26.599956276 +0000 UTC m=+60.834356549 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l4q7j" (UID: "b80556f6-5b32-486d-8459-f17bf90c4e24") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:26 crc kubenswrapper[4789]: I1122 07:57:26.200774 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 07:57:26 crc kubenswrapper[4789]: E1122 07:57:26.200962 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 07:57:26.70093597 +0000 UTC m=+60.935336243 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:26 crc kubenswrapper[4789]: I1122 07:57:26.201081 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:26 crc kubenswrapper[4789]: E1122 07:57:26.201404 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 07:57:26.701396342 +0000 UTC m=+60.935796615 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l4q7j" (UID: "b80556f6-5b32-486d-8459-f17bf90c4e24") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:26 crc kubenswrapper[4789]: I1122 07:57:26.302155 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 07:57:26 crc kubenswrapper[4789]: E1122 07:57:26.302335 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 07:57:26.802308863 +0000 UTC m=+61.036709136 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:26 crc kubenswrapper[4789]: I1122 07:57:26.302455 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:26 crc kubenswrapper[4789]: E1122 07:57:26.303006 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 07:57:26.802997243 +0000 UTC m=+61.037397516 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l4q7j" (UID: "b80556f6-5b32-486d-8459-f17bf90c4e24") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:26 crc kubenswrapper[4789]: I1122 07:57:26.356000 4789 patch_prober.go:28] interesting pod/router-default-5444994796-x5b9n container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 22 07:57:26 crc kubenswrapper[4789]: [-]has-synced failed: reason withheld Nov 22 07:57:26 crc kubenswrapper[4789]: [+]process-running ok Nov 22 07:57:26 crc kubenswrapper[4789]: healthz check failed Nov 22 07:57:26 crc kubenswrapper[4789]: I1122 07:57:26.356433 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-x5b9n" podUID="352693a0-531b-4017-9582-5424c4691795" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 22 07:57:26 crc kubenswrapper[4789]: I1122 07:57:26.404223 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 07:57:26 crc kubenswrapper[4789]: E1122 07:57:26.404356 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 07:57:26.904338755 +0000 UTC m=+61.138739028 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:26 crc kubenswrapper[4789]: I1122 07:57:26.404404 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:26 crc kubenswrapper[4789]: E1122 07:57:26.404867 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 07:57:26.904860569 +0000 UTC m=+61.139260842 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l4q7j" (UID: "b80556f6-5b32-486d-8459-f17bf90c4e24") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:26 crc kubenswrapper[4789]: I1122 07:57:26.476791 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-spnzw" event={"ID":"d086acbe-d982-4880-9538-04a184cb4148","Type":"ContainerStarted","Data":"7584a73692c92bdb0c7840725a3a2f1c68ad75945ace09a23e8ce70c66c7ceba"} Nov 22 07:57:26 crc kubenswrapper[4789]: I1122 07:57:26.496597 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-vh974" event={"ID":"94294737-067b-4a4f-b59b-f9f987fa9127","Type":"ContainerStarted","Data":"ecea2aff86e49cbd25bb124941780394a45cca2766d4bba3a933c35e7aa83654"} Nov 22 07:57:26 crc kubenswrapper[4789]: I1122 07:57:26.507134 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 07:57:26 crc kubenswrapper[4789]: E1122 07:57:26.507353 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 07:57:27.007317752 +0000 UTC m=+61.241718035 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:26 crc kubenswrapper[4789]: I1122 07:57:26.507469 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:26 crc kubenswrapper[4789]: E1122 07:57:26.507790 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 07:57:27.007773465 +0000 UTC m=+61.242173738 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l4q7j" (UID: "b80556f6-5b32-486d-8459-f17bf90c4e24") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:26 crc kubenswrapper[4789]: I1122 07:57:26.513466 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-fr4zx" event={"ID":"10112213-cdd9-4bab-a7d5-bc7a3fab5a48","Type":"ContainerStarted","Data":"8c26d12d9a145b6b734012365b89a101bf4361cd96d87dfb2145e95e15b2c486"} Nov 22 07:57:26 crc kubenswrapper[4789]: I1122 07:57:26.516059 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-hszd6" event={"ID":"bf1ce57a-0abc-425b-a274-ca4681cd487d","Type":"ContainerStarted","Data":"eb7d42eb0c81ed16e27888746b32cf13854768182d84ce9b73a97c9ce878ee29"} Nov 22 07:57:26 crc kubenswrapper[4789]: I1122 07:57:26.516097 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-hszd6" event={"ID":"bf1ce57a-0abc-425b-a274-ca4681cd487d","Type":"ContainerStarted","Data":"00f7315bd0440fefedf4c1d4783fa6de8f617281879fd9f0f4f7f455add0fadc"} Nov 22 07:57:26 crc kubenswrapper[4789]: I1122 07:57:26.518597 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-fb49f" event={"ID":"3cfa3a64-3e7a-4336-9ae0-86822095a1b4","Type":"ContainerStarted","Data":"f5c838289ac46b6e5d8072968de41fc70ed24b8b6e45f41619b8c9f8ecbd0d68"} Nov 22 07:57:26 crc kubenswrapper[4789]: I1122 07:57:26.520885 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-k5c2k" event={"ID":"0776ae85-c7b7-4d22-b1fb-40b56785205a","Type":"ContainerStarted","Data":"0a3773dd31bf80bd8965bd971a845f98827f675982f084c8624249e444a8dc33"} Nov 22 07:57:26 crc kubenswrapper[4789]: I1122 07:57:26.522674 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-szb9d" event={"ID":"85468946-5921-40d2-bf72-14e0a18ebe6f","Type":"ContainerStarted","Data":"6e33eaf04c12de2a0bef16c7f4e3c73fcd902e1e216fbd5091fcf5fd05f708c6"} Nov 22 07:57:26 crc kubenswrapper[4789]: I1122 07:57:26.523883 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-77svl" event={"ID":"9eb8d5d8-3ad7-41d1-9bb2-2d6865619967","Type":"ContainerStarted","Data":"fb23172fe6847c647dc6b91554782ea07d2b1b68ed3374ee83e720a42b9a3f8a"} Nov 22 07:57:26 crc kubenswrapper[4789]: I1122 07:57:26.524074 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-77svl" Nov 22 07:57:26 crc kubenswrapper[4789]: I1122 07:57:26.525698 4789 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-77svl container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.38:8080/healthz\": dial tcp 10.217.0.38:8080: connect: connection refused" start-of-body= Nov 22 07:57:26 crc kubenswrapper[4789]: I1122 07:57:26.525744 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-77svl" podUID="9eb8d5d8-3ad7-41d1-9bb2-2d6865619967" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.38:8080/healthz\": dial tcp 10.217.0.38:8080: connect: connection refused" Nov 22 07:57:26 crc kubenswrapper[4789]: I1122 07:57:26.526004 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-qwhkx" event={"ID":"ce67cd3e-c2d2-4562-bc26-0a68bce1a1f0","Type":"ContainerStarted","Data":"b4bdd35da9f1322b7344112683c4da1e607c97f17ab0b228c2215d504b449ae0"} Nov 22 07:57:26 crc kubenswrapper[4789]: I1122 07:57:26.526134 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-qwhkx" Nov 22 07:57:26 crc kubenswrapper[4789]: I1122 07:57:26.540441 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-sszl6" event={"ID":"f8d25be9-c80c-4057-b84f-7e78598d2e75","Type":"ContainerStarted","Data":"d6ed88370ed71f7c252e1b79e6a851e28cc27658d3eec755dbcacfac6817ccad"} Nov 22 07:57:26 crc kubenswrapper[4789]: I1122 07:57:26.541033 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-sszl6" Nov 22 07:57:26 crc kubenswrapper[4789]: I1122 07:57:26.548113 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"95837c8ce3960dc770e875fcf00fd02b8f45d2ddde9ee6d10dbbcded52d07555"} Nov 22 07:57:26 crc kubenswrapper[4789]: I1122 07:57:26.549581 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-fr4zx" podStartSLOduration=35.549569203 podStartE2EDuration="35.549569203s" podCreationTimestamp="2025-11-22 07:56:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 07:57:26.547807225 +0000 UTC m=+60.782207498" watchObservedRunningTime="2025-11-22 07:57:26.549569203 +0000 UTC m=+60.783969476" Nov 22 07:57:26 crc kubenswrapper[4789]: I1122 07:57:26.554197 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-h6f5v" event={"ID":"cb9c37ae-6151-452f-b467-6b7d14bbb2f8","Type":"ContainerStarted","Data":"346bc0e2601b2acc5a6ddd5a5d5cc4bdffe550708fca2ef39d9ef7adbeb23d46"} Nov 22 07:57:26 crc kubenswrapper[4789]: I1122 07:57:26.555323 4789 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-h6f5v container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.30:8443/healthz\": dial tcp 10.217.0.30:8443: connect: connection refused" start-of-body= Nov 22 07:57:26 crc kubenswrapper[4789]: I1122 07:57:26.555379 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-h6f5v" podUID="cb9c37ae-6151-452f-b467-6b7d14bbb2f8" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.30:8443/healthz\": dial tcp 10.217.0.30:8443: connect: connection refused" Nov 22 07:57:26 crc kubenswrapper[4789]: I1122 07:57:26.560303 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-q7rc4" event={"ID":"8adfba92-180a-4f0a-adc0-d1e2bdd62dba","Type":"ContainerStarted","Data":"bc97d552ddcd6b7223fb0bf44ee01fdaba73e301b04d3470dfa669d8935a0723"} Nov 22 07:57:26 crc kubenswrapper[4789]: I1122 07:57:26.560345 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-q7rc4" event={"ID":"8adfba92-180a-4f0a-adc0-d1e2bdd62dba","Type":"ContainerStarted","Data":"8fc89b69c8b9339e69c0ef7b7b1e167dc56a4109ed2e8f0c023be2a2f7f3dac7"} Nov 22 07:57:26 crc kubenswrapper[4789]: I1122 07:57:26.563223 4789 generic.go:334] "Generic (PLEG): container finished" podID="7dc5fba4-2ba8-4fa2-a76f-0ad8d1196838" containerID="e1bb071a94c1e44a1cc713582c4befa7733d417dba001256a21e65570454e665" exitCode=0 Nov 22 07:57:26 crc kubenswrapper[4789]: I1122 07:57:26.563287 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-frls8" event={"ID":"7dc5fba4-2ba8-4fa2-a76f-0ad8d1196838","Type":"ContainerDied","Data":"e1bb071a94c1e44a1cc713582c4befa7733d417dba001256a21e65570454e665"} Nov 22 07:57:26 crc kubenswrapper[4789]: I1122 07:57:26.563307 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-frls8" event={"ID":"7dc5fba4-2ba8-4fa2-a76f-0ad8d1196838","Type":"ContainerStarted","Data":"31644e79b8165acc61da92eba626af39d9786fe11a8f400472d3ec54e61fdb62"} Nov 22 07:57:26 crc kubenswrapper[4789]: I1122 07:57:26.564549 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-t5d22" event={"ID":"ffef360f-8678-442d-9b27-b8de7646f1a3","Type":"ContainerStarted","Data":"3378b5943bc90f61073c1ea26d52bd66c16c519554b572e0db2fdfbce3f98fb5"} Nov 22 07:57:26 crc kubenswrapper[4789]: I1122 07:57:26.569154 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"a716261ed3980d87de72ff6c439e4de58a029524cdd91e047983fd36947193f9"} Nov 22 07:57:26 crc kubenswrapper[4789]: I1122 07:57:26.577661 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"ce3d47144f0f59231425f0992207411084a64cf3c645812f8bab50fe28f5f84e"} Nov 22 07:57:26 crc kubenswrapper[4789]: I1122 07:57:26.608296 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 07:57:26 crc kubenswrapper[4789]: E1122 07:57:26.609448 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 07:57:27.109432357 +0000 UTC m=+61.343832630 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:26 crc kubenswrapper[4789]: I1122 07:57:26.614421 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zl87n" event={"ID":"f2c80a41-3ceb-4420-a46e-566ff4a28dcb","Type":"ContainerStarted","Data":"486a0c3138e258c4633f1edbd7c3a7a11f13083e607a84b17f9ec747f10d5d7e"} Nov 22 07:57:26 crc kubenswrapper[4789]: I1122 07:57:26.620139 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-hszd6" podStartSLOduration=35.620120595 podStartE2EDuration="35.620120595s" podCreationTimestamp="2025-11-22 07:56:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 07:57:26.577764013 +0000 UTC m=+60.812164296" watchObservedRunningTime="2025-11-22 07:57:26.620120595 +0000 UTC m=+60.854520868" Nov 22 07:57:26 crc kubenswrapper[4789]: I1122 07:57:26.657474 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-xlf6m" event={"ID":"e2a09474-4a22-4f1f-8906-95779f62e570","Type":"ContainerStarted","Data":"1d288eb87961933a65c0e8531d9cc0d4318d4de940cc71100f2b8a6884037aa5"} Nov 22 07:57:26 crc kubenswrapper[4789]: I1122 07:57:26.670888 4789 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-2dxsq container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.36:5443/healthz\": dial tcp 10.217.0.36:5443: connect: connection refused" start-of-body= Nov 22 07:57:26 crc kubenswrapper[4789]: I1122 07:57:26.670955 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2dxsq" podUID="1c517779-505a-4371-9cd7-9537c2fd8763" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.36:5443/healthz\": dial tcp 10.217.0.36:5443: connect: connection refused" Nov 22 07:57:26 crc kubenswrapper[4789]: I1122 07:57:26.671789 4789 patch_prober.go:28] interesting pod/downloads-7954f5f757-tdpht container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Nov 22 07:57:26 crc kubenswrapper[4789]: I1122 07:57:26.671864 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-tdpht" podUID="83ac1d78-cc4c-48b7-b975-2f88802e2d17" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Nov 22 07:57:26 crc kubenswrapper[4789]: I1122 07:57:26.698114 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-4s4sh" Nov 22 07:57:26 crc kubenswrapper[4789]: I1122 07:57:26.706982 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-77svl" podStartSLOduration=35.706962878 podStartE2EDuration="35.706962878s" podCreationTimestamp="2025-11-22 07:56:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 07:57:26.705141119 +0000 UTC m=+60.939541382" watchObservedRunningTime="2025-11-22 07:57:26.706962878 +0000 UTC m=+60.941363151" Nov 22 07:57:26 crc kubenswrapper[4789]: I1122 07:57:26.707868 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-sszl6" podStartSLOduration=35.707862722 podStartE2EDuration="35.707862722s" podCreationTimestamp="2025-11-22 07:56:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 07:57:26.620939778 +0000 UTC m=+60.855340041" watchObservedRunningTime="2025-11-22 07:57:26.707862722 +0000 UTC m=+60.942262995" Nov 22 07:57:26 crc kubenswrapper[4789]: I1122 07:57:26.711469 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:26 crc kubenswrapper[4789]: E1122 07:57:26.715340 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 07:57:27.215326453 +0000 UTC m=+61.449726726 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l4q7j" (UID: "b80556f6-5b32-486d-8459-f17bf90c4e24") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:26 crc kubenswrapper[4789]: I1122 07:57:26.787356 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-qwhkx" podStartSLOduration=36.787341515 podStartE2EDuration="36.787341515s" podCreationTimestamp="2025-11-22 07:56:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 07:57:26.751988212 +0000 UTC m=+60.986388485" watchObservedRunningTime="2025-11-22 07:57:26.787341515 +0000 UTC m=+61.021741788" Nov 22 07:57:26 crc kubenswrapper[4789]: I1122 07:57:26.820207 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 07:57:26 crc kubenswrapper[4789]: E1122 07:57:26.821996 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 07:57:27.32197774 +0000 UTC m=+61.556378013 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:26 crc kubenswrapper[4789]: I1122 07:57:26.838711 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-fb49f" podStartSLOduration=35.83868827 podStartE2EDuration="35.83868827s" podCreationTimestamp="2025-11-22 07:56:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 07:57:26.788043625 +0000 UTC m=+61.022443898" watchObservedRunningTime="2025-11-22 07:57:26.83868827 +0000 UTC m=+61.073088543" Nov 22 07:57:26 crc kubenswrapper[4789]: I1122 07:57:26.889019 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-k5c2k" podStartSLOduration=35.888998917 podStartE2EDuration="35.888998917s" podCreationTimestamp="2025-11-22 07:56:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 07:57:26.842494503 +0000 UTC m=+61.076894786" watchObservedRunningTime="2025-11-22 07:57:26.888998917 +0000 UTC m=+61.123399190" Nov 22 07:57:26 crc kubenswrapper[4789]: I1122 07:57:26.921005 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-d4l6q" podStartSLOduration=35.92098784 podStartE2EDuration="35.92098784s" podCreationTimestamp="2025-11-22 07:56:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 07:57:26.888967367 +0000 UTC m=+61.123367660" watchObservedRunningTime="2025-11-22 07:57:26.92098784 +0000 UTC m=+61.155388113" Nov 22 07:57:26 crc kubenswrapper[4789]: I1122 07:57:26.921850 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:26 crc kubenswrapper[4789]: E1122 07:57:26.922231 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 07:57:27.422218334 +0000 UTC m=+61.656618607 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l4q7j" (UID: "b80556f6-5b32-486d-8459-f17bf90c4e24") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:26 crc kubenswrapper[4789]: I1122 07:57:26.960845 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-q7rc4" podStartSLOduration=36.960825454 podStartE2EDuration="36.960825454s" podCreationTimestamp="2025-11-22 07:56:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 07:57:26.959404576 +0000 UTC m=+61.193804849" watchObservedRunningTime="2025-11-22 07:57:26.960825454 +0000 UTC m=+61.195225727" Nov 22 07:57:27 crc kubenswrapper[4789]: I1122 07:57:27.025500 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 07:57:27 crc kubenswrapper[4789]: E1122 07:57:27.025776 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 07:57:27.525760246 +0000 UTC m=+61.760160519 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:27 crc kubenswrapper[4789]: I1122 07:57:27.025886 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:27 crc kubenswrapper[4789]: E1122 07:57:27.026205 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 07:57:27.526196138 +0000 UTC m=+61.760596411 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l4q7j" (UID: "b80556f6-5b32-486d-8459-f17bf90c4e24") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:27 crc kubenswrapper[4789]: I1122 07:57:27.072651 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zl87n" podStartSLOduration=36.07262672 podStartE2EDuration="36.07262672s" podCreationTimestamp="2025-11-22 07:56:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 07:57:27.02628707 +0000 UTC m=+61.260687353" watchObservedRunningTime="2025-11-22 07:57:27.07262672 +0000 UTC m=+61.307027013" Nov 22 07:57:27 crc kubenswrapper[4789]: I1122 07:57:27.127097 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 07:57:27 crc kubenswrapper[4789]: E1122 07:57:27.127256 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 07:57:27.627234793 +0000 UTC m=+61.861635076 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:27 crc kubenswrapper[4789]: I1122 07:57:27.127460 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:27 crc kubenswrapper[4789]: E1122 07:57:27.127810 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 07:57:27.627800199 +0000 UTC m=+61.862200472 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l4q7j" (UID: "b80556f6-5b32-486d-8459-f17bf90c4e24") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:27 crc kubenswrapper[4789]: I1122 07:57:27.228772 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 07:57:27 crc kubenswrapper[4789]: E1122 07:57:27.228984 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 07:57:27.728952457 +0000 UTC m=+61.963352740 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:27 crc kubenswrapper[4789]: I1122 07:57:27.229030 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:27 crc kubenswrapper[4789]: E1122 07:57:27.229405 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 07:57:27.729393778 +0000 UTC m=+61.963794051 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l4q7j" (UID: "b80556f6-5b32-486d-8459-f17bf90c4e24") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:27 crc kubenswrapper[4789]: I1122 07:57:27.284352 4789 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-qwhkx container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" start-of-body= Nov 22 07:57:27 crc kubenswrapper[4789]: I1122 07:57:27.284401 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-qwhkx" podUID="ce67cd3e-c2d2-4562-bc26-0a68bce1a1f0" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" Nov 22 07:57:27 crc kubenswrapper[4789]: I1122 07:57:27.284403 4789 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-qwhkx container/openshift-config-operator namespace/openshift-config-operator: Liveness probe status=failure output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" start-of-body= Nov 22 07:57:27 crc kubenswrapper[4789]: I1122 07:57:27.284478 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-qwhkx" podUID="ce67cd3e-c2d2-4562-bc26-0a68bce1a1f0" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" Nov 22 07:57:27 crc kubenswrapper[4789]: I1122 07:57:27.329685 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 07:57:27 crc kubenswrapper[4789]: E1122 07:57:27.329863 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 07:57:27.829837528 +0000 UTC m=+62.064237801 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:27 crc kubenswrapper[4789]: I1122 07:57:27.329955 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:27 crc kubenswrapper[4789]: E1122 07:57:27.330289 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 07:57:27.830272979 +0000 UTC m=+62.064673252 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l4q7j" (UID: "b80556f6-5b32-486d-8459-f17bf90c4e24") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:27 crc kubenswrapper[4789]: I1122 07:57:27.350841 4789 patch_prober.go:28] interesting pod/router-default-5444994796-x5b9n container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 22 07:57:27 crc kubenswrapper[4789]: [-]has-synced failed: reason withheld Nov 22 07:57:27 crc kubenswrapper[4789]: [+]process-running ok Nov 22 07:57:27 crc kubenswrapper[4789]: healthz check failed Nov 22 07:57:27 crc kubenswrapper[4789]: I1122 07:57:27.350903 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-x5b9n" podUID="352693a0-531b-4017-9582-5424c4691795" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 22 07:57:27 crc kubenswrapper[4789]: I1122 07:57:27.431044 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 07:57:27 crc kubenswrapper[4789]: E1122 07:57:27.431518 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 07:57:27.93149768 +0000 UTC m=+62.165897963 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:27 crc kubenswrapper[4789]: I1122 07:57:27.532638 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:27 crc kubenswrapper[4789]: E1122 07:57:27.533227 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 07:57:28.033207463 +0000 UTC m=+62.267607746 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l4q7j" (UID: "b80556f6-5b32-486d-8459-f17bf90c4e24") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:27 crc kubenswrapper[4789]: I1122 07:57:27.634279 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 07:57:27 crc kubenswrapper[4789]: E1122 07:57:27.634699 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 07:57:28.13467977 +0000 UTC m=+62.369080043 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:27 crc kubenswrapper[4789]: I1122 07:57:27.676213 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-szb9d" event={"ID":"85468946-5921-40d2-bf72-14e0a18ebe6f","Type":"ContainerStarted","Data":"b69ce0f97099d8d2f3690c9c22a1d8c9cfbaf8b3c381b6ee2db98ee51f889daf"} Nov 22 07:57:27 crc kubenswrapper[4789]: I1122 07:57:27.677198 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-szb9d" Nov 22 07:57:27 crc kubenswrapper[4789]: I1122 07:57:27.688195 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-xlf6m" event={"ID":"e2a09474-4a22-4f1f-8906-95779f62e570","Type":"ContainerStarted","Data":"18d60dfb0f44c634f49a2afa8662a0060289c0b0c4e137aa2aab76c6d15b2be7"} Nov 22 07:57:27 crc kubenswrapper[4789]: I1122 07:57:27.701856 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-szb9d" podStartSLOduration=9.701825781 podStartE2EDuration="9.701825781s" podCreationTimestamp="2025-11-22 07:57:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 07:57:27.696155698 +0000 UTC m=+61.930555971" watchObservedRunningTime="2025-11-22 07:57:27.701825781 +0000 UTC m=+61.936226064" Nov 22 07:57:27 crc kubenswrapper[4789]: I1122 07:57:27.702277 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-t5d22" podStartSLOduration=37.702268253 podStartE2EDuration="37.702268253s" podCreationTimestamp="2025-11-22 07:56:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 07:57:27.072810995 +0000 UTC m=+61.307211268" watchObservedRunningTime="2025-11-22 07:57:27.702268253 +0000 UTC m=+61.936668536" Nov 22 07:57:27 crc kubenswrapper[4789]: I1122 07:57:27.703940 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-multus/cni-sysctl-allowlist-ds-gcnts" podUID="75fa6283-acd8-41bc-9803-2f55119f6828" containerName="kube-multus-additional-cni-plugins" containerID="cri-o://405ac72a4fd327b3c44cbc8cd4e15338078a9cfceba08604c5a9f00fd7e61592" gracePeriod=30 Nov 22 07:57:27 crc kubenswrapper[4789]: I1122 07:57:27.704999 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-vh974" event={"ID":"94294737-067b-4a4f-b59b-f9f987fa9127","Type":"ContainerStarted","Data":"8ef9f22962634e2cf325cb667ca65424681e23dd18b0e8254c7a310f723d18fb"} Nov 22 07:57:27 crc kubenswrapper[4789]: I1122 07:57:27.706876 4789 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-77svl container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.38:8080/healthz\": dial tcp 10.217.0.38:8080: connect: connection refused" start-of-body= Nov 22 07:57:27 crc kubenswrapper[4789]: I1122 07:57:27.706921 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-77svl" podUID="9eb8d5d8-3ad7-41d1-9bb2-2d6865619967" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.38:8080/healthz\": dial tcp 10.217.0.38:8080: connect: connection refused" Nov 22 07:57:27 crc kubenswrapper[4789]: I1122 07:57:27.708321 4789 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-qwhkx container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" start-of-body= Nov 22 07:57:27 crc kubenswrapper[4789]: I1122 07:57:27.708371 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-qwhkx" podUID="ce67cd3e-c2d2-4562-bc26-0a68bce1a1f0" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" Nov 22 07:57:27 crc kubenswrapper[4789]: I1122 07:57:27.724667 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-xlf6m" podStartSLOduration=36.724649197 podStartE2EDuration="36.724649197s" podCreationTimestamp="2025-11-22 07:56:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 07:57:27.723572007 +0000 UTC m=+61.957972280" watchObservedRunningTime="2025-11-22 07:57:27.724649197 +0000 UTC m=+61.959049470" Nov 22 07:57:27 crc kubenswrapper[4789]: I1122 07:57:27.736174 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:27 crc kubenswrapper[4789]: E1122 07:57:27.736573 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 07:57:28.236558757 +0000 UTC m=+62.470959030 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l4q7j" (UID: "b80556f6-5b32-486d-8459-f17bf90c4e24") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:27 crc kubenswrapper[4789]: I1122 07:57:27.774286 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-vh974" podStartSLOduration=37.774262474 podStartE2EDuration="37.774262474s" podCreationTimestamp="2025-11-22 07:56:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 07:57:27.772062516 +0000 UTC m=+62.006462809" watchObservedRunningTime="2025-11-22 07:57:27.774262474 +0000 UTC m=+62.008662747" Nov 22 07:57:27 crc kubenswrapper[4789]: I1122 07:57:27.784616 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-h6f5v" Nov 22 07:57:27 crc kubenswrapper[4789]: I1122 07:57:27.838696 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 07:57:27 crc kubenswrapper[4789]: E1122 07:57:27.839271 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 07:57:28.339250378 +0000 UTC m=+62.573650651 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:27 crc kubenswrapper[4789]: I1122 07:57:27.860377 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-frls8" podStartSLOduration=36.860362137 podStartE2EDuration="36.860362137s" podCreationTimestamp="2025-11-22 07:56:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 07:57:27.821627362 +0000 UTC m=+62.056027655" watchObservedRunningTime="2025-11-22 07:57:27.860362137 +0000 UTC m=+62.094762400" Nov 22 07:57:27 crc kubenswrapper[4789]: I1122 07:57:27.940374 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:27 crc kubenswrapper[4789]: E1122 07:57:27.940903 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 07:57:28.440891089 +0000 UTC m=+62.675291362 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l4q7j" (UID: "b80556f6-5b32-486d-8459-f17bf90c4e24") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:28 crc kubenswrapper[4789]: I1122 07:57:28.041429 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 07:57:28 crc kubenswrapper[4789]: E1122 07:57:28.041783 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 07:57:28.541767379 +0000 UTC m=+62.776167642 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:28 crc kubenswrapper[4789]: I1122 07:57:28.143061 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:28 crc kubenswrapper[4789]: E1122 07:57:28.143440 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 07:57:28.643416521 +0000 UTC m=+62.877816794 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l4q7j" (UID: "b80556f6-5b32-486d-8459-f17bf90c4e24") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:28 crc kubenswrapper[4789]: I1122 07:57:28.244049 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 07:57:28 crc kubenswrapper[4789]: E1122 07:57:28.244227 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 07:57:28.74420206 +0000 UTC m=+62.978602333 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:28 crc kubenswrapper[4789]: I1122 07:57:28.244296 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:28 crc kubenswrapper[4789]: E1122 07:57:28.244650 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 07:57:28.744635781 +0000 UTC m=+62.979036054 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l4q7j" (UID: "b80556f6-5b32-486d-8459-f17bf90c4e24") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:28 crc kubenswrapper[4789]: I1122 07:57:28.345388 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 07:57:28 crc kubenswrapper[4789]: E1122 07:57:28.345558 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 07:57:28.845535242 +0000 UTC m=+63.079935525 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:28 crc kubenswrapper[4789]: I1122 07:57:28.345583 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:28 crc kubenswrapper[4789]: E1122 07:57:28.345959 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 07:57:28.845947084 +0000 UTC m=+63.080347357 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l4q7j" (UID: "b80556f6-5b32-486d-8459-f17bf90c4e24") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:28 crc kubenswrapper[4789]: I1122 07:57:28.353714 4789 patch_prober.go:28] interesting pod/router-default-5444994796-x5b9n container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 22 07:57:28 crc kubenswrapper[4789]: [-]has-synced failed: reason withheld Nov 22 07:57:28 crc kubenswrapper[4789]: [+]process-running ok Nov 22 07:57:28 crc kubenswrapper[4789]: healthz check failed Nov 22 07:57:28 crc kubenswrapper[4789]: I1122 07:57:28.353821 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-x5b9n" podUID="352693a0-531b-4017-9582-5424c4691795" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 22 07:57:28 crc kubenswrapper[4789]: E1122 07:57:28.447418 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 07:57:28.94738236 +0000 UTC m=+63.181782633 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:28 crc kubenswrapper[4789]: I1122 07:57:28.447740 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 07:57:28 crc kubenswrapper[4789]: I1122 07:57:28.448187 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:28 crc kubenswrapper[4789]: E1122 07:57:28.448690 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 07:57:28.948680125 +0000 UTC m=+63.183080398 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l4q7j" (UID: "b80556f6-5b32-486d-8459-f17bf90c4e24") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:28 crc kubenswrapper[4789]: I1122 07:57:28.549286 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 07:57:28 crc kubenswrapper[4789]: E1122 07:57:28.549463 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 07:57:29.049437672 +0000 UTC m=+63.283837945 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:28 crc kubenswrapper[4789]: I1122 07:57:28.549706 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:28 crc kubenswrapper[4789]: E1122 07:57:28.550029 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 07:57:29.050021158 +0000 UTC m=+63.284421431 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l4q7j" (UID: "b80556f6-5b32-486d-8459-f17bf90c4e24") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:28 crc kubenswrapper[4789]: I1122 07:57:28.651014 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 07:57:28 crc kubenswrapper[4789]: E1122 07:57:28.651361 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 07:57:29.151342881 +0000 UTC m=+63.385743154 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:28 crc kubenswrapper[4789]: I1122 07:57:28.710095 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-n27w4" event={"ID":"d58d000b-3f2c-44cb-90c5-8023e9f4f5db","Type":"ContainerStarted","Data":"b291bc64ebbba55d688fd4416894d4b313329bc921084faa36cadbadce08894d"} Nov 22 07:57:28 crc kubenswrapper[4789]: I1122 07:57:28.752340 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:28 crc kubenswrapper[4789]: E1122 07:57:28.752644 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 07:57:29.252631583 +0000 UTC m=+63.487031856 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l4q7j" (UID: "b80556f6-5b32-486d-8459-f17bf90c4e24") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:28 crc kubenswrapper[4789]: I1122 07:57:28.853831 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 07:57:28 crc kubenswrapper[4789]: E1122 07:57:28.854058 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 07:57:29.354025158 +0000 UTC m=+63.588425441 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:28 crc kubenswrapper[4789]: I1122 07:57:28.855788 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:28 crc kubenswrapper[4789]: E1122 07:57:28.856084 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 07:57:29.356069663 +0000 UTC m=+63.590469936 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l4q7j" (UID: "b80556f6-5b32-486d-8459-f17bf90c4e24") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:28 crc kubenswrapper[4789]: I1122 07:57:28.958671 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 07:57:28 crc kubenswrapper[4789]: E1122 07:57:28.958903 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 07:57:29.458872735 +0000 UTC m=+63.693273008 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:28 crc kubenswrapper[4789]: I1122 07:57:28.959017 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:28 crc kubenswrapper[4789]: E1122 07:57:28.959321 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 07:57:29.459310588 +0000 UTC m=+63.693710861 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l4q7j" (UID: "b80556f6-5b32-486d-8459-f17bf90c4e24") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:29 crc kubenswrapper[4789]: I1122 07:57:29.034782 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 22 07:57:29 crc kubenswrapper[4789]: I1122 07:57:29.035519 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 22 07:57:29 crc kubenswrapper[4789]: I1122 07:57:29.037277 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Nov 22 07:57:29 crc kubenswrapper[4789]: I1122 07:57:29.042486 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 22 07:57:29 crc kubenswrapper[4789]: I1122 07:57:29.044185 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Nov 22 07:57:29 crc kubenswrapper[4789]: I1122 07:57:29.060595 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 07:57:29 crc kubenswrapper[4789]: E1122 07:57:29.060823 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 07:57:29.560803225 +0000 UTC m=+63.795203508 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:29 crc kubenswrapper[4789]: I1122 07:57:29.061389 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:29 crc kubenswrapper[4789]: E1122 07:57:29.061914 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 07:57:29.561882134 +0000 UTC m=+63.796282407 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l4q7j" (UID: "b80556f6-5b32-486d-8459-f17bf90c4e24") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:29 crc kubenswrapper[4789]: I1122 07:57:29.162724 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 07:57:29 crc kubenswrapper[4789]: I1122 07:57:29.162979 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3d44f23c-b509-470a-a5e4-6ef336d32d5a-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"3d44f23c-b509-470a-a5e4-6ef336d32d5a\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 22 07:57:29 crc kubenswrapper[4789]: I1122 07:57:29.163001 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3d44f23c-b509-470a-a5e4-6ef336d32d5a-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"3d44f23c-b509-470a-a5e4-6ef336d32d5a\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 22 07:57:29 crc kubenswrapper[4789]: E1122 07:57:29.163105 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 07:57:29.663089423 +0000 UTC m=+63.897489696 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:29 crc kubenswrapper[4789]: I1122 07:57:29.263985 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3d44f23c-b509-470a-a5e4-6ef336d32d5a-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"3d44f23c-b509-470a-a5e4-6ef336d32d5a\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 22 07:57:29 crc kubenswrapper[4789]: I1122 07:57:29.264035 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3d44f23c-b509-470a-a5e4-6ef336d32d5a-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"3d44f23c-b509-470a-a5e4-6ef336d32d5a\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 22 07:57:29 crc kubenswrapper[4789]: I1122 07:57:29.264082 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:29 crc kubenswrapper[4789]: E1122 07:57:29.264435 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 07:57:29.764420197 +0000 UTC m=+63.998820470 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l4q7j" (UID: "b80556f6-5b32-486d-8459-f17bf90c4e24") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:29 crc kubenswrapper[4789]: I1122 07:57:29.264650 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3d44f23c-b509-470a-a5e4-6ef336d32d5a-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"3d44f23c-b509-470a-a5e4-6ef336d32d5a\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 22 07:57:29 crc kubenswrapper[4789]: I1122 07:57:29.293619 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3d44f23c-b509-470a-a5e4-6ef336d32d5a-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"3d44f23c-b509-470a-a5e4-6ef336d32d5a\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 22 07:57:29 crc kubenswrapper[4789]: I1122 07:57:29.350577 4789 patch_prober.go:28] interesting pod/router-default-5444994796-x5b9n container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 22 07:57:29 crc kubenswrapper[4789]: [-]has-synced failed: reason withheld Nov 22 07:57:29 crc kubenswrapper[4789]: [+]process-running ok Nov 22 07:57:29 crc kubenswrapper[4789]: healthz check failed Nov 22 07:57:29 crc kubenswrapper[4789]: I1122 07:57:29.350656 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-x5b9n" podUID="352693a0-531b-4017-9582-5424c4691795" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 22 07:57:29 crc kubenswrapper[4789]: I1122 07:57:29.360921 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 22 07:57:29 crc kubenswrapper[4789]: I1122 07:57:29.365162 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 07:57:29 crc kubenswrapper[4789]: E1122 07:57:29.365338 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 07:57:29.865309078 +0000 UTC m=+64.099709351 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:29 crc kubenswrapper[4789]: I1122 07:57:29.365414 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:29 crc kubenswrapper[4789]: E1122 07:57:29.365685 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 07:57:29.865674688 +0000 UTC m=+64.100074961 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l4q7j" (UID: "b80556f6-5b32-486d-8459-f17bf90c4e24") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:29 crc kubenswrapper[4789]: I1122 07:57:29.466113 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 07:57:29 crc kubenswrapper[4789]: E1122 07:57:29.466664 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 07:57:29.966633801 +0000 UTC m=+64.201034074 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:29 crc kubenswrapper[4789]: I1122 07:57:29.466717 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:29 crc kubenswrapper[4789]: E1122 07:57:29.467235 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 07:57:29.967223456 +0000 UTC m=+64.201623729 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l4q7j" (UID: "b80556f6-5b32-486d-8459-f17bf90c4e24") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:29 crc kubenswrapper[4789]: I1122 07:57:29.568426 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 07:57:29 crc kubenswrapper[4789]: E1122 07:57:29.568800 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 07:57:30.068779206 +0000 UTC m=+64.303179479 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:29 crc kubenswrapper[4789]: I1122 07:57:29.613925 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 22 07:57:29 crc kubenswrapper[4789]: I1122 07:57:29.669819 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:29 crc kubenswrapper[4789]: E1122 07:57:29.670172 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 07:57:30.1701571 +0000 UTC m=+64.404557373 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l4q7j" (UID: "b80556f6-5b32-486d-8459-f17bf90c4e24") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:29 crc kubenswrapper[4789]: I1122 07:57:29.698714 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-7t8sl"] Nov 22 07:57:29 crc kubenswrapper[4789]: I1122 07:57:29.699976 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7t8sl" Nov 22 07:57:29 crc kubenswrapper[4789]: I1122 07:57:29.708062 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 22 07:57:29 crc kubenswrapper[4789]: I1122 07:57:29.710814 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7t8sl"] Nov 22 07:57:29 crc kubenswrapper[4789]: I1122 07:57:29.762529 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"3d44f23c-b509-470a-a5e4-6ef336d32d5a","Type":"ContainerStarted","Data":"66cfdbc5e1bd531eeecd4e8527243e6d443f7cf6bbc3af861c3912ba6e8cd86c"} Nov 22 07:57:29 crc kubenswrapper[4789]: I1122 07:57:29.766485 4789 generic.go:334] "Generic (PLEG): container finished" podID="ac3d1daa-cb6d-41fc-9798-e687d9b2a21d" containerID="5cc6b25550f1ec45cf8e7b0340383254d666cc9cb3aa291b0a9b7d2543f04a4e" exitCode=0 Nov 22 07:57:29 crc kubenswrapper[4789]: I1122 07:57:29.766659 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29396625-nrhcv" event={"ID":"ac3d1daa-cb6d-41fc-9798-e687d9b2a21d","Type":"ContainerDied","Data":"5cc6b25550f1ec45cf8e7b0340383254d666cc9cb3aa291b0a9b7d2543f04a4e"} Nov 22 07:57:29 crc kubenswrapper[4789]: I1122 07:57:29.773452 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 07:57:29 crc kubenswrapper[4789]: E1122 07:57:29.773742 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 07:57:30.273725113 +0000 UTC m=+64.508125386 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:29 crc kubenswrapper[4789]: I1122 07:57:29.779058 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-745s6\" (UniqueName: \"kubernetes.io/projected/7426acc3-0cbe-43f6-a327-886c9de10a94-kube-api-access-745s6\") pod \"community-operators-7t8sl\" (UID: \"7426acc3-0cbe-43f6-a327-886c9de10a94\") " pod="openshift-marketplace/community-operators-7t8sl" Nov 22 07:57:29 crc kubenswrapper[4789]: I1122 07:57:29.779211 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7426acc3-0cbe-43f6-a327-886c9de10a94-catalog-content\") pod \"community-operators-7t8sl\" (UID: \"7426acc3-0cbe-43f6-a327-886c9de10a94\") " pod="openshift-marketplace/community-operators-7t8sl" Nov 22 07:57:29 crc kubenswrapper[4789]: I1122 07:57:29.779295 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:29 crc kubenswrapper[4789]: I1122 07:57:29.779355 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7426acc3-0cbe-43f6-a327-886c9de10a94-utilities\") pod \"community-operators-7t8sl\" (UID: \"7426acc3-0cbe-43f6-a327-886c9de10a94\") " pod="openshift-marketplace/community-operators-7t8sl" Nov 22 07:57:29 crc kubenswrapper[4789]: E1122 07:57:29.779923 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 07:57:30.279908511 +0000 UTC m=+64.514308784 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l4q7j" (UID: "b80556f6-5b32-486d-8459-f17bf90c4e24") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:29 crc kubenswrapper[4789]: I1122 07:57:29.880783 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 07:57:29 crc kubenswrapper[4789]: I1122 07:57:29.880974 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7426acc3-0cbe-43f6-a327-886c9de10a94-catalog-content\") pod \"community-operators-7t8sl\" (UID: \"7426acc3-0cbe-43f6-a327-886c9de10a94\") " pod="openshift-marketplace/community-operators-7t8sl" Nov 22 07:57:29 crc kubenswrapper[4789]: I1122 07:57:29.881064 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7426acc3-0cbe-43f6-a327-886c9de10a94-utilities\") pod \"community-operators-7t8sl\" (UID: \"7426acc3-0cbe-43f6-a327-886c9de10a94\") " pod="openshift-marketplace/community-operators-7t8sl" Nov 22 07:57:29 crc kubenswrapper[4789]: E1122 07:57:29.881176 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 07:57:30.381140881 +0000 UTC m=+64.615541154 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:29 crc kubenswrapper[4789]: I1122 07:57:29.881339 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-745s6\" (UniqueName: \"kubernetes.io/projected/7426acc3-0cbe-43f6-a327-886c9de10a94-kube-api-access-745s6\") pod \"community-operators-7t8sl\" (UID: \"7426acc3-0cbe-43f6-a327-886c9de10a94\") " pod="openshift-marketplace/community-operators-7t8sl" Nov 22 07:57:29 crc kubenswrapper[4789]: I1122 07:57:29.881557 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7426acc3-0cbe-43f6-a327-886c9de10a94-catalog-content\") pod \"community-operators-7t8sl\" (UID: \"7426acc3-0cbe-43f6-a327-886c9de10a94\") " pod="openshift-marketplace/community-operators-7t8sl" Nov 22 07:57:29 crc kubenswrapper[4789]: I1122 07:57:29.881829 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7426acc3-0cbe-43f6-a327-886c9de10a94-utilities\") pod \"community-operators-7t8sl\" (UID: \"7426acc3-0cbe-43f6-a327-886c9de10a94\") " pod="openshift-marketplace/community-operators-7t8sl" Nov 22 07:57:29 crc kubenswrapper[4789]: I1122 07:57:29.901858 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-gbjc4"] Nov 22 07:57:29 crc kubenswrapper[4789]: I1122 07:57:29.903506 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-745s6\" (UniqueName: \"kubernetes.io/projected/7426acc3-0cbe-43f6-a327-886c9de10a94-kube-api-access-745s6\") pod \"community-operators-7t8sl\" (UID: \"7426acc3-0cbe-43f6-a327-886c9de10a94\") " pod="openshift-marketplace/community-operators-7t8sl" Nov 22 07:57:29 crc kubenswrapper[4789]: I1122 07:57:29.905591 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gbjc4" Nov 22 07:57:29 crc kubenswrapper[4789]: I1122 07:57:29.908239 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 22 07:57:29 crc kubenswrapper[4789]: I1122 07:57:29.920146 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-gbjc4"] Nov 22 07:57:29 crc kubenswrapper[4789]: I1122 07:57:29.982650 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4a4f530-1afc-402b-8026-584bba4f8f1a-utilities\") pod \"certified-operators-gbjc4\" (UID: \"d4a4f530-1afc-402b-8026-584bba4f8f1a\") " pod="openshift-marketplace/certified-operators-gbjc4" Nov 22 07:57:29 crc kubenswrapper[4789]: I1122 07:57:29.983109 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:29 crc kubenswrapper[4789]: I1122 07:57:29.983311 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4a4f530-1afc-402b-8026-584bba4f8f1a-catalog-content\") pod \"certified-operators-gbjc4\" (UID: \"d4a4f530-1afc-402b-8026-584bba4f8f1a\") " pod="openshift-marketplace/certified-operators-gbjc4" Nov 22 07:57:29 crc kubenswrapper[4789]: I1122 07:57:29.983433 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pf4qs\" (UniqueName: \"kubernetes.io/projected/d4a4f530-1afc-402b-8026-584bba4f8f1a-kube-api-access-pf4qs\") pod \"certified-operators-gbjc4\" (UID: \"d4a4f530-1afc-402b-8026-584bba4f8f1a\") " pod="openshift-marketplace/certified-operators-gbjc4" Nov 22 07:57:29 crc kubenswrapper[4789]: E1122 07:57:29.983517 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 07:57:30.483494986 +0000 UTC m=+64.717895339 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l4q7j" (UID: "b80556f6-5b32-486d-8459-f17bf90c4e24") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:30 crc kubenswrapper[4789]: I1122 07:57:30.061733 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7t8sl" Nov 22 07:57:30 crc kubenswrapper[4789]: I1122 07:57:30.084172 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 07:57:30 crc kubenswrapper[4789]: E1122 07:57:30.084390 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 07:57:30.584362305 +0000 UTC m=+64.818762578 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:30 crc kubenswrapper[4789]: I1122 07:57:30.084458 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4a4f530-1afc-402b-8026-584bba4f8f1a-catalog-content\") pod \"certified-operators-gbjc4\" (UID: \"d4a4f530-1afc-402b-8026-584bba4f8f1a\") " pod="openshift-marketplace/certified-operators-gbjc4" Nov 22 07:57:30 crc kubenswrapper[4789]: I1122 07:57:30.084495 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pf4qs\" (UniqueName: \"kubernetes.io/projected/d4a4f530-1afc-402b-8026-584bba4f8f1a-kube-api-access-pf4qs\") pod \"certified-operators-gbjc4\" (UID: \"d4a4f530-1afc-402b-8026-584bba4f8f1a\") " pod="openshift-marketplace/certified-operators-gbjc4" Nov 22 07:57:30 crc kubenswrapper[4789]: I1122 07:57:30.084540 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4a4f530-1afc-402b-8026-584bba4f8f1a-utilities\") pod \"certified-operators-gbjc4\" (UID: \"d4a4f530-1afc-402b-8026-584bba4f8f1a\") " pod="openshift-marketplace/certified-operators-gbjc4" Nov 22 07:57:30 crc kubenswrapper[4789]: I1122 07:57:30.084560 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:30 crc kubenswrapper[4789]: E1122 07:57:30.084865 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 07:57:30.584851539 +0000 UTC m=+64.819251812 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l4q7j" (UID: "b80556f6-5b32-486d-8459-f17bf90c4e24") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:30 crc kubenswrapper[4789]: I1122 07:57:30.085262 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4a4f530-1afc-402b-8026-584bba4f8f1a-catalog-content\") pod \"certified-operators-gbjc4\" (UID: \"d4a4f530-1afc-402b-8026-584bba4f8f1a\") " pod="openshift-marketplace/certified-operators-gbjc4" Nov 22 07:57:30 crc kubenswrapper[4789]: I1122 07:57:30.087710 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4a4f530-1afc-402b-8026-584bba4f8f1a-utilities\") pod \"certified-operators-gbjc4\" (UID: \"d4a4f530-1afc-402b-8026-584bba4f8f1a\") " pod="openshift-marketplace/certified-operators-gbjc4" Nov 22 07:57:30 crc kubenswrapper[4789]: I1122 07:57:30.104485 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pf4qs\" (UniqueName: \"kubernetes.io/projected/d4a4f530-1afc-402b-8026-584bba4f8f1a-kube-api-access-pf4qs\") pod \"certified-operators-gbjc4\" (UID: \"d4a4f530-1afc-402b-8026-584bba4f8f1a\") " pod="openshift-marketplace/certified-operators-gbjc4" Nov 22 07:57:30 crc kubenswrapper[4789]: I1122 07:57:30.108545 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-rnkxn"] Nov 22 07:57:30 crc kubenswrapper[4789]: I1122 07:57:30.109623 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rnkxn" Nov 22 07:57:30 crc kubenswrapper[4789]: I1122 07:57:30.124339 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rnkxn"] Nov 22 07:57:30 crc kubenswrapper[4789]: I1122 07:57:30.189722 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 07:57:30 crc kubenswrapper[4789]: I1122 07:57:30.190426 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a078eea-384c-41cc-ae2e-30e6696422a2-catalog-content\") pod \"community-operators-rnkxn\" (UID: \"6a078eea-384c-41cc-ae2e-30e6696422a2\") " pod="openshift-marketplace/community-operators-rnkxn" Nov 22 07:57:30 crc kubenswrapper[4789]: I1122 07:57:30.190471 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a078eea-384c-41cc-ae2e-30e6696422a2-utilities\") pod \"community-operators-rnkxn\" (UID: \"6a078eea-384c-41cc-ae2e-30e6696422a2\") " pod="openshift-marketplace/community-operators-rnkxn" Nov 22 07:57:30 crc kubenswrapper[4789]: I1122 07:57:30.190551 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5hrkf\" (UniqueName: \"kubernetes.io/projected/6a078eea-384c-41cc-ae2e-30e6696422a2-kube-api-access-5hrkf\") pod \"community-operators-rnkxn\" (UID: \"6a078eea-384c-41cc-ae2e-30e6696422a2\") " pod="openshift-marketplace/community-operators-rnkxn" Nov 22 07:57:30 crc kubenswrapper[4789]: E1122 07:57:30.190677 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 07:57:30.690661452 +0000 UTC m=+64.925061725 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:30 crc kubenswrapper[4789]: I1122 07:57:30.231941 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gbjc4" Nov 22 07:57:30 crc kubenswrapper[4789]: I1122 07:57:30.292468 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:30 crc kubenswrapper[4789]: I1122 07:57:30.292506 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a078eea-384c-41cc-ae2e-30e6696422a2-catalog-content\") pod \"community-operators-rnkxn\" (UID: \"6a078eea-384c-41cc-ae2e-30e6696422a2\") " pod="openshift-marketplace/community-operators-rnkxn" Nov 22 07:57:30 crc kubenswrapper[4789]: I1122 07:57:30.292554 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a078eea-384c-41cc-ae2e-30e6696422a2-utilities\") pod \"community-operators-rnkxn\" (UID: \"6a078eea-384c-41cc-ae2e-30e6696422a2\") " pod="openshift-marketplace/community-operators-rnkxn" Nov 22 07:57:30 crc kubenswrapper[4789]: I1122 07:57:30.292576 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5hrkf\" (UniqueName: \"kubernetes.io/projected/6a078eea-384c-41cc-ae2e-30e6696422a2-kube-api-access-5hrkf\") pod \"community-operators-rnkxn\" (UID: \"6a078eea-384c-41cc-ae2e-30e6696422a2\") " pod="openshift-marketplace/community-operators-rnkxn" Nov 22 07:57:30 crc kubenswrapper[4789]: I1122 07:57:30.293235 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a078eea-384c-41cc-ae2e-30e6696422a2-catalog-content\") pod \"community-operators-rnkxn\" (UID: \"6a078eea-384c-41cc-ae2e-30e6696422a2\") " pod="openshift-marketplace/community-operators-rnkxn" Nov 22 07:57:30 crc kubenswrapper[4789]: I1122 07:57:30.293279 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a078eea-384c-41cc-ae2e-30e6696422a2-utilities\") pod \"community-operators-rnkxn\" (UID: \"6a078eea-384c-41cc-ae2e-30e6696422a2\") " pod="openshift-marketplace/community-operators-rnkxn" Nov 22 07:57:30 crc kubenswrapper[4789]: E1122 07:57:30.293309 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 07:57:30.793291989 +0000 UTC m=+65.027692262 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l4q7j" (UID: "b80556f6-5b32-486d-8459-f17bf90c4e24") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:30 crc kubenswrapper[4789]: I1122 07:57:30.299065 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-fnbtg"] Nov 22 07:57:30 crc kubenswrapper[4789]: I1122 07:57:30.299962 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fnbtg" Nov 22 07:57:30 crc kubenswrapper[4789]: I1122 07:57:30.303129 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-qwhkx" Nov 22 07:57:30 crc kubenswrapper[4789]: I1122 07:57:30.310806 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5hrkf\" (UniqueName: \"kubernetes.io/projected/6a078eea-384c-41cc-ae2e-30e6696422a2-kube-api-access-5hrkf\") pod \"community-operators-rnkxn\" (UID: \"6a078eea-384c-41cc-ae2e-30e6696422a2\") " pod="openshift-marketplace/community-operators-rnkxn" Nov 22 07:57:30 crc kubenswrapper[4789]: I1122 07:57:30.328092 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fnbtg"] Nov 22 07:57:30 crc kubenswrapper[4789]: I1122 07:57:30.377728 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 22 07:57:30 crc kubenswrapper[4789]: I1122 07:57:30.378520 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 22 07:57:30 crc kubenswrapper[4789]: I1122 07:57:30.381539 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 22 07:57:30 crc kubenswrapper[4789]: I1122 07:57:30.385437 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Nov 22 07:57:30 crc kubenswrapper[4789]: I1122 07:57:30.385834 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Nov 22 07:57:30 crc kubenswrapper[4789]: I1122 07:57:30.392177 4789 patch_prober.go:28] interesting pod/router-default-5444994796-x5b9n container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 22 07:57:30 crc kubenswrapper[4789]: [-]has-synced failed: reason withheld Nov 22 07:57:30 crc kubenswrapper[4789]: [+]process-running ok Nov 22 07:57:30 crc kubenswrapper[4789]: healthz check failed Nov 22 07:57:30 crc kubenswrapper[4789]: I1122 07:57:30.392238 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-x5b9n" podUID="352693a0-531b-4017-9582-5424c4691795" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 22 07:57:30 crc kubenswrapper[4789]: I1122 07:57:30.394016 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 07:57:30 crc kubenswrapper[4789]: I1122 07:57:30.394173 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ef38f088-0b6a-49a8-811f-3973b8207308-utilities\") pod \"certified-operators-fnbtg\" (UID: \"ef38f088-0b6a-49a8-811f-3973b8207308\") " pod="openshift-marketplace/certified-operators-fnbtg" Nov 22 07:57:30 crc kubenswrapper[4789]: I1122 07:57:30.394206 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ef38f088-0b6a-49a8-811f-3973b8207308-catalog-content\") pod \"certified-operators-fnbtg\" (UID: \"ef38f088-0b6a-49a8-811f-3973b8207308\") " pod="openshift-marketplace/certified-operators-fnbtg" Nov 22 07:57:30 crc kubenswrapper[4789]: I1122 07:57:30.394243 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cfnx7\" (UniqueName: \"kubernetes.io/projected/ef38f088-0b6a-49a8-811f-3973b8207308-kube-api-access-cfnx7\") pod \"certified-operators-fnbtg\" (UID: \"ef38f088-0b6a-49a8-811f-3973b8207308\") " pod="openshift-marketplace/certified-operators-fnbtg" Nov 22 07:57:30 crc kubenswrapper[4789]: E1122 07:57:30.395105 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 07:57:30.895076174 +0000 UTC m=+65.129476457 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:30 crc kubenswrapper[4789]: I1122 07:57:30.426952 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rnkxn" Nov 22 07:57:30 crc kubenswrapper[4789]: I1122 07:57:30.464648 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7t8sl"] Nov 22 07:57:30 crc kubenswrapper[4789]: I1122 07:57:30.496583 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f42980d5-344a-4cf7-b937-431ea412e9ab-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"f42980d5-344a-4cf7-b937-431ea412e9ab\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 22 07:57:30 crc kubenswrapper[4789]: I1122 07:57:30.496643 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f42980d5-344a-4cf7-b937-431ea412e9ab-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"f42980d5-344a-4cf7-b937-431ea412e9ab\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 22 07:57:30 crc kubenswrapper[4789]: I1122 07:57:30.496701 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ef38f088-0b6a-49a8-811f-3973b8207308-utilities\") pod \"certified-operators-fnbtg\" (UID: \"ef38f088-0b6a-49a8-811f-3973b8207308\") " pod="openshift-marketplace/certified-operators-fnbtg" Nov 22 07:57:30 crc kubenswrapper[4789]: I1122 07:57:30.496738 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ef38f088-0b6a-49a8-811f-3973b8207308-catalog-content\") pod \"certified-operators-fnbtg\" (UID: \"ef38f088-0b6a-49a8-811f-3973b8207308\") " pod="openshift-marketplace/certified-operators-fnbtg" Nov 22 07:57:30 crc kubenswrapper[4789]: I1122 07:57:30.496810 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:30 crc kubenswrapper[4789]: I1122 07:57:30.496844 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cfnx7\" (UniqueName: \"kubernetes.io/projected/ef38f088-0b6a-49a8-811f-3973b8207308-kube-api-access-cfnx7\") pod \"certified-operators-fnbtg\" (UID: \"ef38f088-0b6a-49a8-811f-3973b8207308\") " pod="openshift-marketplace/certified-operators-fnbtg" Nov 22 07:57:30 crc kubenswrapper[4789]: I1122 07:57:30.501357 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ef38f088-0b6a-49a8-811f-3973b8207308-catalog-content\") pod \"certified-operators-fnbtg\" (UID: \"ef38f088-0b6a-49a8-811f-3973b8207308\") " pod="openshift-marketplace/certified-operators-fnbtg" Nov 22 07:57:30 crc kubenswrapper[4789]: E1122 07:57:30.501897 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 07:57:31.001877424 +0000 UTC m=+65.236277697 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l4q7j" (UID: "b80556f6-5b32-486d-8459-f17bf90c4e24") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:30 crc kubenswrapper[4789]: I1122 07:57:30.505739 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ef38f088-0b6a-49a8-811f-3973b8207308-utilities\") pod \"certified-operators-fnbtg\" (UID: \"ef38f088-0b6a-49a8-811f-3973b8207308\") " pod="openshift-marketplace/certified-operators-fnbtg" Nov 22 07:57:30 crc kubenswrapper[4789]: I1122 07:57:30.539303 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cfnx7\" (UniqueName: \"kubernetes.io/projected/ef38f088-0b6a-49a8-811f-3973b8207308-kube-api-access-cfnx7\") pod \"certified-operators-fnbtg\" (UID: \"ef38f088-0b6a-49a8-811f-3973b8207308\") " pod="openshift-marketplace/certified-operators-fnbtg" Nov 22 07:57:30 crc kubenswrapper[4789]: I1122 07:57:30.598286 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 07:57:30 crc kubenswrapper[4789]: I1122 07:57:30.598562 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f42980d5-344a-4cf7-b937-431ea412e9ab-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"f42980d5-344a-4cf7-b937-431ea412e9ab\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 22 07:57:30 crc kubenswrapper[4789]: I1122 07:57:30.598665 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f42980d5-344a-4cf7-b937-431ea412e9ab-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"f42980d5-344a-4cf7-b937-431ea412e9ab\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 22 07:57:30 crc kubenswrapper[4789]: I1122 07:57:30.598802 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f42980d5-344a-4cf7-b937-431ea412e9ab-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"f42980d5-344a-4cf7-b937-431ea412e9ab\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 22 07:57:30 crc kubenswrapper[4789]: E1122 07:57:30.598916 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 07:57:31.098859909 +0000 UTC m=+65.333260182 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:30 crc kubenswrapper[4789]: I1122 07:57:30.620073 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f42980d5-344a-4cf7-b937-431ea412e9ab-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"f42980d5-344a-4cf7-b937-431ea412e9ab\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 22 07:57:30 crc kubenswrapper[4789]: I1122 07:57:30.646355 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fnbtg" Nov 22 07:57:30 crc kubenswrapper[4789]: I1122 07:57:30.680625 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-gbjc4"] Nov 22 07:57:30 crc kubenswrapper[4789]: W1122 07:57:30.694333 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd4a4f530_1afc_402b_8026_584bba4f8f1a.slice/crio-d8a0191b9e09a1bbbe54a395bc17206a8c48b10df45cb728e0da0366f87ec72c WatchSource:0}: Error finding container d8a0191b9e09a1bbbe54a395bc17206a8c48b10df45cb728e0da0366f87ec72c: Status 404 returned error can't find the container with id d8a0191b9e09a1bbbe54a395bc17206a8c48b10df45cb728e0da0366f87ec72c Nov 22 07:57:30 crc kubenswrapper[4789]: I1122 07:57:30.720671 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:30 crc kubenswrapper[4789]: I1122 07:57:30.720839 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 22 07:57:30 crc kubenswrapper[4789]: E1122 07:57:30.721284 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 07:57:31.221262002 +0000 UTC m=+65.455662265 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l4q7j" (UID: "b80556f6-5b32-486d-8459-f17bf90c4e24") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:30 crc kubenswrapper[4789]: I1122 07:57:30.781919 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7t8sl" event={"ID":"7426acc3-0cbe-43f6-a327-886c9de10a94","Type":"ContainerStarted","Data":"c0ae897dc3391c4856d5ad1d55fe30b19df432432d8e7eef5135d1cf4e1f7504"} Nov 22 07:57:30 crc kubenswrapper[4789]: I1122 07:57:30.783452 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"3d44f23c-b509-470a-a5e4-6ef336d32d5a","Type":"ContainerStarted","Data":"6538b79f931b1247b40d2d1bd30b618899987616f6ffecf285b5e7c79a10602e"} Nov 22 07:57:30 crc kubenswrapper[4789]: I1122 07:57:30.785380 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gbjc4" event={"ID":"d4a4f530-1afc-402b-8026-584bba4f8f1a","Type":"ContainerStarted","Data":"d8a0191b9e09a1bbbe54a395bc17206a8c48b10df45cb728e0da0366f87ec72c"} Nov 22 07:57:30 crc kubenswrapper[4789]: I1122 07:57:30.804199 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=1.804179604 podStartE2EDuration="1.804179604s" podCreationTimestamp="2025-11-22 07:57:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 07:57:30.802872488 +0000 UTC m=+65.037272761" watchObservedRunningTime="2025-11-22 07:57:30.804179604 +0000 UTC m=+65.038579877" Nov 22 07:57:30 crc kubenswrapper[4789]: I1122 07:57:30.822599 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 07:57:30 crc kubenswrapper[4789]: E1122 07:57:30.823082 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 07:57:31.323058527 +0000 UTC m=+65.557458800 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:30 crc kubenswrapper[4789]: I1122 07:57:30.923957 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:30 crc kubenswrapper[4789]: E1122 07:57:30.924318 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 07:57:31.424302386 +0000 UTC m=+65.658702659 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l4q7j" (UID: "b80556f6-5b32-486d-8459-f17bf90c4e24") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:30 crc kubenswrapper[4789]: I1122 07:57:30.997181 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rnkxn"] Nov 22 07:57:31 crc kubenswrapper[4789]: I1122 07:57:31.018018 4789 patch_prober.go:28] interesting pod/downloads-7954f5f757-tdpht container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Nov 22 07:57:31 crc kubenswrapper[4789]: I1122 07:57:31.018076 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-tdpht" podUID="83ac1d78-cc4c-48b7-b975-2f88802e2d17" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Nov 22 07:57:31 crc kubenswrapper[4789]: I1122 07:57:31.018087 4789 patch_prober.go:28] interesting pod/downloads-7954f5f757-tdpht container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Nov 22 07:57:31 crc kubenswrapper[4789]: I1122 07:57:31.018133 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-tdpht" podUID="83ac1d78-cc4c-48b7-b975-2f88802e2d17" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Nov 22 07:57:31 crc kubenswrapper[4789]: I1122 07:57:31.024956 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 07:57:31 crc kubenswrapper[4789]: E1122 07:57:31.026531 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 07:57:31.526509292 +0000 UTC m=+65.760909565 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:31 crc kubenswrapper[4789]: W1122 07:57:31.032807 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6a078eea_384c_41cc_ae2e_30e6696422a2.slice/crio-8cfc825de34d8dc7d7bfd849784ebe1f9e7a88b1a7382c7bdd503d546ec40b41 WatchSource:0}: Error finding container 8cfc825de34d8dc7d7bfd849784ebe1f9e7a88b1a7382c7bdd503d546ec40b41: Status 404 returned error can't find the container with id 8cfc825de34d8dc7d7bfd849784ebe1f9e7a88b1a7382c7bdd503d546ec40b41 Nov 22 07:57:31 crc kubenswrapper[4789]: I1122 07:57:31.072085 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 22 07:57:31 crc kubenswrapper[4789]: I1122 07:57:31.076079 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-pb98x" Nov 22 07:57:31 crc kubenswrapper[4789]: I1122 07:57:31.076152 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-pb98x" Nov 22 07:57:31 crc kubenswrapper[4789]: I1122 07:57:31.090914 4789 patch_prober.go:28] interesting pod/console-f9d7485db-pb98x container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.19:8443/health\": dial tcp 10.217.0.19:8443: connect: connection refused" start-of-body= Nov 22 07:57:31 crc kubenswrapper[4789]: I1122 07:57:31.090974 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-pb98x" podUID="9b861750-2517-4cf3-9c78-930480711af0" containerName="console" probeResult="failure" output="Get \"https://10.217.0.19:8443/health\": dial tcp 10.217.0.19:8443: connect: connection refused" Nov 22 07:57:31 crc kubenswrapper[4789]: I1122 07:57:31.128059 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:31 crc kubenswrapper[4789]: E1122 07:57:31.128832 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 07:57:31.62881164 +0000 UTC m=+65.863211913 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l4q7j" (UID: "b80556f6-5b32-486d-8459-f17bf90c4e24") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:31 crc kubenswrapper[4789]: W1122 07:57:31.132034 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-podf42980d5_344a_4cf7_b937_431ea412e9ab.slice/crio-284e22ef4912638d79025dc74afc24f8ee70a4b2c98c4fb266bff9373664a1dc WatchSource:0}: Error finding container 284e22ef4912638d79025dc74afc24f8ee70a4b2c98c4fb266bff9373664a1dc: Status 404 returned error can't find the container with id 284e22ef4912638d79025dc74afc24f8ee70a4b2c98c4fb266bff9373664a1dc Nov 22 07:57:31 crc kubenswrapper[4789]: I1122 07:57:31.147154 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396625-nrhcv" Nov 22 07:57:31 crc kubenswrapper[4789]: I1122 07:57:31.194357 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fnbtg"] Nov 22 07:57:31 crc kubenswrapper[4789]: I1122 07:57:31.229511 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ac3d1daa-cb6d-41fc-9798-e687d9b2a21d-secret-volume\") pod \"ac3d1daa-cb6d-41fc-9798-e687d9b2a21d\" (UID: \"ac3d1daa-cb6d-41fc-9798-e687d9b2a21d\") " Nov 22 07:57:31 crc kubenswrapper[4789]: I1122 07:57:31.229587 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kznfz\" (UniqueName: \"kubernetes.io/projected/ac3d1daa-cb6d-41fc-9798-e687d9b2a21d-kube-api-access-kznfz\") pod \"ac3d1daa-cb6d-41fc-9798-e687d9b2a21d\" (UID: \"ac3d1daa-cb6d-41fc-9798-e687d9b2a21d\") " Nov 22 07:57:31 crc kubenswrapper[4789]: I1122 07:57:31.229700 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ac3d1daa-cb6d-41fc-9798-e687d9b2a21d-config-volume\") pod \"ac3d1daa-cb6d-41fc-9798-e687d9b2a21d\" (UID: \"ac3d1daa-cb6d-41fc-9798-e687d9b2a21d\") " Nov 22 07:57:31 crc kubenswrapper[4789]: I1122 07:57:31.231196 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 07:57:31 crc kubenswrapper[4789]: E1122 07:57:31.231346 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 07:57:31.731321224 +0000 UTC m=+65.965721507 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:31 crc kubenswrapper[4789]: I1122 07:57:31.231401 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:31 crc kubenswrapper[4789]: E1122 07:57:31.231821 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 07:57:31.731807628 +0000 UTC m=+65.966207891 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l4q7j" (UID: "b80556f6-5b32-486d-8459-f17bf90c4e24") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:31 crc kubenswrapper[4789]: I1122 07:57:31.232315 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ac3d1daa-cb6d-41fc-9798-e687d9b2a21d-config-volume" (OuterVolumeSpecName: "config-volume") pod "ac3d1daa-cb6d-41fc-9798-e687d9b2a21d" (UID: "ac3d1daa-cb6d-41fc-9798-e687d9b2a21d"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 07:57:31 crc kubenswrapper[4789]: I1122 07:57:31.237012 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac3d1daa-cb6d-41fc-9798-e687d9b2a21d-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "ac3d1daa-cb6d-41fc-9798-e687d9b2a21d" (UID: "ac3d1daa-cb6d-41fc-9798-e687d9b2a21d"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 07:57:31 crc kubenswrapper[4789]: I1122 07:57:31.237119 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac3d1daa-cb6d-41fc-9798-e687d9b2a21d-kube-api-access-kznfz" (OuterVolumeSpecName: "kube-api-access-kznfz") pod "ac3d1daa-cb6d-41fc-9798-e687d9b2a21d" (UID: "ac3d1daa-cb6d-41fc-9798-e687d9b2a21d"). InnerVolumeSpecName "kube-api-access-kznfz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 07:57:31 crc kubenswrapper[4789]: W1122 07:57:31.324205 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef38f088_0b6a_49a8_811f_3973b8207308.slice/crio-95d4319f107342c47d17dbed8a879bb0c84f8a395182231e2c5a3b2ba25c6668 WatchSource:0}: Error finding container 95d4319f107342c47d17dbed8a879bb0c84f8a395182231e2c5a3b2ba25c6668: Status 404 returned error can't find the container with id 95d4319f107342c47d17dbed8a879bb0c84f8a395182231e2c5a3b2ba25c6668 Nov 22 07:57:31 crc kubenswrapper[4789]: I1122 07:57:31.333141 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 07:57:31 crc kubenswrapper[4789]: E1122 07:57:31.333304 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 07:57:31.833277443 +0000 UTC m=+66.067677716 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:31 crc kubenswrapper[4789]: I1122 07:57:31.333570 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:31 crc kubenswrapper[4789]: I1122 07:57:31.333614 4789 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ac3d1daa-cb6d-41fc-9798-e687d9b2a21d-config-volume\") on node \"crc\" DevicePath \"\"" Nov 22 07:57:31 crc kubenswrapper[4789]: I1122 07:57:31.333626 4789 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ac3d1daa-cb6d-41fc-9798-e687d9b2a21d-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 22 07:57:31 crc kubenswrapper[4789]: I1122 07:57:31.333636 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kznfz\" (UniqueName: \"kubernetes.io/projected/ac3d1daa-cb6d-41fc-9798-e687d9b2a21d-kube-api-access-kznfz\") on node \"crc\" DevicePath \"\"" Nov 22 07:57:31 crc kubenswrapper[4789]: E1122 07:57:31.334180 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 07:57:31.834168378 +0000 UTC m=+66.068568711 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l4q7j" (UID: "b80556f6-5b32-486d-8459-f17bf90c4e24") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:31 crc kubenswrapper[4789]: I1122 07:57:31.347954 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-x5b9n" Nov 22 07:57:31 crc kubenswrapper[4789]: I1122 07:57:31.350926 4789 patch_prober.go:28] interesting pod/router-default-5444994796-x5b9n container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 22 07:57:31 crc kubenswrapper[4789]: [-]has-synced failed: reason withheld Nov 22 07:57:31 crc kubenswrapper[4789]: [+]process-running ok Nov 22 07:57:31 crc kubenswrapper[4789]: healthz check failed Nov 22 07:57:31 crc kubenswrapper[4789]: I1122 07:57:31.350997 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-x5b9n" podUID="352693a0-531b-4017-9582-5424c4691795" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 22 07:57:31 crc kubenswrapper[4789]: I1122 07:57:31.435203 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 07:57:31 crc kubenswrapper[4789]: E1122 07:57:31.435383 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 07:57:31.935353595 +0000 UTC m=+66.169753868 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:31 crc kubenswrapper[4789]: I1122 07:57:31.435488 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:31 crc kubenswrapper[4789]: E1122 07:57:31.435911 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 07:57:31.935900151 +0000 UTC m=+66.170300424 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l4q7j" (UID: "b80556f6-5b32-486d-8459-f17bf90c4e24") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:31 crc kubenswrapper[4789]: I1122 07:57:31.467197 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-vh974" Nov 22 07:57:31 crc kubenswrapper[4789]: I1122 07:57:31.467259 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-vh974" Nov 22 07:57:31 crc kubenswrapper[4789]: I1122 07:57:31.537198 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 07:57:31 crc kubenswrapper[4789]: E1122 07:57:31.537366 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 07:57:32.037339335 +0000 UTC m=+66.271739608 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:31 crc kubenswrapper[4789]: I1122 07:57:31.537683 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:31 crc kubenswrapper[4789]: E1122 07:57:31.538029 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 07:57:32.038015813 +0000 UTC m=+66.272416096 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l4q7j" (UID: "b80556f6-5b32-486d-8459-f17bf90c4e24") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:31 crc kubenswrapper[4789]: I1122 07:57:31.538154 4789 patch_prober.go:28] interesting pod/apiserver-76f77b778f-vh974 container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Nov 22 07:57:31 crc kubenswrapper[4789]: [+]log ok Nov 22 07:57:31 crc kubenswrapper[4789]: [+]etcd ok Nov 22 07:57:31 crc kubenswrapper[4789]: [+]poststarthook/start-apiserver-admission-initializer ok Nov 22 07:57:31 crc kubenswrapper[4789]: [+]poststarthook/generic-apiserver-start-informers ok Nov 22 07:57:31 crc kubenswrapper[4789]: [+]poststarthook/max-in-flight-filter ok Nov 22 07:57:31 crc kubenswrapper[4789]: [+]poststarthook/storage-object-count-tracker-hook ok Nov 22 07:57:31 crc kubenswrapper[4789]: [+]poststarthook/image.openshift.io-apiserver-caches ok Nov 22 07:57:31 crc kubenswrapper[4789]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Nov 22 07:57:31 crc kubenswrapper[4789]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Nov 22 07:57:31 crc kubenswrapper[4789]: [+]poststarthook/project.openshift.io-projectcache ok Nov 22 07:57:31 crc kubenswrapper[4789]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Nov 22 07:57:31 crc kubenswrapper[4789]: [-]poststarthook/openshift.io-startinformers failed: reason withheld Nov 22 07:57:31 crc kubenswrapper[4789]: [+]poststarthook/openshift.io-restmapperupdater ok Nov 22 07:57:31 crc kubenswrapper[4789]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Nov 22 07:57:31 crc kubenswrapper[4789]: livez check failed Nov 22 07:57:31 crc kubenswrapper[4789]: I1122 07:57:31.538192 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-vh974" podUID="94294737-067b-4a4f-b59b-f9f987fa9127" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 22 07:57:31 crc kubenswrapper[4789]: I1122 07:57:31.638857 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 07:57:31 crc kubenswrapper[4789]: E1122 07:57:31.639003 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 07:57:32.138971836 +0000 UTC m=+66.373372109 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:31 crc kubenswrapper[4789]: I1122 07:57:31.639269 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:31 crc kubenswrapper[4789]: E1122 07:57:31.640394 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 07:57:32.140372464 +0000 UTC m=+66.374772827 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l4q7j" (UID: "b80556f6-5b32-486d-8459-f17bf90c4e24") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:31 crc kubenswrapper[4789]: I1122 07:57:31.646555 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2dxsq" Nov 22 07:57:31 crc kubenswrapper[4789]: I1122 07:57:31.696076 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-5wgxf"] Nov 22 07:57:31 crc kubenswrapper[4789]: E1122 07:57:31.696400 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac3d1daa-cb6d-41fc-9798-e687d9b2a21d" containerName="collect-profiles" Nov 22 07:57:31 crc kubenswrapper[4789]: I1122 07:57:31.696433 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac3d1daa-cb6d-41fc-9798-e687d9b2a21d" containerName="collect-profiles" Nov 22 07:57:31 crc kubenswrapper[4789]: I1122 07:57:31.696545 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac3d1daa-cb6d-41fc-9798-e687d9b2a21d" containerName="collect-profiles" Nov 22 07:57:31 crc kubenswrapper[4789]: I1122 07:57:31.697481 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5wgxf" Nov 22 07:57:31 crc kubenswrapper[4789]: I1122 07:57:31.700418 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 22 07:57:31 crc kubenswrapper[4789]: I1122 07:57:31.710987 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5wgxf"] Nov 22 07:57:31 crc kubenswrapper[4789]: I1122 07:57:31.741434 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 07:57:31 crc kubenswrapper[4789]: E1122 07:57:31.742347 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 07:57:32.242322992 +0000 UTC m=+66.476723265 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:31 crc kubenswrapper[4789]: I1122 07:57:31.751807 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-frls8" Nov 22 07:57:31 crc kubenswrapper[4789]: I1122 07:57:31.752126 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-frls8" Nov 22 07:57:31 crc kubenswrapper[4789]: I1122 07:57:31.766332 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-frls8" Nov 22 07:57:31 crc kubenswrapper[4789]: I1122 07:57:31.793703 4789 generic.go:334] "Generic (PLEG): container finished" podID="7426acc3-0cbe-43f6-a327-886c9de10a94" containerID="3db8b3ea616e93792b38156af198afd4a8b8c1d8ce0cc030b03c35ad16d449c8" exitCode=0 Nov 22 07:57:31 crc kubenswrapper[4789]: I1122 07:57:31.793829 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7t8sl" event={"ID":"7426acc3-0cbe-43f6-a327-886c9de10a94","Type":"ContainerDied","Data":"3db8b3ea616e93792b38156af198afd4a8b8c1d8ce0cc030b03c35ad16d449c8"} Nov 22 07:57:31 crc kubenswrapper[4789]: I1122 07:57:31.801052 4789 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 22 07:57:31 crc kubenswrapper[4789]: I1122 07:57:31.808738 4789 generic.go:334] "Generic (PLEG): container finished" podID="3d44f23c-b509-470a-a5e4-6ef336d32d5a" containerID="6538b79f931b1247b40d2d1bd30b618899987616f6ffecf285b5e7c79a10602e" exitCode=0 Nov 22 07:57:31 crc kubenswrapper[4789]: I1122 07:57:31.809326 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"3d44f23c-b509-470a-a5e4-6ef336d32d5a","Type":"ContainerDied","Data":"6538b79f931b1247b40d2d1bd30b618899987616f6ffecf285b5e7c79a10602e"} Nov 22 07:57:31 crc kubenswrapper[4789]: I1122 07:57:31.811950 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fnbtg" event={"ID":"ef38f088-0b6a-49a8-811f-3973b8207308","Type":"ContainerStarted","Data":"909ffa9c112c40bf527b814f367d47a9af890485b374aefe90d7f37a1613c140"} Nov 22 07:57:31 crc kubenswrapper[4789]: I1122 07:57:31.812101 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fnbtg" event={"ID":"ef38f088-0b6a-49a8-811f-3973b8207308","Type":"ContainerStarted","Data":"95d4319f107342c47d17dbed8a879bb0c84f8a395182231e2c5a3b2ba25c6668"} Nov 22 07:57:31 crc kubenswrapper[4789]: I1122 07:57:31.813425 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"f42980d5-344a-4cf7-b937-431ea412e9ab","Type":"ContainerStarted","Data":"adcd9e6728b8bdf50a436b27fdf82c0918f7b37c9bd07bbac9eb2397c6646676"} Nov 22 07:57:31 crc kubenswrapper[4789]: I1122 07:57:31.813463 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"f42980d5-344a-4cf7-b937-431ea412e9ab","Type":"ContainerStarted","Data":"284e22ef4912638d79025dc74afc24f8ee70a4b2c98c4fb266bff9373664a1dc"} Nov 22 07:57:31 crc kubenswrapper[4789]: I1122 07:57:31.814925 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29396625-nrhcv" event={"ID":"ac3d1daa-cb6d-41fc-9798-e687d9b2a21d","Type":"ContainerDied","Data":"12999710c702d658eae70903ce4222dc88be544709994e1063ed779ff6d938e1"} Nov 22 07:57:31 crc kubenswrapper[4789]: I1122 07:57:31.814983 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="12999710c702d658eae70903ce4222dc88be544709994e1063ed779ff6d938e1" Nov 22 07:57:31 crc kubenswrapper[4789]: I1122 07:57:31.814958 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396625-nrhcv" Nov 22 07:57:31 crc kubenswrapper[4789]: I1122 07:57:31.820072 4789 generic.go:334] "Generic (PLEG): container finished" podID="d4a4f530-1afc-402b-8026-584bba4f8f1a" containerID="13fa65cbf77f7199881e2a4fa541ef485009f272481faf8e2f0f5f882f228268" exitCode=0 Nov 22 07:57:31 crc kubenswrapper[4789]: I1122 07:57:31.821107 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gbjc4" event={"ID":"d4a4f530-1afc-402b-8026-584bba4f8f1a","Type":"ContainerDied","Data":"13fa65cbf77f7199881e2a4fa541ef485009f272481faf8e2f0f5f882f228268"} Nov 22 07:57:31 crc kubenswrapper[4789]: I1122 07:57:31.826415 4789 generic.go:334] "Generic (PLEG): container finished" podID="6a078eea-384c-41cc-ae2e-30e6696422a2" containerID="85b63870e53a2e3ec8be518bf3b3f93585cac0bf20b5e9cb6305b0fed81c9a69" exitCode=0 Nov 22 07:57:31 crc kubenswrapper[4789]: I1122 07:57:31.826825 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rnkxn" event={"ID":"6a078eea-384c-41cc-ae2e-30e6696422a2","Type":"ContainerDied","Data":"85b63870e53a2e3ec8be518bf3b3f93585cac0bf20b5e9cb6305b0fed81c9a69"} Nov 22 07:57:31 crc kubenswrapper[4789]: I1122 07:57:31.826867 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rnkxn" event={"ID":"6a078eea-384c-41cc-ae2e-30e6696422a2","Type":"ContainerStarted","Data":"8cfc825de34d8dc7d7bfd849784ebe1f9e7a88b1a7382c7bdd503d546ec40b41"} Nov 22 07:57:31 crc kubenswrapper[4789]: I1122 07:57:31.834673 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-frls8" Nov 22 07:57:31 crc kubenswrapper[4789]: I1122 07:57:31.842657 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8235ef4a-0670-4b55-bfe6-4cc3cb1b24de-catalog-content\") pod \"redhat-marketplace-5wgxf\" (UID: \"8235ef4a-0670-4b55-bfe6-4cc3cb1b24de\") " pod="openshift-marketplace/redhat-marketplace-5wgxf" Nov 22 07:57:31 crc kubenswrapper[4789]: I1122 07:57:31.842708 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8235ef4a-0670-4b55-bfe6-4cc3cb1b24de-utilities\") pod \"redhat-marketplace-5wgxf\" (UID: \"8235ef4a-0670-4b55-bfe6-4cc3cb1b24de\") " pod="openshift-marketplace/redhat-marketplace-5wgxf" Nov 22 07:57:31 crc kubenswrapper[4789]: I1122 07:57:31.842764 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vlv48\" (UniqueName: \"kubernetes.io/projected/8235ef4a-0670-4b55-bfe6-4cc3cb1b24de-kube-api-access-vlv48\") pod \"redhat-marketplace-5wgxf\" (UID: \"8235ef4a-0670-4b55-bfe6-4cc3cb1b24de\") " pod="openshift-marketplace/redhat-marketplace-5wgxf" Nov 22 07:57:31 crc kubenswrapper[4789]: I1122 07:57:31.842852 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:31 crc kubenswrapper[4789]: E1122 07:57:31.843656 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 07:57:32.343642284 +0000 UTC m=+66.578042557 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l4q7j" (UID: "b80556f6-5b32-486d-8459-f17bf90c4e24") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:31 crc kubenswrapper[4789]: I1122 07:57:31.929556 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-77svl" Nov 22 07:57:31 crc kubenswrapper[4789]: I1122 07:57:31.944252 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 07:57:31 crc kubenswrapper[4789]: I1122 07:57:31.944546 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vlv48\" (UniqueName: \"kubernetes.io/projected/8235ef4a-0670-4b55-bfe6-4cc3cb1b24de-kube-api-access-vlv48\") pod \"redhat-marketplace-5wgxf\" (UID: \"8235ef4a-0670-4b55-bfe6-4cc3cb1b24de\") " pod="openshift-marketplace/redhat-marketplace-5wgxf" Nov 22 07:57:31 crc kubenswrapper[4789]: I1122 07:57:31.944716 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8235ef4a-0670-4b55-bfe6-4cc3cb1b24de-catalog-content\") pod \"redhat-marketplace-5wgxf\" (UID: \"8235ef4a-0670-4b55-bfe6-4cc3cb1b24de\") " pod="openshift-marketplace/redhat-marketplace-5wgxf" Nov 22 07:57:31 crc kubenswrapper[4789]: I1122 07:57:31.944801 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8235ef4a-0670-4b55-bfe6-4cc3cb1b24de-utilities\") pod \"redhat-marketplace-5wgxf\" (UID: \"8235ef4a-0670-4b55-bfe6-4cc3cb1b24de\") " pod="openshift-marketplace/redhat-marketplace-5wgxf" Nov 22 07:57:31 crc kubenswrapper[4789]: E1122 07:57:31.945251 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 07:57:32.445232223 +0000 UTC m=+66.679632506 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:31 crc kubenswrapper[4789]: I1122 07:57:31.946325 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8235ef4a-0670-4b55-bfe6-4cc3cb1b24de-catalog-content\") pod \"redhat-marketplace-5wgxf\" (UID: \"8235ef4a-0670-4b55-bfe6-4cc3cb1b24de\") " pod="openshift-marketplace/redhat-marketplace-5wgxf" Nov 22 07:57:31 crc kubenswrapper[4789]: I1122 07:57:31.946862 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8235ef4a-0670-4b55-bfe6-4cc3cb1b24de-utilities\") pod \"redhat-marketplace-5wgxf\" (UID: \"8235ef4a-0670-4b55-bfe6-4cc3cb1b24de\") " pod="openshift-marketplace/redhat-marketplace-5wgxf" Nov 22 07:57:31 crc kubenswrapper[4789]: E1122 07:57:31.989555 4789 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="405ac72a4fd327b3c44cbc8cd4e15338078a9cfceba08604c5a9f00fd7e61592" cmd=["/bin/bash","-c","test -f /ready/ready"] Nov 22 07:57:31 crc kubenswrapper[4789]: E1122 07:57:31.993343 4789 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="405ac72a4fd327b3c44cbc8cd4e15338078a9cfceba08604c5a9f00fd7e61592" cmd=["/bin/bash","-c","test -f /ready/ready"] Nov 22 07:57:31 crc kubenswrapper[4789]: E1122 07:57:31.995302 4789 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="405ac72a4fd327b3c44cbc8cd4e15338078a9cfceba08604c5a9f00fd7e61592" cmd=["/bin/bash","-c","test -f /ready/ready"] Nov 22 07:57:31 crc kubenswrapper[4789]: E1122 07:57:31.995381 4789 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openshift-multus/cni-sysctl-allowlist-ds-gcnts" podUID="75fa6283-acd8-41bc-9803-2f55119f6828" containerName="kube-multus-additional-cni-plugins" Nov 22 07:57:32 crc kubenswrapper[4789]: I1122 07:57:32.002376 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vlv48\" (UniqueName: \"kubernetes.io/projected/8235ef4a-0670-4b55-bfe6-4cc3cb1b24de-kube-api-access-vlv48\") pod \"redhat-marketplace-5wgxf\" (UID: \"8235ef4a-0670-4b55-bfe6-4cc3cb1b24de\") " pod="openshift-marketplace/redhat-marketplace-5wgxf" Nov 22 07:57:32 crc kubenswrapper[4789]: I1122 07:57:32.015686 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5wgxf" Nov 22 07:57:32 crc kubenswrapper[4789]: I1122 07:57:32.047698 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:32 crc kubenswrapper[4789]: E1122 07:57:32.048115 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 07:57:32.548099287 +0000 UTC m=+66.782499560 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l4q7j" (UID: "b80556f6-5b32-486d-8459-f17bf90c4e24") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:32 crc kubenswrapper[4789]: I1122 07:57:32.096272 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-xxw9z"] Nov 22 07:57:32 crc kubenswrapper[4789]: I1122 07:57:32.097215 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xxw9z" Nov 22 07:57:32 crc kubenswrapper[4789]: I1122 07:57:32.106835 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xxw9z"] Nov 22 07:57:32 crc kubenswrapper[4789]: I1122 07:57:32.148861 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 07:57:32 crc kubenswrapper[4789]: E1122 07:57:32.148995 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 07:57:32.648972596 +0000 UTC m=+66.883372869 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:32 crc kubenswrapper[4789]: I1122 07:57:32.149639 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:32 crc kubenswrapper[4789]: E1122 07:57:32.149990 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 07:57:32.649978904 +0000 UTC m=+66.884379187 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l4q7j" (UID: "b80556f6-5b32-486d-8459-f17bf90c4e24") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:32 crc kubenswrapper[4789]: I1122 07:57:32.250976 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 07:57:32 crc kubenswrapper[4789]: I1122 07:57:32.251210 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xb8h9\" (UniqueName: \"kubernetes.io/projected/0d93195d-1cef-4b5e-bc10-c696e82d5e38-kube-api-access-xb8h9\") pod \"redhat-marketplace-xxw9z\" (UID: \"0d93195d-1cef-4b5e-bc10-c696e82d5e38\") " pod="openshift-marketplace/redhat-marketplace-xxw9z" Nov 22 07:57:32 crc kubenswrapper[4789]: I1122 07:57:32.251264 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0d93195d-1cef-4b5e-bc10-c696e82d5e38-catalog-content\") pod \"redhat-marketplace-xxw9z\" (UID: \"0d93195d-1cef-4b5e-bc10-c696e82d5e38\") " pod="openshift-marketplace/redhat-marketplace-xxw9z" Nov 22 07:57:32 crc kubenswrapper[4789]: I1122 07:57:32.251338 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0d93195d-1cef-4b5e-bc10-c696e82d5e38-utilities\") pod \"redhat-marketplace-xxw9z\" (UID: \"0d93195d-1cef-4b5e-bc10-c696e82d5e38\") " pod="openshift-marketplace/redhat-marketplace-xxw9z" Nov 22 07:57:32 crc kubenswrapper[4789]: E1122 07:57:32.251503 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 07:57:32.751453889 +0000 UTC m=+66.985854162 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:32 crc kubenswrapper[4789]: I1122 07:57:32.356590 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0d93195d-1cef-4b5e-bc10-c696e82d5e38-catalog-content\") pod \"redhat-marketplace-xxw9z\" (UID: \"0d93195d-1cef-4b5e-bc10-c696e82d5e38\") " pod="openshift-marketplace/redhat-marketplace-xxw9z" Nov 22 07:57:32 crc kubenswrapper[4789]: I1122 07:57:32.356669 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:32 crc kubenswrapper[4789]: I1122 07:57:32.356713 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0d93195d-1cef-4b5e-bc10-c696e82d5e38-utilities\") pod \"redhat-marketplace-xxw9z\" (UID: \"0d93195d-1cef-4b5e-bc10-c696e82d5e38\") " pod="openshift-marketplace/redhat-marketplace-xxw9z" Nov 22 07:57:32 crc kubenswrapper[4789]: I1122 07:57:32.356785 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xb8h9\" (UniqueName: \"kubernetes.io/projected/0d93195d-1cef-4b5e-bc10-c696e82d5e38-kube-api-access-xb8h9\") pod \"redhat-marketplace-xxw9z\" (UID: \"0d93195d-1cef-4b5e-bc10-c696e82d5e38\") " pod="openshift-marketplace/redhat-marketplace-xxw9z" Nov 22 07:57:32 crc kubenswrapper[4789]: E1122 07:57:32.357217 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 07:57:32.857198312 +0000 UTC m=+67.091598605 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l4q7j" (UID: "b80556f6-5b32-486d-8459-f17bf90c4e24") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:32 crc kubenswrapper[4789]: I1122 07:57:32.357885 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0d93195d-1cef-4b5e-bc10-c696e82d5e38-catalog-content\") pod \"redhat-marketplace-xxw9z\" (UID: \"0d93195d-1cef-4b5e-bc10-c696e82d5e38\") " pod="openshift-marketplace/redhat-marketplace-xxw9z" Nov 22 07:57:32 crc kubenswrapper[4789]: I1122 07:57:32.357891 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0d93195d-1cef-4b5e-bc10-c696e82d5e38-utilities\") pod \"redhat-marketplace-xxw9z\" (UID: \"0d93195d-1cef-4b5e-bc10-c696e82d5e38\") " pod="openshift-marketplace/redhat-marketplace-xxw9z" Nov 22 07:57:32 crc kubenswrapper[4789]: I1122 07:57:32.359139 4789 patch_prober.go:28] interesting pod/router-default-5444994796-x5b9n container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 22 07:57:32 crc kubenswrapper[4789]: [-]has-synced failed: reason withheld Nov 22 07:57:32 crc kubenswrapper[4789]: [+]process-running ok Nov 22 07:57:32 crc kubenswrapper[4789]: healthz check failed Nov 22 07:57:32 crc kubenswrapper[4789]: I1122 07:57:32.359184 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-x5b9n" podUID="352693a0-531b-4017-9582-5424c4691795" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 22 07:57:32 crc kubenswrapper[4789]: I1122 07:57:32.389473 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xb8h9\" (UniqueName: \"kubernetes.io/projected/0d93195d-1cef-4b5e-bc10-c696e82d5e38-kube-api-access-xb8h9\") pod \"redhat-marketplace-xxw9z\" (UID: \"0d93195d-1cef-4b5e-bc10-c696e82d5e38\") " pod="openshift-marketplace/redhat-marketplace-xxw9z" Nov 22 07:57:32 crc kubenswrapper[4789]: I1122 07:57:32.425554 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xxw9z" Nov 22 07:57:32 crc kubenswrapper[4789]: I1122 07:57:32.443529 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5wgxf"] Nov 22 07:57:32 crc kubenswrapper[4789]: W1122 07:57:32.449088 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8235ef4a_0670_4b55_bfe6_4cc3cb1b24de.slice/crio-51126406b32d5dfb23c27551f52a44b5452197883de71ef092d207dad0aa8053 WatchSource:0}: Error finding container 51126406b32d5dfb23c27551f52a44b5452197883de71ef092d207dad0aa8053: Status 404 returned error can't find the container with id 51126406b32d5dfb23c27551f52a44b5452197883de71ef092d207dad0aa8053 Nov 22 07:57:32 crc kubenswrapper[4789]: I1122 07:57:32.457526 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 07:57:32 crc kubenswrapper[4789]: E1122 07:57:32.457635 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 07:57:32.957608838 +0000 UTC m=+67.192009111 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:32 crc kubenswrapper[4789]: I1122 07:57:32.458189 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:32 crc kubenswrapper[4789]: E1122 07:57:32.458783 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 07:57:32.958741259 +0000 UTC m=+67.193141542 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l4q7j" (UID: "b80556f6-5b32-486d-8459-f17bf90c4e24") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:32 crc kubenswrapper[4789]: I1122 07:57:32.559703 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 07:57:32 crc kubenswrapper[4789]: E1122 07:57:32.560117 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 07:57:33.060086661 +0000 UTC m=+67.294486934 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:32 crc kubenswrapper[4789]: I1122 07:57:32.560324 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:32 crc kubenswrapper[4789]: E1122 07:57:32.560652 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 07:57:33.060638356 +0000 UTC m=+67.295038629 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l4q7j" (UID: "b80556f6-5b32-486d-8459-f17bf90c4e24") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:32 crc kubenswrapper[4789]: I1122 07:57:32.630471 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xxw9z"] Nov 22 07:57:32 crc kubenswrapper[4789]: I1122 07:57:32.661604 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 07:57:32 crc kubenswrapper[4789]: E1122 07:57:32.662394 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 07:57:33.162376799 +0000 UTC m=+67.396777072 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:32 crc kubenswrapper[4789]: I1122 07:57:32.764290 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:32 crc kubenswrapper[4789]: E1122 07:57:32.765150 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 07:57:33.265131 +0000 UTC m=+67.499531273 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l4q7j" (UID: "b80556f6-5b32-486d-8459-f17bf90c4e24") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:32 crc kubenswrapper[4789]: I1122 07:57:32.832336 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xxw9z" event={"ID":"0d93195d-1cef-4b5e-bc10-c696e82d5e38","Type":"ContainerStarted","Data":"d6fa1d063398f944333d2ff2272485c3a3cd2da792824f624d724404191a1754"} Nov 22 07:57:32 crc kubenswrapper[4789]: I1122 07:57:32.833593 4789 generic.go:334] "Generic (PLEG): container finished" podID="f42980d5-344a-4cf7-b937-431ea412e9ab" containerID="adcd9e6728b8bdf50a436b27fdf82c0918f7b37c9bd07bbac9eb2397c6646676" exitCode=0 Nov 22 07:57:32 crc kubenswrapper[4789]: I1122 07:57:32.833638 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"f42980d5-344a-4cf7-b937-431ea412e9ab","Type":"ContainerDied","Data":"adcd9e6728b8bdf50a436b27fdf82c0918f7b37c9bd07bbac9eb2397c6646676"} Nov 22 07:57:32 crc kubenswrapper[4789]: I1122 07:57:32.834886 4789 generic.go:334] "Generic (PLEG): container finished" podID="8235ef4a-0670-4b55-bfe6-4cc3cb1b24de" containerID="63766f140e768be82aa548a637ff889922d5555ed2edbee551dcc16cf013a7ba" exitCode=0 Nov 22 07:57:32 crc kubenswrapper[4789]: I1122 07:57:32.834929 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5wgxf" event={"ID":"8235ef4a-0670-4b55-bfe6-4cc3cb1b24de","Type":"ContainerDied","Data":"63766f140e768be82aa548a637ff889922d5555ed2edbee551dcc16cf013a7ba"} Nov 22 07:57:32 crc kubenswrapper[4789]: I1122 07:57:32.834943 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5wgxf" event={"ID":"8235ef4a-0670-4b55-bfe6-4cc3cb1b24de","Type":"ContainerStarted","Data":"51126406b32d5dfb23c27551f52a44b5452197883de71ef092d207dad0aa8053"} Nov 22 07:57:32 crc kubenswrapper[4789]: I1122 07:57:32.835918 4789 generic.go:334] "Generic (PLEG): container finished" podID="ef38f088-0b6a-49a8-811f-3973b8207308" containerID="909ffa9c112c40bf527b814f367d47a9af890485b374aefe90d7f37a1613c140" exitCode=0 Nov 22 07:57:32 crc kubenswrapper[4789]: I1122 07:57:32.835954 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fnbtg" event={"ID":"ef38f088-0b6a-49a8-811f-3973b8207308","Type":"ContainerDied","Data":"909ffa9c112c40bf527b814f367d47a9af890485b374aefe90d7f37a1613c140"} Nov 22 07:57:32 crc kubenswrapper[4789]: I1122 07:57:32.868589 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 07:57:32 crc kubenswrapper[4789]: E1122 07:57:32.868933 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 07:57:33.368919058 +0000 UTC m=+67.603319331 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:32 crc kubenswrapper[4789]: I1122 07:57:32.891881 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-jdz2m"] Nov 22 07:57:32 crc kubenswrapper[4789]: I1122 07:57:32.892845 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jdz2m" Nov 22 07:57:32 crc kubenswrapper[4789]: I1122 07:57:32.895395 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 22 07:57:32 crc kubenswrapper[4789]: I1122 07:57:32.903031 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jdz2m"] Nov 22 07:57:32 crc kubenswrapper[4789]: I1122 07:57:32.971669 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:32 crc kubenswrapper[4789]: I1122 07:57:32.971719 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2bbcd7cf-ccb1-4b91-9e71-2aa87677be19-utilities\") pod \"redhat-operators-jdz2m\" (UID: \"2bbcd7cf-ccb1-4b91-9e71-2aa87677be19\") " pod="openshift-marketplace/redhat-operators-jdz2m" Nov 22 07:57:32 crc kubenswrapper[4789]: I1122 07:57:32.971767 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vs8fh\" (UniqueName: \"kubernetes.io/projected/2bbcd7cf-ccb1-4b91-9e71-2aa87677be19-kube-api-access-vs8fh\") pod \"redhat-operators-jdz2m\" (UID: \"2bbcd7cf-ccb1-4b91-9e71-2aa87677be19\") " pod="openshift-marketplace/redhat-operators-jdz2m" Nov 22 07:57:32 crc kubenswrapper[4789]: I1122 07:57:32.971857 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2bbcd7cf-ccb1-4b91-9e71-2aa87677be19-catalog-content\") pod \"redhat-operators-jdz2m\" (UID: \"2bbcd7cf-ccb1-4b91-9e71-2aa87677be19\") " pod="openshift-marketplace/redhat-operators-jdz2m" Nov 22 07:57:32 crc kubenswrapper[4789]: E1122 07:57:32.972275 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 07:57:33.472256555 +0000 UTC m=+67.706656868 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l4q7j" (UID: "b80556f6-5b32-486d-8459-f17bf90c4e24") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:33 crc kubenswrapper[4789]: I1122 07:57:33.073002 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 07:57:33 crc kubenswrapper[4789]: E1122 07:57:33.073049 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 07:57:33.573033912 +0000 UTC m=+67.807434185 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:33 crc kubenswrapper[4789]: I1122 07:57:33.073381 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:33 crc kubenswrapper[4789]: I1122 07:57:33.073421 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2bbcd7cf-ccb1-4b91-9e71-2aa87677be19-utilities\") pod \"redhat-operators-jdz2m\" (UID: \"2bbcd7cf-ccb1-4b91-9e71-2aa87677be19\") " pod="openshift-marketplace/redhat-operators-jdz2m" Nov 22 07:57:33 crc kubenswrapper[4789]: I1122 07:57:33.073453 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vs8fh\" (UniqueName: \"kubernetes.io/projected/2bbcd7cf-ccb1-4b91-9e71-2aa87677be19-kube-api-access-vs8fh\") pod \"redhat-operators-jdz2m\" (UID: \"2bbcd7cf-ccb1-4b91-9e71-2aa87677be19\") " pod="openshift-marketplace/redhat-operators-jdz2m" Nov 22 07:57:33 crc kubenswrapper[4789]: I1122 07:57:33.073516 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2bbcd7cf-ccb1-4b91-9e71-2aa87677be19-catalog-content\") pod \"redhat-operators-jdz2m\" (UID: \"2bbcd7cf-ccb1-4b91-9e71-2aa87677be19\") " pod="openshift-marketplace/redhat-operators-jdz2m" Nov 22 07:57:33 crc kubenswrapper[4789]: I1122 07:57:33.073966 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2bbcd7cf-ccb1-4b91-9e71-2aa87677be19-catalog-content\") pod \"redhat-operators-jdz2m\" (UID: \"2bbcd7cf-ccb1-4b91-9e71-2aa87677be19\") " pod="openshift-marketplace/redhat-operators-jdz2m" Nov 22 07:57:33 crc kubenswrapper[4789]: I1122 07:57:33.074006 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2bbcd7cf-ccb1-4b91-9e71-2aa87677be19-utilities\") pod \"redhat-operators-jdz2m\" (UID: \"2bbcd7cf-ccb1-4b91-9e71-2aa87677be19\") " pod="openshift-marketplace/redhat-operators-jdz2m" Nov 22 07:57:33 crc kubenswrapper[4789]: E1122 07:57:33.074231 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 07:57:33.574221504 +0000 UTC m=+67.808621777 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l4q7j" (UID: "b80556f6-5b32-486d-8459-f17bf90c4e24") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:33 crc kubenswrapper[4789]: I1122 07:57:33.098286 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vs8fh\" (UniqueName: \"kubernetes.io/projected/2bbcd7cf-ccb1-4b91-9e71-2aa87677be19-kube-api-access-vs8fh\") pod \"redhat-operators-jdz2m\" (UID: \"2bbcd7cf-ccb1-4b91-9e71-2aa87677be19\") " pod="openshift-marketplace/redhat-operators-jdz2m" Nov 22 07:57:33 crc kubenswrapper[4789]: I1122 07:57:33.110488 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 22 07:57:33 crc kubenswrapper[4789]: I1122 07:57:33.174523 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 07:57:33 crc kubenswrapper[4789]: E1122 07:57:33.174726 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 07:57:33.674695132 +0000 UTC m=+67.909095415 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:33 crc kubenswrapper[4789]: I1122 07:57:33.174843 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:33 crc kubenswrapper[4789]: E1122 07:57:33.175141 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 07:57:33.675129405 +0000 UTC m=+67.909529678 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l4q7j" (UID: "b80556f6-5b32-486d-8459-f17bf90c4e24") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:33 crc kubenswrapper[4789]: I1122 07:57:33.249147 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jdz2m" Nov 22 07:57:33 crc kubenswrapper[4789]: I1122 07:57:33.276125 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 07:57:33 crc kubenswrapper[4789]: I1122 07:57:33.276222 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3d44f23c-b509-470a-a5e4-6ef336d32d5a-kube-api-access\") pod \"3d44f23c-b509-470a-a5e4-6ef336d32d5a\" (UID: \"3d44f23c-b509-470a-a5e4-6ef336d32d5a\") " Nov 22 07:57:33 crc kubenswrapper[4789]: I1122 07:57:33.276265 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3d44f23c-b509-470a-a5e4-6ef336d32d5a-kubelet-dir\") pod \"3d44f23c-b509-470a-a5e4-6ef336d32d5a\" (UID: \"3d44f23c-b509-470a-a5e4-6ef336d32d5a\") " Nov 22 07:57:33 crc kubenswrapper[4789]: E1122 07:57:33.276299 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 07:57:33.776269211 +0000 UTC m=+68.010669524 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:33 crc kubenswrapper[4789]: I1122 07:57:33.276407 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3d44f23c-b509-470a-a5e4-6ef336d32d5a-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "3d44f23c-b509-470a-a5e4-6ef336d32d5a" (UID: "3d44f23c-b509-470a-a5e4-6ef336d32d5a"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 07:57:33 crc kubenswrapper[4789]: I1122 07:57:33.276426 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:33 crc kubenswrapper[4789]: I1122 07:57:33.276562 4789 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3d44f23c-b509-470a-a5e4-6ef336d32d5a-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 22 07:57:33 crc kubenswrapper[4789]: E1122 07:57:33.277284 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 07:57:33.777267399 +0000 UTC m=+68.011667742 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l4q7j" (UID: "b80556f6-5b32-486d-8459-f17bf90c4e24") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:33 crc kubenswrapper[4789]: I1122 07:57:33.294916 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d44f23c-b509-470a-a5e4-6ef336d32d5a-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "3d44f23c-b509-470a-a5e4-6ef336d32d5a" (UID: "3d44f23c-b509-470a-a5e4-6ef336d32d5a"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 07:57:33 crc kubenswrapper[4789]: I1122 07:57:33.298404 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-4gz55"] Nov 22 07:57:33 crc kubenswrapper[4789]: E1122 07:57:33.298945 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d44f23c-b509-470a-a5e4-6ef336d32d5a" containerName="pruner" Nov 22 07:57:33 crc kubenswrapper[4789]: I1122 07:57:33.299047 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d44f23c-b509-470a-a5e4-6ef336d32d5a" containerName="pruner" Nov 22 07:57:33 crc kubenswrapper[4789]: I1122 07:57:33.299203 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d44f23c-b509-470a-a5e4-6ef336d32d5a" containerName="pruner" Nov 22 07:57:33 crc kubenswrapper[4789]: I1122 07:57:33.300645 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4gz55" Nov 22 07:57:33 crc kubenswrapper[4789]: I1122 07:57:33.306190 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4gz55"] Nov 22 07:57:33 crc kubenswrapper[4789]: I1122 07:57:33.351382 4789 patch_prober.go:28] interesting pod/router-default-5444994796-x5b9n container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 22 07:57:33 crc kubenswrapper[4789]: [-]has-synced failed: reason withheld Nov 22 07:57:33 crc kubenswrapper[4789]: [+]process-running ok Nov 22 07:57:33 crc kubenswrapper[4789]: healthz check failed Nov 22 07:57:33 crc kubenswrapper[4789]: I1122 07:57:33.351436 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-x5b9n" podUID="352693a0-531b-4017-9582-5424c4691795" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 22 07:57:33 crc kubenswrapper[4789]: I1122 07:57:33.377717 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 07:57:33 crc kubenswrapper[4789]: E1122 07:57:33.377892 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 07:57:33.87786304 +0000 UTC m=+68.112263313 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:33 crc kubenswrapper[4789]: I1122 07:57:33.377994 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92287f13-f1d5-422a-9107-f0495287e5ff-catalog-content\") pod \"redhat-operators-4gz55\" (UID: \"92287f13-f1d5-422a-9107-f0495287e5ff\") " pod="openshift-marketplace/redhat-operators-4gz55" Nov 22 07:57:33 crc kubenswrapper[4789]: I1122 07:57:33.378041 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92287f13-f1d5-422a-9107-f0495287e5ff-utilities\") pod \"redhat-operators-4gz55\" (UID: \"92287f13-f1d5-422a-9107-f0495287e5ff\") " pod="openshift-marketplace/redhat-operators-4gz55" Nov 22 07:57:33 crc kubenswrapper[4789]: I1122 07:57:33.378170 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lnhh8\" (UniqueName: \"kubernetes.io/projected/92287f13-f1d5-422a-9107-f0495287e5ff-kube-api-access-lnhh8\") pod \"redhat-operators-4gz55\" (UID: \"92287f13-f1d5-422a-9107-f0495287e5ff\") " pod="openshift-marketplace/redhat-operators-4gz55" Nov 22 07:57:33 crc kubenswrapper[4789]: I1122 07:57:33.378234 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:33 crc kubenswrapper[4789]: I1122 07:57:33.378282 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3d44f23c-b509-470a-a5e4-6ef336d32d5a-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 22 07:57:33 crc kubenswrapper[4789]: E1122 07:57:33.378693 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 07:57:33.878675872 +0000 UTC m=+68.113076145 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l4q7j" (UID: "b80556f6-5b32-486d-8459-f17bf90c4e24") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:33 crc kubenswrapper[4789]: I1122 07:57:33.464123 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jdz2m"] Nov 22 07:57:33 crc kubenswrapper[4789]: I1122 07:57:33.478815 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 07:57:33 crc kubenswrapper[4789]: I1122 07:57:33.478947 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92287f13-f1d5-422a-9107-f0495287e5ff-catalog-content\") pod \"redhat-operators-4gz55\" (UID: \"92287f13-f1d5-422a-9107-f0495287e5ff\") " pod="openshift-marketplace/redhat-operators-4gz55" Nov 22 07:57:33 crc kubenswrapper[4789]: I1122 07:57:33.478973 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92287f13-f1d5-422a-9107-f0495287e5ff-utilities\") pod \"redhat-operators-4gz55\" (UID: \"92287f13-f1d5-422a-9107-f0495287e5ff\") " pod="openshift-marketplace/redhat-operators-4gz55" Nov 22 07:57:33 crc kubenswrapper[4789]: I1122 07:57:33.479022 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lnhh8\" (UniqueName: \"kubernetes.io/projected/92287f13-f1d5-422a-9107-f0495287e5ff-kube-api-access-lnhh8\") pod \"redhat-operators-4gz55\" (UID: \"92287f13-f1d5-422a-9107-f0495287e5ff\") " pod="openshift-marketplace/redhat-operators-4gz55" Nov 22 07:57:33 crc kubenswrapper[4789]: E1122 07:57:33.479486 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 07:57:33.979432249 +0000 UTC m=+68.213832552 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:33 crc kubenswrapper[4789]: I1122 07:57:33.480048 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92287f13-f1d5-422a-9107-f0495287e5ff-catalog-content\") pod \"redhat-operators-4gz55\" (UID: \"92287f13-f1d5-422a-9107-f0495287e5ff\") " pod="openshift-marketplace/redhat-operators-4gz55" Nov 22 07:57:33 crc kubenswrapper[4789]: I1122 07:57:33.480438 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92287f13-f1d5-422a-9107-f0495287e5ff-utilities\") pod \"redhat-operators-4gz55\" (UID: \"92287f13-f1d5-422a-9107-f0495287e5ff\") " pod="openshift-marketplace/redhat-operators-4gz55" Nov 22 07:57:33 crc kubenswrapper[4789]: I1122 07:57:33.505006 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lnhh8\" (UniqueName: \"kubernetes.io/projected/92287f13-f1d5-422a-9107-f0495287e5ff-kube-api-access-lnhh8\") pod \"redhat-operators-4gz55\" (UID: \"92287f13-f1d5-422a-9107-f0495287e5ff\") " pod="openshift-marketplace/redhat-operators-4gz55" Nov 22 07:57:33 crc kubenswrapper[4789]: I1122 07:57:33.580372 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:33 crc kubenswrapper[4789]: E1122 07:57:33.580885 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 07:57:34.080864983 +0000 UTC m=+68.315265266 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l4q7j" (UID: "b80556f6-5b32-486d-8459-f17bf90c4e24") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:33 crc kubenswrapper[4789]: I1122 07:57:33.622505 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4gz55" Nov 22 07:57:33 crc kubenswrapper[4789]: I1122 07:57:33.681447 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 07:57:33 crc kubenswrapper[4789]: E1122 07:57:33.681610 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 07:57:34.181585028 +0000 UTC m=+68.415985301 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:33 crc kubenswrapper[4789]: I1122 07:57:33.681683 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:33 crc kubenswrapper[4789]: E1122 07:57:33.682146 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 07:57:34.182133584 +0000 UTC m=+68.416533857 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l4q7j" (UID: "b80556f6-5b32-486d-8459-f17bf90c4e24") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:33 crc kubenswrapper[4789]: I1122 07:57:33.782869 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 07:57:33 crc kubenswrapper[4789]: E1122 07:57:33.782997 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 07:57:34.282973142 +0000 UTC m=+68.517373415 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:33 crc kubenswrapper[4789]: I1122 07:57:33.783819 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:33 crc kubenswrapper[4789]: E1122 07:57:33.784111 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 07:57:34.284100013 +0000 UTC m=+68.518500296 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l4q7j" (UID: "b80556f6-5b32-486d-8459-f17bf90c4e24") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:33 crc kubenswrapper[4789]: I1122 07:57:33.857148 4789 generic.go:334] "Generic (PLEG): container finished" podID="2bbcd7cf-ccb1-4b91-9e71-2aa87677be19" containerID="210d915204b25ed0acd603a467395328ce394aa17e753337ad9d6f5889e0bcde" exitCode=0 Nov 22 07:57:33 crc kubenswrapper[4789]: I1122 07:57:33.857232 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jdz2m" event={"ID":"2bbcd7cf-ccb1-4b91-9e71-2aa87677be19","Type":"ContainerDied","Data":"210d915204b25ed0acd603a467395328ce394aa17e753337ad9d6f5889e0bcde"} Nov 22 07:57:33 crc kubenswrapper[4789]: I1122 07:57:33.857267 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jdz2m" event={"ID":"2bbcd7cf-ccb1-4b91-9e71-2aa87677be19","Type":"ContainerStarted","Data":"2f5b97af05a9223958e3057a9f1b53f9d711571f91f0d1669d11c99f24c9e9d5"} Nov 22 07:57:33 crc kubenswrapper[4789]: I1122 07:57:33.862812 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-n27w4" event={"ID":"d58d000b-3f2c-44cb-90c5-8023e9f4f5db","Type":"ContainerStarted","Data":"019f9f618e51c0aa4cb640a3b910f7c490bf944960fa202fe4fd3b9ca2377f8d"} Nov 22 07:57:33 crc kubenswrapper[4789]: I1122 07:57:33.862857 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-n27w4" event={"ID":"d58d000b-3f2c-44cb-90c5-8023e9f4f5db","Type":"ContainerStarted","Data":"17fd3452cc9d2156717fed96744f512884f2ca370d0d3a890c49a26cc278bcda"} Nov 22 07:57:33 crc kubenswrapper[4789]: I1122 07:57:33.863959 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4gz55"] Nov 22 07:57:33 crc kubenswrapper[4789]: I1122 07:57:33.865679 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"3d44f23c-b509-470a-a5e4-6ef336d32d5a","Type":"ContainerDied","Data":"66cfdbc5e1bd531eeecd4e8527243e6d443f7cf6bbc3af861c3912ba6e8cd86c"} Nov 22 07:57:33 crc kubenswrapper[4789]: I1122 07:57:33.865731 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="66cfdbc5e1bd531eeecd4e8527243e6d443f7cf6bbc3af861c3912ba6e8cd86c" Nov 22 07:57:33 crc kubenswrapper[4789]: I1122 07:57:33.865770 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 22 07:57:33 crc kubenswrapper[4789]: I1122 07:57:33.867459 4789 generic.go:334] "Generic (PLEG): container finished" podID="0d93195d-1cef-4b5e-bc10-c696e82d5e38" containerID="d1740965270fe0eb6a052fefdfa72dfede028461d06e159fef3321ae477efd04" exitCode=0 Nov 22 07:57:33 crc kubenswrapper[4789]: I1122 07:57:33.867559 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xxw9z" event={"ID":"0d93195d-1cef-4b5e-bc10-c696e82d5e38","Type":"ContainerDied","Data":"d1740965270fe0eb6a052fefdfa72dfede028461d06e159fef3321ae477efd04"} Nov 22 07:57:33 crc kubenswrapper[4789]: I1122 07:57:33.885461 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 07:57:33 crc kubenswrapper[4789]: E1122 07:57:33.885682 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 07:57:34.385661411 +0000 UTC m=+68.620061684 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:33 crc kubenswrapper[4789]: I1122 07:57:33.885813 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:33 crc kubenswrapper[4789]: E1122 07:57:33.886217 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 07:57:34.386207986 +0000 UTC m=+68.620608269 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l4q7j" (UID: "b80556f6-5b32-486d-8459-f17bf90c4e24") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:33 crc kubenswrapper[4789]: I1122 07:57:33.987426 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 07:57:34 crc kubenswrapper[4789]: E1122 07:57:33.989025 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 07:57:34.488996527 +0000 UTC m=+68.723396880 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:34 crc kubenswrapper[4789]: I1122 07:57:34.089848 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:34 crc kubenswrapper[4789]: E1122 07:57:34.090235 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 07:57:34.590219626 +0000 UTC m=+68.824619899 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l4q7j" (UID: "b80556f6-5b32-486d-8459-f17bf90c4e24") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:34 crc kubenswrapper[4789]: I1122 07:57:34.116861 4789 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Nov 22 07:57:34 crc kubenswrapper[4789]: I1122 07:57:34.161148 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 22 07:57:34 crc kubenswrapper[4789]: I1122 07:57:34.191524 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 07:57:34 crc kubenswrapper[4789]: E1122 07:57:34.191951 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 07:57:34.691936339 +0000 UTC m=+68.926336612 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:34 crc kubenswrapper[4789]: I1122 07:57:34.293243 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f42980d5-344a-4cf7-b937-431ea412e9ab-kube-api-access\") pod \"f42980d5-344a-4cf7-b937-431ea412e9ab\" (UID: \"f42980d5-344a-4cf7-b937-431ea412e9ab\") " Nov 22 07:57:34 crc kubenswrapper[4789]: I1122 07:57:34.293499 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f42980d5-344a-4cf7-b937-431ea412e9ab-kubelet-dir\") pod \"f42980d5-344a-4cf7-b937-431ea412e9ab\" (UID: \"f42980d5-344a-4cf7-b937-431ea412e9ab\") " Nov 22 07:57:34 crc kubenswrapper[4789]: I1122 07:57:34.293587 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f42980d5-344a-4cf7-b937-431ea412e9ab-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "f42980d5-344a-4cf7-b937-431ea412e9ab" (UID: "f42980d5-344a-4cf7-b937-431ea412e9ab"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 07:57:34 crc kubenswrapper[4789]: I1122 07:57:34.293638 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:34 crc kubenswrapper[4789]: I1122 07:57:34.293812 4789 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f42980d5-344a-4cf7-b937-431ea412e9ab-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 22 07:57:34 crc kubenswrapper[4789]: E1122 07:57:34.294218 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 07:57:34.794201226 +0000 UTC m=+69.028601589 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l4q7j" (UID: "b80556f6-5b32-486d-8459-f17bf90c4e24") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:34 crc kubenswrapper[4789]: I1122 07:57:34.305679 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f42980d5-344a-4cf7-b937-431ea412e9ab-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "f42980d5-344a-4cf7-b937-431ea412e9ab" (UID: "f42980d5-344a-4cf7-b937-431ea412e9ab"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 07:57:34 crc kubenswrapper[4789]: I1122 07:57:34.353385 4789 patch_prober.go:28] interesting pod/router-default-5444994796-x5b9n container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 22 07:57:34 crc kubenswrapper[4789]: [-]has-synced failed: reason withheld Nov 22 07:57:34 crc kubenswrapper[4789]: [+]process-running ok Nov 22 07:57:34 crc kubenswrapper[4789]: healthz check failed Nov 22 07:57:34 crc kubenswrapper[4789]: I1122 07:57:34.353455 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-x5b9n" podUID="352693a0-531b-4017-9582-5424c4691795" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 22 07:57:34 crc kubenswrapper[4789]: I1122 07:57:34.394576 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 07:57:34 crc kubenswrapper[4789]: E1122 07:57:34.395109 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 07:57:34.895076906 +0000 UTC m=+69.129477179 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:34 crc kubenswrapper[4789]: I1122 07:57:34.395168 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f42980d5-344a-4cf7-b937-431ea412e9ab-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 22 07:57:34 crc kubenswrapper[4789]: I1122 07:57:34.496360 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:34 crc kubenswrapper[4789]: E1122 07:57:34.496774 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 07:57:34.996744137 +0000 UTC m=+69.231144410 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l4q7j" (UID: "b80556f6-5b32-486d-8459-f17bf90c4e24") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:34 crc kubenswrapper[4789]: E1122 07:57:34.607428 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 07:57:35.107400662 +0000 UTC m=+69.341800935 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:34 crc kubenswrapper[4789]: I1122 07:57:34.612792 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 07:57:34 crc kubenswrapper[4789]: I1122 07:57:34.613332 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:34 crc kubenswrapper[4789]: E1122 07:57:34.613862 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 07:57:35.113823016 +0000 UTC m=+69.348223289 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l4q7j" (UID: "b80556f6-5b32-486d-8459-f17bf90c4e24") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:34 crc kubenswrapper[4789]: I1122 07:57:34.715982 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 07:57:34 crc kubenswrapper[4789]: E1122 07:57:34.716186 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 07:57:35.216155635 +0000 UTC m=+69.450555908 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:34 crc kubenswrapper[4789]: I1122 07:57:34.716273 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:34 crc kubenswrapper[4789]: E1122 07:57:34.716627 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 07:57:35.216612068 +0000 UTC m=+69.451012341 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l4q7j" (UID: "b80556f6-5b32-486d-8459-f17bf90c4e24") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:34 crc kubenswrapper[4789]: I1122 07:57:34.822183 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 07:57:34 crc kubenswrapper[4789]: E1122 07:57:34.822409 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 07:57:35.32237382 +0000 UTC m=+69.556774103 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:34 crc kubenswrapper[4789]: I1122 07:57:34.822647 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:34 crc kubenswrapper[4789]: E1122 07:57:34.823090 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 07:57:35.32307824 +0000 UTC m=+69.557478513 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-l4q7j" (UID: "b80556f6-5b32-486d-8459-f17bf90c4e24") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 07:57:34 crc kubenswrapper[4789]: I1122 07:57:34.877022 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 22 07:57:34 crc kubenswrapper[4789]: I1122 07:57:34.877330 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"f42980d5-344a-4cf7-b937-431ea412e9ab","Type":"ContainerDied","Data":"284e22ef4912638d79025dc74afc24f8ee70a4b2c98c4fb266bff9373664a1dc"} Nov 22 07:57:34 crc kubenswrapper[4789]: I1122 07:57:34.877373 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="284e22ef4912638d79025dc74afc24f8ee70a4b2c98c4fb266bff9373664a1dc" Nov 22 07:57:34 crc kubenswrapper[4789]: I1122 07:57:34.880186 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-n27w4" event={"ID":"d58d000b-3f2c-44cb-90c5-8023e9f4f5db","Type":"ContainerStarted","Data":"afe66ee14db16579d2cbfd8f6f1081523906ff7ce3e42d9a4e997a2503e67f7f"} Nov 22 07:57:34 crc kubenswrapper[4789]: I1122 07:57:34.885212 4789 generic.go:334] "Generic (PLEG): container finished" podID="92287f13-f1d5-422a-9107-f0495287e5ff" containerID="69ad89253c458146e9be3bc545098e967c7ac62b48b82435e9ca7ebab6e2e7d6" exitCode=0 Nov 22 07:57:34 crc kubenswrapper[4789]: I1122 07:57:34.885242 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4gz55" event={"ID":"92287f13-f1d5-422a-9107-f0495287e5ff","Type":"ContainerDied","Data":"69ad89253c458146e9be3bc545098e967c7ac62b48b82435e9ca7ebab6e2e7d6"} Nov 22 07:57:34 crc kubenswrapper[4789]: I1122 07:57:34.885255 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4gz55" event={"ID":"92287f13-f1d5-422a-9107-f0495287e5ff","Type":"ContainerStarted","Data":"9a48770e409dd0e1f2acf609d1dfa08cfa32d424f329edbbd6e56a4c3f89aa53"} Nov 22 07:57:34 crc kubenswrapper[4789]: I1122 07:57:34.901560 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-n27w4" podStartSLOduration=16.90154105 podStartE2EDuration="16.90154105s" podCreationTimestamp="2025-11-22 07:57:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 07:57:34.896518734 +0000 UTC m=+69.130919007" watchObservedRunningTime="2025-11-22 07:57:34.90154105 +0000 UTC m=+69.135941323" Nov 22 07:57:34 crc kubenswrapper[4789]: I1122 07:57:34.903640 4789 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-11-22T07:57:34.116901541Z","Handler":null,"Name":""} Nov 22 07:57:34 crc kubenswrapper[4789]: I1122 07:57:34.909886 4789 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Nov 22 07:57:34 crc kubenswrapper[4789]: I1122 07:57:34.909924 4789 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Nov 22 07:57:34 crc kubenswrapper[4789]: I1122 07:57:34.923702 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 07:57:34 crc kubenswrapper[4789]: I1122 07:57:34.977117 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Nov 22 07:57:35 crc kubenswrapper[4789]: I1122 07:57:35.053652 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 22 07:57:35 crc kubenswrapper[4789]: I1122 07:57:35.126452 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:35 crc kubenswrapper[4789]: I1122 07:57:35.159869 4789 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 22 07:57:35 crc kubenswrapper[4789]: I1122 07:57:35.159940 4789 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:35 crc kubenswrapper[4789]: I1122 07:57:35.202599 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-l4q7j\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:35 crc kubenswrapper[4789]: I1122 07:57:35.350282 4789 patch_prober.go:28] interesting pod/router-default-5444994796-x5b9n container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 22 07:57:35 crc kubenswrapper[4789]: [-]has-synced failed: reason withheld Nov 22 07:57:35 crc kubenswrapper[4789]: [+]process-running ok Nov 22 07:57:35 crc kubenswrapper[4789]: healthz check failed Nov 22 07:57:35 crc kubenswrapper[4789]: I1122 07:57:35.350365 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-x5b9n" podUID="352693a0-531b-4017-9582-5424c4691795" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 22 07:57:35 crc kubenswrapper[4789]: I1122 07:57:35.437586 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Nov 22 07:57:35 crc kubenswrapper[4789]: I1122 07:57:35.444363 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:35 crc kubenswrapper[4789]: I1122 07:57:35.790766 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-l4q7j"] Nov 22 07:57:35 crc kubenswrapper[4789]: I1122 07:57:35.896080 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" event={"ID":"b80556f6-5b32-486d-8459-f17bf90c4e24","Type":"ContainerStarted","Data":"4f6b1ab90738b52b1c9f0ee01449e02798904a435fa37930c555980928ef50ee"} Nov 22 07:57:35 crc kubenswrapper[4789]: I1122 07:57:35.991094 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Nov 22 07:57:36 crc kubenswrapper[4789]: I1122 07:57:36.351646 4789 patch_prober.go:28] interesting pod/router-default-5444994796-x5b9n container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 22 07:57:36 crc kubenswrapper[4789]: [-]has-synced failed: reason withheld Nov 22 07:57:36 crc kubenswrapper[4789]: [+]process-running ok Nov 22 07:57:36 crc kubenswrapper[4789]: healthz check failed Nov 22 07:57:36 crc kubenswrapper[4789]: I1122 07:57:36.351728 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-x5b9n" podUID="352693a0-531b-4017-9582-5424c4691795" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 22 07:57:36 crc kubenswrapper[4789]: I1122 07:57:36.473116 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-vh974" Nov 22 07:57:36 crc kubenswrapper[4789]: I1122 07:57:36.479027 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-vh974" Nov 22 07:57:36 crc kubenswrapper[4789]: I1122 07:57:36.519337 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=2.519308285 podStartE2EDuration="2.519308285s" podCreationTimestamp="2025-11-22 07:57:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 07:57:36.509808917 +0000 UTC m=+70.744209210" watchObservedRunningTime="2025-11-22 07:57:36.519308285 +0000 UTC m=+70.753708558" Nov 22 07:57:36 crc kubenswrapper[4789]: I1122 07:57:36.903858 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" event={"ID":"b80556f6-5b32-486d-8459-f17bf90c4e24","Type":"ContainerStarted","Data":"4035555c9cf0e724ce2aa131cf4571b2740d4a9f623cc09f130d9eef32829946"} Nov 22 07:57:37 crc kubenswrapper[4789]: I1122 07:57:37.350845 4789 patch_prober.go:28] interesting pod/router-default-5444994796-x5b9n container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 22 07:57:37 crc kubenswrapper[4789]: [-]has-synced failed: reason withheld Nov 22 07:57:37 crc kubenswrapper[4789]: [+]process-running ok Nov 22 07:57:37 crc kubenswrapper[4789]: healthz check failed Nov 22 07:57:37 crc kubenswrapper[4789]: I1122 07:57:37.351057 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-x5b9n" podUID="352693a0-531b-4017-9582-5424c4691795" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 22 07:57:37 crc kubenswrapper[4789]: I1122 07:57:37.362143 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-szb9d" Nov 22 07:57:37 crc kubenswrapper[4789]: I1122 07:57:37.910820 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:37 crc kubenswrapper[4789]: I1122 07:57:37.934927 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" podStartSLOduration=46.93490471 podStartE2EDuration="46.93490471s" podCreationTimestamp="2025-11-22 07:56:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 07:57:37.926798819 +0000 UTC m=+72.161199112" watchObservedRunningTime="2025-11-22 07:57:37.93490471 +0000 UTC m=+72.169304983" Nov 22 07:57:38 crc kubenswrapper[4789]: I1122 07:57:38.358395 4789 patch_prober.go:28] interesting pod/router-default-5444994796-x5b9n container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 22 07:57:38 crc kubenswrapper[4789]: [-]has-synced failed: reason withheld Nov 22 07:57:38 crc kubenswrapper[4789]: [+]process-running ok Nov 22 07:57:38 crc kubenswrapper[4789]: healthz check failed Nov 22 07:57:38 crc kubenswrapper[4789]: I1122 07:57:38.358449 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-x5b9n" podUID="352693a0-531b-4017-9582-5424c4691795" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 22 07:57:39 crc kubenswrapper[4789]: I1122 07:57:39.351891 4789 patch_prober.go:28] interesting pod/router-default-5444994796-x5b9n container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 22 07:57:39 crc kubenswrapper[4789]: [-]has-synced failed: reason withheld Nov 22 07:57:39 crc kubenswrapper[4789]: [+]process-running ok Nov 22 07:57:39 crc kubenswrapper[4789]: healthz check failed Nov 22 07:57:39 crc kubenswrapper[4789]: I1122 07:57:39.351978 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-x5b9n" podUID="352693a0-531b-4017-9582-5424c4691795" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 22 07:57:40 crc kubenswrapper[4789]: I1122 07:57:40.366665 4789 patch_prober.go:28] interesting pod/router-default-5444994796-x5b9n container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 22 07:57:40 crc kubenswrapper[4789]: [-]has-synced failed: reason withheld Nov 22 07:57:40 crc kubenswrapper[4789]: [+]process-running ok Nov 22 07:57:40 crc kubenswrapper[4789]: healthz check failed Nov 22 07:57:40 crc kubenswrapper[4789]: I1122 07:57:40.367080 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-x5b9n" podUID="352693a0-531b-4017-9582-5424c4691795" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 22 07:57:41 crc kubenswrapper[4789]: I1122 07:57:41.018192 4789 patch_prober.go:28] interesting pod/downloads-7954f5f757-tdpht container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Nov 22 07:57:41 crc kubenswrapper[4789]: I1122 07:57:41.018200 4789 patch_prober.go:28] interesting pod/downloads-7954f5f757-tdpht container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Nov 22 07:57:41 crc kubenswrapper[4789]: I1122 07:57:41.018262 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-tdpht" podUID="83ac1d78-cc4c-48b7-b975-2f88802e2d17" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Nov 22 07:57:41 crc kubenswrapper[4789]: I1122 07:57:41.018311 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-tdpht" podUID="83ac1d78-cc4c-48b7-b975-2f88802e2d17" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Nov 22 07:57:41 crc kubenswrapper[4789]: I1122 07:57:41.072652 4789 patch_prober.go:28] interesting pod/console-f9d7485db-pb98x container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.19:8443/health\": dial tcp 10.217.0.19:8443: connect: connection refused" start-of-body= Nov 22 07:57:41 crc kubenswrapper[4789]: I1122 07:57:41.072826 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-pb98x" podUID="9b861750-2517-4cf3-9c78-930480711af0" containerName="console" probeResult="failure" output="Get \"https://10.217.0.19:8443/health\": dial tcp 10.217.0.19:8443: connect: connection refused" Nov 22 07:57:41 crc kubenswrapper[4789]: I1122 07:57:41.349694 4789 patch_prober.go:28] interesting pod/router-default-5444994796-x5b9n container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 22 07:57:41 crc kubenswrapper[4789]: [-]has-synced failed: reason withheld Nov 22 07:57:41 crc kubenswrapper[4789]: [+]process-running ok Nov 22 07:57:41 crc kubenswrapper[4789]: healthz check failed Nov 22 07:57:41 crc kubenswrapper[4789]: I1122 07:57:41.349842 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-x5b9n" podUID="352693a0-531b-4017-9582-5424c4691795" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 22 07:57:41 crc kubenswrapper[4789]: E1122 07:57:41.987384 4789 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="405ac72a4fd327b3c44cbc8cd4e15338078a9cfceba08604c5a9f00fd7e61592" cmd=["/bin/bash","-c","test -f /ready/ready"] Nov 22 07:57:41 crc kubenswrapper[4789]: E1122 07:57:41.989717 4789 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="405ac72a4fd327b3c44cbc8cd4e15338078a9cfceba08604c5a9f00fd7e61592" cmd=["/bin/bash","-c","test -f /ready/ready"] Nov 22 07:57:41 crc kubenswrapper[4789]: E1122 07:57:41.990961 4789 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="405ac72a4fd327b3c44cbc8cd4e15338078a9cfceba08604c5a9f00fd7e61592" cmd=["/bin/bash","-c","test -f /ready/ready"] Nov 22 07:57:41 crc kubenswrapper[4789]: E1122 07:57:41.990996 4789 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openshift-multus/cni-sysctl-allowlist-ds-gcnts" podUID="75fa6283-acd8-41bc-9803-2f55119f6828" containerName="kube-multus-additional-cni-plugins" Nov 22 07:57:42 crc kubenswrapper[4789]: I1122 07:57:42.349797 4789 patch_prober.go:28] interesting pod/router-default-5444994796-x5b9n container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 22 07:57:42 crc kubenswrapper[4789]: [-]has-synced failed: reason withheld Nov 22 07:57:42 crc kubenswrapper[4789]: [+]process-running ok Nov 22 07:57:42 crc kubenswrapper[4789]: healthz check failed Nov 22 07:57:42 crc kubenswrapper[4789]: I1122 07:57:42.349865 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-x5b9n" podUID="352693a0-531b-4017-9582-5424c4691795" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 22 07:57:43 crc kubenswrapper[4789]: I1122 07:57:43.206591 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" Nov 22 07:57:43 crc kubenswrapper[4789]: I1122 07:57:43.351114 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-x5b9n" Nov 22 07:57:43 crc kubenswrapper[4789]: I1122 07:57:43.354115 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-x5b9n" Nov 22 07:57:51 crc kubenswrapper[4789]: I1122 07:57:51.018285 4789 patch_prober.go:28] interesting pod/downloads-7954f5f757-tdpht container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Nov 22 07:57:51 crc kubenswrapper[4789]: I1122 07:57:51.019992 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-tdpht" podUID="83ac1d78-cc4c-48b7-b975-2f88802e2d17" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Nov 22 07:57:51 crc kubenswrapper[4789]: I1122 07:57:51.018482 4789 patch_prober.go:28] interesting pod/downloads-7954f5f757-tdpht container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Nov 22 07:57:51 crc kubenswrapper[4789]: I1122 07:57:51.020188 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-tdpht" podUID="83ac1d78-cc4c-48b7-b975-2f88802e2d17" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Nov 22 07:57:51 crc kubenswrapper[4789]: I1122 07:57:51.020250 4789 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-console/downloads-7954f5f757-tdpht" Nov 22 07:57:51 crc kubenswrapper[4789]: I1122 07:57:51.020898 4789 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="download-server" containerStatusID={"Type":"cri-o","ID":"54620f1f3b277d33854c3b2d2582c694baad3604484132173d061c2656aa3c6e"} pod="openshift-console/downloads-7954f5f757-tdpht" containerMessage="Container download-server failed liveness probe, will be restarted" Nov 22 07:57:51 crc kubenswrapper[4789]: I1122 07:57:51.021010 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/downloads-7954f5f757-tdpht" podUID="83ac1d78-cc4c-48b7-b975-2f88802e2d17" containerName="download-server" containerID="cri-o://54620f1f3b277d33854c3b2d2582c694baad3604484132173d061c2656aa3c6e" gracePeriod=2 Nov 22 07:57:51 crc kubenswrapper[4789]: I1122 07:57:51.021063 4789 patch_prober.go:28] interesting pod/downloads-7954f5f757-tdpht container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Nov 22 07:57:51 crc kubenswrapper[4789]: I1122 07:57:51.024710 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-tdpht" podUID="83ac1d78-cc4c-48b7-b975-2f88802e2d17" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Nov 22 07:57:51 crc kubenswrapper[4789]: I1122 07:57:51.076595 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-pb98x" Nov 22 07:57:51 crc kubenswrapper[4789]: I1122 07:57:51.081642 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-pb98x" Nov 22 07:57:51 crc kubenswrapper[4789]: E1122 07:57:51.988789 4789 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="405ac72a4fd327b3c44cbc8cd4e15338078a9cfceba08604c5a9f00fd7e61592" cmd=["/bin/bash","-c","test -f /ready/ready"] Nov 22 07:57:51 crc kubenswrapper[4789]: E1122 07:57:51.990580 4789 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="405ac72a4fd327b3c44cbc8cd4e15338078a9cfceba08604c5a9f00fd7e61592" cmd=["/bin/bash","-c","test -f /ready/ready"] Nov 22 07:57:51 crc kubenswrapper[4789]: E1122 07:57:51.992494 4789 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="405ac72a4fd327b3c44cbc8cd4e15338078a9cfceba08604c5a9f00fd7e61592" cmd=["/bin/bash","-c","test -f /ready/ready"] Nov 22 07:57:51 crc kubenswrapper[4789]: E1122 07:57:51.992632 4789 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openshift-multus/cni-sysctl-allowlist-ds-gcnts" podUID="75fa6283-acd8-41bc-9803-2f55119f6828" containerName="kube-multus-additional-cni-plugins" Nov 22 07:57:52 crc kubenswrapper[4789]: I1122 07:57:52.389944 4789 patch_prober.go:28] interesting pod/router-default-5444994796-x5b9n container/router namespace/openshift-ingress: Readiness probe status=failure output="Get \"http://localhost:1936/healthz/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 22 07:57:52 crc kubenswrapper[4789]: I1122 07:57:52.390014 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-ingress/router-default-5444994796-x5b9n" podUID="352693a0-531b-4017-9582-5424c4691795" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 22 07:57:55 crc kubenswrapper[4789]: I1122 07:57:55.109026 4789 generic.go:334] "Generic (PLEG): container finished" podID="83ac1d78-cc4c-48b7-b975-2f88802e2d17" containerID="54620f1f3b277d33854c3b2d2582c694baad3604484132173d061c2656aa3c6e" exitCode=0 Nov 22 07:57:55 crc kubenswrapper[4789]: I1122 07:57:55.109144 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-tdpht" event={"ID":"83ac1d78-cc4c-48b7-b975-2f88802e2d17","Type":"ContainerDied","Data":"54620f1f3b277d33854c3b2d2582c694baad3604484132173d061c2656aa3c6e"} Nov 22 07:57:55 crc kubenswrapper[4789]: I1122 07:57:55.451053 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 07:57:59 crc kubenswrapper[4789]: I1122 07:57:59.132104 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_cni-sysctl-allowlist-ds-gcnts_75fa6283-acd8-41bc-9803-2f55119f6828/kube-multus-additional-cni-plugins/0.log" Nov 22 07:57:59 crc kubenswrapper[4789]: I1122 07:57:59.132477 4789 generic.go:334] "Generic (PLEG): container finished" podID="75fa6283-acd8-41bc-9803-2f55119f6828" containerID="405ac72a4fd327b3c44cbc8cd4e15338078a9cfceba08604c5a9f00fd7e61592" exitCode=137 Nov 22 07:57:59 crc kubenswrapper[4789]: I1122 07:57:59.132512 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/cni-sysctl-allowlist-ds-gcnts" event={"ID":"75fa6283-acd8-41bc-9803-2f55119f6828","Type":"ContainerDied","Data":"405ac72a4fd327b3c44cbc8cd4e15338078a9cfceba08604c5a9f00fd7e61592"} Nov 22 07:58:01 crc kubenswrapper[4789]: I1122 07:58:01.018861 4789 patch_prober.go:28] interesting pod/downloads-7954f5f757-tdpht container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Nov 22 07:58:01 crc kubenswrapper[4789]: I1122 07:58:01.019351 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-tdpht" podUID="83ac1d78-cc4c-48b7-b975-2f88802e2d17" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Nov 22 07:58:01 crc kubenswrapper[4789]: I1122 07:58:01.657037 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-sszl6" Nov 22 07:58:01 crc kubenswrapper[4789]: E1122 07:58:01.947185 4789 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Nov 22 07:58:01 crc kubenswrapper[4789]: E1122 07:58:01.947391 4789 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-pf4qs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-gbjc4_openshift-marketplace(d4a4f530-1afc-402b-8026-584bba4f8f1a): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 22 07:58:01 crc kubenswrapper[4789]: E1122 07:58:01.948676 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-gbjc4" podUID="d4a4f530-1afc-402b-8026-584bba4f8f1a" Nov 22 07:58:01 crc kubenswrapper[4789]: E1122 07:58:01.985690 4789 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 405ac72a4fd327b3c44cbc8cd4e15338078a9cfceba08604c5a9f00fd7e61592 is running failed: container process not found" containerID="405ac72a4fd327b3c44cbc8cd4e15338078a9cfceba08604c5a9f00fd7e61592" cmd=["/bin/bash","-c","test -f /ready/ready"] Nov 22 07:58:01 crc kubenswrapper[4789]: E1122 07:58:01.986126 4789 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 405ac72a4fd327b3c44cbc8cd4e15338078a9cfceba08604c5a9f00fd7e61592 is running failed: container process not found" containerID="405ac72a4fd327b3c44cbc8cd4e15338078a9cfceba08604c5a9f00fd7e61592" cmd=["/bin/bash","-c","test -f /ready/ready"] Nov 22 07:58:01 crc kubenswrapper[4789]: E1122 07:58:01.986603 4789 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 405ac72a4fd327b3c44cbc8cd4e15338078a9cfceba08604c5a9f00fd7e61592 is running failed: container process not found" containerID="405ac72a4fd327b3c44cbc8cd4e15338078a9cfceba08604c5a9f00fd7e61592" cmd=["/bin/bash","-c","test -f /ready/ready"] Nov 22 07:58:01 crc kubenswrapper[4789]: E1122 07:58:01.986639 4789 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 405ac72a4fd327b3c44cbc8cd4e15338078a9cfceba08604c5a9f00fd7e61592 is running failed: container process not found" probeType="Readiness" pod="openshift-multus/cni-sysctl-allowlist-ds-gcnts" podUID="75fa6283-acd8-41bc-9803-2f55119f6828" containerName="kube-multus-additional-cni-plugins" Nov 22 07:58:02 crc kubenswrapper[4789]: I1122 07:58:02.898061 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 07:58:03 crc kubenswrapper[4789]: E1122 07:58:03.394711 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-gbjc4" podUID="d4a4f530-1afc-402b-8026-584bba4f8f1a" Nov 22 07:58:08 crc kubenswrapper[4789]: E1122 07:58:08.601655 4789 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Nov 22 07:58:08 crc kubenswrapper[4789]: E1122 07:58:08.602198 4789 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vlv48,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-5wgxf_openshift-marketplace(8235ef4a-0670-4b55-bfe6-4cc3cb1b24de): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 22 07:58:08 crc kubenswrapper[4789]: E1122 07:58:08.603540 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-5wgxf" podUID="8235ef4a-0670-4b55-bfe6-4cc3cb1b24de" Nov 22 07:58:08 crc kubenswrapper[4789]: I1122 07:58:08.664019 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_cni-sysctl-allowlist-ds-gcnts_75fa6283-acd8-41bc-9803-2f55119f6828/kube-multus-additional-cni-plugins/0.log" Nov 22 07:58:08 crc kubenswrapper[4789]: I1122 07:58:08.664082 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-multus/cni-sysctl-allowlist-ds-gcnts" Nov 22 07:58:08 crc kubenswrapper[4789]: I1122 07:58:08.723962 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/75fa6283-acd8-41bc-9803-2f55119f6828-ready\") pod \"75fa6283-acd8-41bc-9803-2f55119f6828\" (UID: \"75fa6283-acd8-41bc-9803-2f55119f6828\") " Nov 22 07:58:08 crc kubenswrapper[4789]: I1122 07:58:08.724027 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/75fa6283-acd8-41bc-9803-2f55119f6828-cni-sysctl-allowlist\") pod \"75fa6283-acd8-41bc-9803-2f55119f6828\" (UID: \"75fa6283-acd8-41bc-9803-2f55119f6828\") " Nov 22 07:58:08 crc kubenswrapper[4789]: I1122 07:58:08.724083 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6r7x6\" (UniqueName: \"kubernetes.io/projected/75fa6283-acd8-41bc-9803-2f55119f6828-kube-api-access-6r7x6\") pod \"75fa6283-acd8-41bc-9803-2f55119f6828\" (UID: \"75fa6283-acd8-41bc-9803-2f55119f6828\") " Nov 22 07:58:08 crc kubenswrapper[4789]: I1122 07:58:08.724116 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/75fa6283-acd8-41bc-9803-2f55119f6828-tuning-conf-dir\") pod \"75fa6283-acd8-41bc-9803-2f55119f6828\" (UID: \"75fa6283-acd8-41bc-9803-2f55119f6828\") " Nov 22 07:58:08 crc kubenswrapper[4789]: I1122 07:58:08.724352 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/75fa6283-acd8-41bc-9803-2f55119f6828-tuning-conf-dir" (OuterVolumeSpecName: "tuning-conf-dir") pod "75fa6283-acd8-41bc-9803-2f55119f6828" (UID: "75fa6283-acd8-41bc-9803-2f55119f6828"). InnerVolumeSpecName "tuning-conf-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 07:58:08 crc kubenswrapper[4789]: I1122 07:58:08.724618 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/75fa6283-acd8-41bc-9803-2f55119f6828-ready" (OuterVolumeSpecName: "ready") pod "75fa6283-acd8-41bc-9803-2f55119f6828" (UID: "75fa6283-acd8-41bc-9803-2f55119f6828"). InnerVolumeSpecName "ready". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 07:58:08 crc kubenswrapper[4789]: I1122 07:58:08.726226 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/75fa6283-acd8-41bc-9803-2f55119f6828-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "75fa6283-acd8-41bc-9803-2f55119f6828" (UID: "75fa6283-acd8-41bc-9803-2f55119f6828"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 07:58:08 crc kubenswrapper[4789]: I1122 07:58:08.731288 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75fa6283-acd8-41bc-9803-2f55119f6828-kube-api-access-6r7x6" (OuterVolumeSpecName: "kube-api-access-6r7x6") pod "75fa6283-acd8-41bc-9803-2f55119f6828" (UID: "75fa6283-acd8-41bc-9803-2f55119f6828"). InnerVolumeSpecName "kube-api-access-6r7x6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 07:58:08 crc kubenswrapper[4789]: I1122 07:58:08.826205 4789 reconciler_common.go:293] "Volume detached for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/75fa6283-acd8-41bc-9803-2f55119f6828-tuning-conf-dir\") on node \"crc\" DevicePath \"\"" Nov 22 07:58:08 crc kubenswrapper[4789]: I1122 07:58:08.826267 4789 reconciler_common.go:293] "Volume detached for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/75fa6283-acd8-41bc-9803-2f55119f6828-ready\") on node \"crc\" DevicePath \"\"" Nov 22 07:58:08 crc kubenswrapper[4789]: I1122 07:58:08.826287 4789 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/75fa6283-acd8-41bc-9803-2f55119f6828-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Nov 22 07:58:08 crc kubenswrapper[4789]: I1122 07:58:08.826310 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6r7x6\" (UniqueName: \"kubernetes.io/projected/75fa6283-acd8-41bc-9803-2f55119f6828-kube-api-access-6r7x6\") on node \"crc\" DevicePath \"\"" Nov 22 07:58:09 crc kubenswrapper[4789]: I1122 07:58:09.192616 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_cni-sysctl-allowlist-ds-gcnts_75fa6283-acd8-41bc-9803-2f55119f6828/kube-multus-additional-cni-plugins/0.log" Nov 22 07:58:09 crc kubenswrapper[4789]: I1122 07:58:09.192684 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/cni-sysctl-allowlist-ds-gcnts" event={"ID":"75fa6283-acd8-41bc-9803-2f55119f6828","Type":"ContainerDied","Data":"7ce73911b860efe1f373ef3f3bed89467e181e04382094247fab65fcc1777c66"} Nov 22 07:58:09 crc kubenswrapper[4789]: I1122 07:58:09.192730 4789 scope.go:117] "RemoveContainer" containerID="405ac72a4fd327b3c44cbc8cd4e15338078a9cfceba08604c5a9f00fd7e61592" Nov 22 07:58:09 crc kubenswrapper[4789]: I1122 07:58:09.192793 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-multus/cni-sysctl-allowlist-ds-gcnts" Nov 22 07:58:09 crc kubenswrapper[4789]: I1122 07:58:09.221985 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-multus/cni-sysctl-allowlist-ds-gcnts"] Nov 22 07:58:09 crc kubenswrapper[4789]: I1122 07:58:09.225614 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-multus/cni-sysctl-allowlist-ds-gcnts"] Nov 22 07:58:09 crc kubenswrapper[4789]: E1122 07:58:09.881770 4789 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Nov 22 07:58:09 crc kubenswrapper[4789]: E1122 07:58:09.882187 4789 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5hrkf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-rnkxn_openshift-marketplace(6a078eea-384c-41cc-ae2e-30e6696422a2): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 22 07:58:09 crc kubenswrapper[4789]: E1122 07:58:09.884270 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-rnkxn" podUID="6a078eea-384c-41cc-ae2e-30e6696422a2" Nov 22 07:58:09 crc kubenswrapper[4789]: I1122 07:58:09.971654 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="75fa6283-acd8-41bc-9803-2f55119f6828" path="/var/lib/kubelet/pods/75fa6283-acd8-41bc-9803-2f55119f6828/volumes" Nov 22 07:58:11 crc kubenswrapper[4789]: I1122 07:58:11.018845 4789 patch_prober.go:28] interesting pod/downloads-7954f5f757-tdpht container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Nov 22 07:58:11 crc kubenswrapper[4789]: I1122 07:58:11.018903 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-tdpht" podUID="83ac1d78-cc4c-48b7-b975-2f88802e2d17" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Nov 22 07:58:12 crc kubenswrapper[4789]: E1122 07:58:12.948183 4789 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Nov 22 07:58:12 crc kubenswrapper[4789]: E1122 07:58:12.948952 4789 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-745s6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-7t8sl_openshift-marketplace(7426acc3-0cbe-43f6-a327-886c9de10a94): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 22 07:58:12 crc kubenswrapper[4789]: E1122 07:58:12.950192 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-7t8sl" podUID="7426acc3-0cbe-43f6-a327-886c9de10a94" Nov 22 07:58:13 crc kubenswrapper[4789]: E1122 07:58:13.628381 4789 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Nov 22 07:58:13 crc kubenswrapper[4789]: E1122 07:58:13.628691 4789 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-cfnx7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-fnbtg_openshift-marketplace(ef38f088-0b6a-49a8-811f-3973b8207308): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 22 07:58:13 crc kubenswrapper[4789]: E1122 07:58:13.630009 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-fnbtg" podUID="ef38f088-0b6a-49a8-811f-3973b8207308" Nov 22 07:58:21 crc kubenswrapper[4789]: I1122 07:58:21.017412 4789 patch_prober.go:28] interesting pod/downloads-7954f5f757-tdpht container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Nov 22 07:58:21 crc kubenswrapper[4789]: I1122 07:58:21.017999 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-tdpht" podUID="83ac1d78-cc4c-48b7-b975-2f88802e2d17" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Nov 22 07:58:22 crc kubenswrapper[4789]: E1122 07:58:22.056565 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-fnbtg" podUID="ef38f088-0b6a-49a8-811f-3973b8207308" Nov 22 07:58:22 crc kubenswrapper[4789]: E1122 07:58:22.057147 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-7t8sl" podUID="7426acc3-0cbe-43f6-a327-886c9de10a94" Nov 22 07:58:22 crc kubenswrapper[4789]: E1122 07:58:22.091874 4789 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Nov 22 07:58:22 crc kubenswrapper[4789]: E1122 07:58:22.092061 4789 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-lnhh8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-4gz55_openshift-marketplace(92287f13-f1d5-422a-9107-f0495287e5ff): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 22 07:58:22 crc kubenswrapper[4789]: E1122 07:58:22.093256 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-4gz55" podUID="92287f13-f1d5-422a-9107-f0495287e5ff" Nov 22 07:58:27 crc kubenswrapper[4789]: E1122 07:58:27.088569 4789 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Nov 22 07:58:27 crc kubenswrapper[4789]: E1122 07:58:27.089055 4789 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vs8fh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-jdz2m_openshift-marketplace(2bbcd7cf-ccb1-4b91-9e71-2aa87677be19): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 22 07:58:27 crc kubenswrapper[4789]: E1122 07:58:27.090296 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-jdz2m" podUID="2bbcd7cf-ccb1-4b91-9e71-2aa87677be19" Nov 22 07:58:31 crc kubenswrapper[4789]: I1122 07:58:31.018224 4789 patch_prober.go:28] interesting pod/downloads-7954f5f757-tdpht container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Nov 22 07:58:31 crc kubenswrapper[4789]: I1122 07:58:31.018559 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-tdpht" podUID="83ac1d78-cc4c-48b7-b975-2f88802e2d17" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Nov 22 07:58:41 crc kubenswrapper[4789]: I1122 07:58:41.018295 4789 patch_prober.go:28] interesting pod/downloads-7954f5f757-tdpht container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Nov 22 07:58:41 crc kubenswrapper[4789]: I1122 07:58:41.018911 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-tdpht" podUID="83ac1d78-cc4c-48b7-b975-2f88802e2d17" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Nov 22 07:58:51 crc kubenswrapper[4789]: I1122 07:58:51.017897 4789 patch_prober.go:28] interesting pod/downloads-7954f5f757-tdpht container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Nov 22 07:58:51 crc kubenswrapper[4789]: I1122 07:58:51.018484 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-tdpht" podUID="83ac1d78-cc4c-48b7-b975-2f88802e2d17" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Nov 22 07:59:01 crc kubenswrapper[4789]: I1122 07:59:01.017883 4789 patch_prober.go:28] interesting pod/downloads-7954f5f757-tdpht container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Nov 22 07:59:01 crc kubenswrapper[4789]: I1122 07:59:01.018554 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-tdpht" podUID="83ac1d78-cc4c-48b7-b975-2f88802e2d17" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Nov 22 07:59:01 crc kubenswrapper[4789]: I1122 07:59:01.504736 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-tdpht" event={"ID":"83ac1d78-cc4c-48b7-b975-2f88802e2d17","Type":"ContainerStarted","Data":"8c00b262a23389bcbc79f2bdd15b526d7d691027207bdf55bda0b923ce44172d"} Nov 22 07:59:01 crc kubenswrapper[4789]: I1122 07:59:01.505926 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-tdpht" Nov 22 07:59:01 crc kubenswrapper[4789]: I1122 07:59:01.506092 4789 patch_prober.go:28] interesting pod/downloads-7954f5f757-tdpht container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Nov 22 07:59:01 crc kubenswrapper[4789]: I1122 07:59:01.506138 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-tdpht" podUID="83ac1d78-cc4c-48b7-b975-2f88802e2d17" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Nov 22 07:59:02 crc kubenswrapper[4789]: I1122 07:59:02.513128 4789 generic.go:334] "Generic (PLEG): container finished" podID="6a078eea-384c-41cc-ae2e-30e6696422a2" containerID="913cc1afb00fa11c8a3f08822b5d01bc71a938a2fa63d2c35c15260cd886f3f8" exitCode=0 Nov 22 07:59:02 crc kubenswrapper[4789]: I1122 07:59:02.513781 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rnkxn" event={"ID":"6a078eea-384c-41cc-ae2e-30e6696422a2","Type":"ContainerDied","Data":"913cc1afb00fa11c8a3f08822b5d01bc71a938a2fa63d2c35c15260cd886f3f8"} Nov 22 07:59:02 crc kubenswrapper[4789]: I1122 07:59:02.524001 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fnbtg" event={"ID":"ef38f088-0b6a-49a8-811f-3973b8207308","Type":"ContainerStarted","Data":"1f4aa4596a863189049dd757318390335eafae227666162ee8b2bacd048ac07d"} Nov 22 07:59:02 crc kubenswrapper[4789]: I1122 07:59:02.535169 4789 generic.go:334] "Generic (PLEG): container finished" podID="0d93195d-1cef-4b5e-bc10-c696e82d5e38" containerID="5778e8d078417c11639db2cd0a95c4ef9b482c2a914543d40321c591fe25049f" exitCode=0 Nov 22 07:59:02 crc kubenswrapper[4789]: I1122 07:59:02.535290 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xxw9z" event={"ID":"0d93195d-1cef-4b5e-bc10-c696e82d5e38","Type":"ContainerDied","Data":"5778e8d078417c11639db2cd0a95c4ef9b482c2a914543d40321c591fe25049f"} Nov 22 07:59:02 crc kubenswrapper[4789]: I1122 07:59:02.547973 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jdz2m" event={"ID":"2bbcd7cf-ccb1-4b91-9e71-2aa87677be19","Type":"ContainerStarted","Data":"9ee1547395e4e03e0ee50d99b4764acdf0a3df06d3f8b323334099913b9c8224"} Nov 22 07:59:02 crc kubenswrapper[4789]: I1122 07:59:02.559045 4789 generic.go:334] "Generic (PLEG): container finished" podID="8235ef4a-0670-4b55-bfe6-4cc3cb1b24de" containerID="6b1f2ba6174a685908041aeae46a98db241deac5d93d783eabcf5df90333b251" exitCode=0 Nov 22 07:59:02 crc kubenswrapper[4789]: I1122 07:59:02.559150 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5wgxf" event={"ID":"8235ef4a-0670-4b55-bfe6-4cc3cb1b24de","Type":"ContainerDied","Data":"6b1f2ba6174a685908041aeae46a98db241deac5d93d783eabcf5df90333b251"} Nov 22 07:59:02 crc kubenswrapper[4789]: I1122 07:59:02.565849 4789 generic.go:334] "Generic (PLEG): container finished" podID="d4a4f530-1afc-402b-8026-584bba4f8f1a" containerID="76501fa438782b1c9700d1539ccbad88ac50a96f05bc101608894bf1711b6f9d" exitCode=0 Nov 22 07:59:02 crc kubenswrapper[4789]: I1122 07:59:02.566318 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gbjc4" event={"ID":"d4a4f530-1afc-402b-8026-584bba4f8f1a","Type":"ContainerDied","Data":"76501fa438782b1c9700d1539ccbad88ac50a96f05bc101608894bf1711b6f9d"} Nov 22 07:59:02 crc kubenswrapper[4789]: I1122 07:59:02.569964 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4gz55" event={"ID":"92287f13-f1d5-422a-9107-f0495287e5ff","Type":"ContainerStarted","Data":"d7685033894cc0e7cdb85fe9c8867032adf67e369e118eec787f6a246d2da3fc"} Nov 22 07:59:02 crc kubenswrapper[4789]: I1122 07:59:02.579924 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7t8sl" event={"ID":"7426acc3-0cbe-43f6-a327-886c9de10a94","Type":"ContainerStarted","Data":"f95d16bef21f99bf0cd2f9e7b4cf410f5d2751c9c8328eeb1981f21907a7e1bf"} Nov 22 07:59:02 crc kubenswrapper[4789]: I1122 07:59:02.580425 4789 patch_prober.go:28] interesting pod/downloads-7954f5f757-tdpht container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Nov 22 07:59:02 crc kubenswrapper[4789]: I1122 07:59:02.580467 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-tdpht" podUID="83ac1d78-cc4c-48b7-b975-2f88802e2d17" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Nov 22 07:59:03 crc kubenswrapper[4789]: I1122 07:59:03.585986 4789 generic.go:334] "Generic (PLEG): container finished" podID="7426acc3-0cbe-43f6-a327-886c9de10a94" containerID="f95d16bef21f99bf0cd2f9e7b4cf410f5d2751c9c8328eeb1981f21907a7e1bf" exitCode=0 Nov 22 07:59:03 crc kubenswrapper[4789]: I1122 07:59:03.586071 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7t8sl" event={"ID":"7426acc3-0cbe-43f6-a327-886c9de10a94","Type":"ContainerDied","Data":"f95d16bef21f99bf0cd2f9e7b4cf410f5d2751c9c8328eeb1981f21907a7e1bf"} Nov 22 07:59:03 crc kubenswrapper[4789]: I1122 07:59:03.589642 4789 generic.go:334] "Generic (PLEG): container finished" podID="ef38f088-0b6a-49a8-811f-3973b8207308" containerID="1f4aa4596a863189049dd757318390335eafae227666162ee8b2bacd048ac07d" exitCode=0 Nov 22 07:59:03 crc kubenswrapper[4789]: I1122 07:59:03.589814 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fnbtg" event={"ID":"ef38f088-0b6a-49a8-811f-3973b8207308","Type":"ContainerDied","Data":"1f4aa4596a863189049dd757318390335eafae227666162ee8b2bacd048ac07d"} Nov 22 07:59:03 crc kubenswrapper[4789]: I1122 07:59:03.595139 4789 generic.go:334] "Generic (PLEG): container finished" podID="2bbcd7cf-ccb1-4b91-9e71-2aa87677be19" containerID="9ee1547395e4e03e0ee50d99b4764acdf0a3df06d3f8b323334099913b9c8224" exitCode=0 Nov 22 07:59:03 crc kubenswrapper[4789]: I1122 07:59:03.595277 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jdz2m" event={"ID":"2bbcd7cf-ccb1-4b91-9e71-2aa87677be19","Type":"ContainerDied","Data":"9ee1547395e4e03e0ee50d99b4764acdf0a3df06d3f8b323334099913b9c8224"} Nov 22 07:59:03 crc kubenswrapper[4789]: I1122 07:59:03.603271 4789 generic.go:334] "Generic (PLEG): container finished" podID="92287f13-f1d5-422a-9107-f0495287e5ff" containerID="d7685033894cc0e7cdb85fe9c8867032adf67e369e118eec787f6a246d2da3fc" exitCode=0 Nov 22 07:59:03 crc kubenswrapper[4789]: I1122 07:59:03.603597 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4gz55" event={"ID":"92287f13-f1d5-422a-9107-f0495287e5ff","Type":"ContainerDied","Data":"d7685033894cc0e7cdb85fe9c8867032adf67e369e118eec787f6a246d2da3fc"} Nov 22 07:59:03 crc kubenswrapper[4789]: I1122 07:59:03.604546 4789 patch_prober.go:28] interesting pod/downloads-7954f5f757-tdpht container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Nov 22 07:59:03 crc kubenswrapper[4789]: I1122 07:59:03.604583 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-tdpht" podUID="83ac1d78-cc4c-48b7-b975-2f88802e2d17" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Nov 22 07:59:11 crc kubenswrapper[4789]: I1122 07:59:11.039941 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-tdpht" Nov 22 07:59:35 crc kubenswrapper[4789]: I1122 07:59:35.372275 4789 patch_prober.go:28] interesting pod/machine-config-daemon-pxvrm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 07:59:35 crc kubenswrapper[4789]: I1122 07:59:35.372869 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 07:59:38 crc kubenswrapper[4789]: I1122 07:59:38.807600 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5wgxf" event={"ID":"8235ef4a-0670-4b55-bfe6-4cc3cb1b24de","Type":"ContainerStarted","Data":"9ac489479b5b330e2d26f3a51496cab2b8298caf3f196d7ef43c98283d7c272b"} Nov 22 07:59:38 crc kubenswrapper[4789]: I1122 07:59:38.810204 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4gz55" event={"ID":"92287f13-f1d5-422a-9107-f0495287e5ff","Type":"ContainerStarted","Data":"4e54a6def062beeba8345cbb1c4b937411a8d5c71f85714d7c6ff651d1b81637"} Nov 22 07:59:38 crc kubenswrapper[4789]: I1122 07:59:38.812629 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7t8sl" event={"ID":"7426acc3-0cbe-43f6-a327-886c9de10a94","Type":"ContainerStarted","Data":"a8182ffc687b14ba917c20bb199af1ad661d56966207b973881b0f5668cbddf3"} Nov 22 07:59:38 crc kubenswrapper[4789]: I1122 07:59:38.814303 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rnkxn" event={"ID":"6a078eea-384c-41cc-ae2e-30e6696422a2","Type":"ContainerStarted","Data":"64c3ed797114f69d42914a3534dc34a4cd2eb8c25c8dd25aa7dd991df65aa3a2"} Nov 22 07:59:38 crc kubenswrapper[4789]: I1122 07:59:38.816049 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fnbtg" event={"ID":"ef38f088-0b6a-49a8-811f-3973b8207308","Type":"ContainerStarted","Data":"c29f9a75de198f73c7a05dee3c5a3986b7172607646e3ace489843f50846e465"} Nov 22 07:59:38 crc kubenswrapper[4789]: I1122 07:59:38.817640 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xxw9z" event={"ID":"0d93195d-1cef-4b5e-bc10-c696e82d5e38","Type":"ContainerStarted","Data":"992bea44011eaed84ef0b6fc54e09b5a9ac97d9e6b70ca932e455a4a945f953f"} Nov 22 07:59:38 crc kubenswrapper[4789]: I1122 07:59:38.820886 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jdz2m" event={"ID":"2bbcd7cf-ccb1-4b91-9e71-2aa87677be19","Type":"ContainerStarted","Data":"98103cc7b23e40c25b0d4918ebc1f0b6f2f9d1abd1bac6bd361eb6b477cf1715"} Nov 22 07:59:38 crc kubenswrapper[4789]: I1122 07:59:38.822826 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gbjc4" event={"ID":"d4a4f530-1afc-402b-8026-584bba4f8f1a","Type":"ContainerStarted","Data":"61a0c521d576645357310609f30f970de20cb2843d225885d81fc416c2fa2ea0"} Nov 22 07:59:38 crc kubenswrapper[4789]: I1122 07:59:38.838854 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-xxw9z" podStartSLOduration=28.979606158 podStartE2EDuration="2m6.838839145s" podCreationTimestamp="2025-11-22 07:57:32 +0000 UTC" firstStartedPulling="2025-11-22 07:57:33.870109399 +0000 UTC m=+68.104509672" lastFinishedPulling="2025-11-22 07:59:11.729342386 +0000 UTC m=+165.963742659" observedRunningTime="2025-11-22 07:59:38.836065858 +0000 UTC m=+193.070466141" watchObservedRunningTime="2025-11-22 07:59:38.838839145 +0000 UTC m=+193.073239418" Nov 22 07:59:39 crc kubenswrapper[4789]: I1122 07:59:39.851688 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-jdz2m" podStartSLOduration=5.708204183 podStartE2EDuration="2m7.851666251s" podCreationTimestamp="2025-11-22 07:57:32 +0000 UTC" firstStartedPulling="2025-11-22 07:57:33.858907834 +0000 UTC m=+68.093308107" lastFinishedPulling="2025-11-22 07:59:36.002369902 +0000 UTC m=+190.236770175" observedRunningTime="2025-11-22 07:59:39.84912946 +0000 UTC m=+194.083529753" watchObservedRunningTime="2025-11-22 07:59:39.851666251 +0000 UTC m=+194.086066534" Nov 22 07:59:39 crc kubenswrapper[4789]: I1122 07:59:39.868280 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-7t8sl" podStartSLOduration=10.001214124 podStartE2EDuration="2m10.868262447s" podCreationTimestamp="2025-11-22 07:57:29 +0000 UTC" firstStartedPulling="2025-11-22 07:57:31.800645926 +0000 UTC m=+66.035046189" lastFinishedPulling="2025-11-22 07:59:32.667694239 +0000 UTC m=+186.902094512" observedRunningTime="2025-11-22 07:59:39.866092237 +0000 UTC m=+194.100492540" watchObservedRunningTime="2025-11-22 07:59:39.868262447 +0000 UTC m=+194.102662720" Nov 22 07:59:39 crc kubenswrapper[4789]: I1122 07:59:39.905939 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-gbjc4" podStartSLOduration=6.716219927 podStartE2EDuration="2m10.905917896s" podCreationTimestamp="2025-11-22 07:57:29 +0000 UTC" firstStartedPulling="2025-11-22 07:57:31.826016875 +0000 UTC m=+66.060417148" lastFinishedPulling="2025-11-22 07:59:36.015714844 +0000 UTC m=+190.250115117" observedRunningTime="2025-11-22 07:59:39.887974701 +0000 UTC m=+194.122374984" watchObservedRunningTime="2025-11-22 07:59:39.905917896 +0000 UTC m=+194.140318169" Nov 22 07:59:39 crc kubenswrapper[4789]: I1122 07:59:39.906627 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-4gz55" podStartSLOduration=11.318621554 podStartE2EDuration="2m6.906621996s" podCreationTimestamp="2025-11-22 07:57:33 +0000 UTC" firstStartedPulling="2025-11-22 07:57:34.887240942 +0000 UTC m=+69.121641215" lastFinishedPulling="2025-11-22 07:59:30.475241384 +0000 UTC m=+184.709641657" observedRunningTime="2025-11-22 07:59:39.904813155 +0000 UTC m=+194.139213458" watchObservedRunningTime="2025-11-22 07:59:39.906621996 +0000 UTC m=+194.141022279" Nov 22 07:59:39 crc kubenswrapper[4789]: I1122 07:59:39.925988 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-rnkxn" podStartSLOduration=5.753096734 podStartE2EDuration="2m9.92596687s" podCreationTimestamp="2025-11-22 07:57:30 +0000 UTC" firstStartedPulling="2025-11-22 07:57:31.835189145 +0000 UTC m=+66.069589418" lastFinishedPulling="2025-11-22 07:59:36.008059281 +0000 UTC m=+190.242459554" observedRunningTime="2025-11-22 07:59:39.92100632 +0000 UTC m=+194.155406613" watchObservedRunningTime="2025-11-22 07:59:39.92596687 +0000 UTC m=+194.160367133" Nov 22 07:59:39 crc kubenswrapper[4789]: I1122 07:59:39.943313 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-5wgxf" podStartSLOduration=6.810391285 podStartE2EDuration="2m8.943293496s" podCreationTimestamp="2025-11-22 07:57:31 +0000 UTC" firstStartedPulling="2025-11-22 07:57:33.869492532 +0000 UTC m=+68.103892805" lastFinishedPulling="2025-11-22 07:59:36.002394733 +0000 UTC m=+190.236795016" observedRunningTime="2025-11-22 07:59:39.938497442 +0000 UTC m=+194.172897725" watchObservedRunningTime="2025-11-22 07:59:39.943293496 +0000 UTC m=+194.177693769" Nov 22 07:59:39 crc kubenswrapper[4789]: I1122 07:59:39.958780 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-fnbtg" podStartSLOduration=7.553039908 podStartE2EDuration="2m9.958741491s" podCreationTimestamp="2025-11-22 07:57:30 +0000 UTC" firstStartedPulling="2025-11-22 07:57:32.837450444 +0000 UTC m=+67.071850717" lastFinishedPulling="2025-11-22 07:59:35.243152027 +0000 UTC m=+189.477552300" observedRunningTime="2025-11-22 07:59:39.957314801 +0000 UTC m=+194.191715094" watchObservedRunningTime="2025-11-22 07:59:39.958741491 +0000 UTC m=+194.193141764" Nov 22 07:59:40 crc kubenswrapper[4789]: I1122 07:59:40.062310 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-7t8sl" Nov 22 07:59:40 crc kubenswrapper[4789]: I1122 07:59:40.062673 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-7t8sl" Nov 22 07:59:40 crc kubenswrapper[4789]: I1122 07:59:40.233093 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-gbjc4" Nov 22 07:59:40 crc kubenswrapper[4789]: I1122 07:59:40.233154 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-gbjc4" Nov 22 07:59:40 crc kubenswrapper[4789]: I1122 07:59:40.428510 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-rnkxn" Nov 22 07:59:40 crc kubenswrapper[4789]: I1122 07:59:40.428568 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-rnkxn" Nov 22 07:59:40 crc kubenswrapper[4789]: I1122 07:59:40.647101 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-fnbtg" Nov 22 07:59:40 crc kubenswrapper[4789]: I1122 07:59:40.647268 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-fnbtg" Nov 22 07:59:42 crc kubenswrapper[4789]: I1122 07:59:42.016597 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-5wgxf" Nov 22 07:59:42 crc kubenswrapper[4789]: I1122 07:59:42.016665 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-5wgxf" Nov 22 07:59:42 crc kubenswrapper[4789]: I1122 07:59:42.368197 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-gbjc4" podUID="d4a4f530-1afc-402b-8026-584bba4f8f1a" containerName="registry-server" probeResult="failure" output=< Nov 22 07:59:42 crc kubenswrapper[4789]: timeout: failed to connect service ":50051" within 1s Nov 22 07:59:42 crc kubenswrapper[4789]: > Nov 22 07:59:42 crc kubenswrapper[4789]: I1122 07:59:42.368418 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-fnbtg" podUID="ef38f088-0b6a-49a8-811f-3973b8207308" containerName="registry-server" probeResult="failure" output=< Nov 22 07:59:42 crc kubenswrapper[4789]: timeout: failed to connect service ":50051" within 1s Nov 22 07:59:42 crc kubenswrapper[4789]: > Nov 22 07:59:42 crc kubenswrapper[4789]: I1122 07:59:42.369353 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-7t8sl" podUID="7426acc3-0cbe-43f6-a327-886c9de10a94" containerName="registry-server" probeResult="failure" output=< Nov 22 07:59:42 crc kubenswrapper[4789]: timeout: failed to connect service ":50051" within 1s Nov 22 07:59:42 crc kubenswrapper[4789]: > Nov 22 07:59:42 crc kubenswrapper[4789]: I1122 07:59:42.370876 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-rnkxn" podUID="6a078eea-384c-41cc-ae2e-30e6696422a2" containerName="registry-server" probeResult="failure" output=< Nov 22 07:59:42 crc kubenswrapper[4789]: timeout: failed to connect service ":50051" within 1s Nov 22 07:59:42 crc kubenswrapper[4789]: > Nov 22 07:59:42 crc kubenswrapper[4789]: I1122 07:59:42.426669 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-xxw9z" Nov 22 07:59:42 crc kubenswrapper[4789]: I1122 07:59:42.426723 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-xxw9z" Nov 22 07:59:43 crc kubenswrapper[4789]: I1122 07:59:43.060863 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-5wgxf" podUID="8235ef4a-0670-4b55-bfe6-4cc3cb1b24de" containerName="registry-server" probeResult="failure" output=< Nov 22 07:59:43 crc kubenswrapper[4789]: timeout: failed to connect service ":50051" within 1s Nov 22 07:59:43 crc kubenswrapper[4789]: > Nov 22 07:59:43 crc kubenswrapper[4789]: I1122 07:59:43.249996 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-jdz2m" Nov 22 07:59:43 crc kubenswrapper[4789]: I1122 07:59:43.250061 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-jdz2m" Nov 22 07:59:43 crc kubenswrapper[4789]: I1122 07:59:43.470632 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-xxw9z" podUID="0d93195d-1cef-4b5e-bc10-c696e82d5e38" containerName="registry-server" probeResult="failure" output=< Nov 22 07:59:43 crc kubenswrapper[4789]: timeout: failed to connect service ":50051" within 1s Nov 22 07:59:43 crc kubenswrapper[4789]: > Nov 22 07:59:43 crc kubenswrapper[4789]: I1122 07:59:43.622648 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-4gz55" Nov 22 07:59:43 crc kubenswrapper[4789]: I1122 07:59:43.622697 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-4gz55" Nov 22 07:59:44 crc kubenswrapper[4789]: I1122 07:59:44.283106 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-jdz2m" podUID="2bbcd7cf-ccb1-4b91-9e71-2aa87677be19" containerName="registry-server" probeResult="failure" output=< Nov 22 07:59:44 crc kubenswrapper[4789]: timeout: failed to connect service ":50051" within 1s Nov 22 07:59:44 crc kubenswrapper[4789]: > Nov 22 07:59:44 crc kubenswrapper[4789]: I1122 07:59:44.666886 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-4gz55" podUID="92287f13-f1d5-422a-9107-f0495287e5ff" containerName="registry-server" probeResult="failure" output=< Nov 22 07:59:44 crc kubenswrapper[4789]: timeout: failed to connect service ":50051" within 1s Nov 22 07:59:44 crc kubenswrapper[4789]: > Nov 22 07:59:50 crc kubenswrapper[4789]: I1122 07:59:50.116261 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-7t8sl" Nov 22 07:59:50 crc kubenswrapper[4789]: I1122 07:59:50.169882 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-7t8sl" Nov 22 07:59:50 crc kubenswrapper[4789]: I1122 07:59:50.269051 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-gbjc4" Nov 22 07:59:50 crc kubenswrapper[4789]: I1122 07:59:50.306108 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-gbjc4" Nov 22 07:59:50 crc kubenswrapper[4789]: I1122 07:59:50.471571 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-rnkxn" Nov 22 07:59:50 crc kubenswrapper[4789]: I1122 07:59:50.507685 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-rnkxn" Nov 22 07:59:50 crc kubenswrapper[4789]: I1122 07:59:50.686167 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-fnbtg" Nov 22 07:59:50 crc kubenswrapper[4789]: I1122 07:59:50.729351 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-fnbtg" Nov 22 07:59:52 crc kubenswrapper[4789]: I1122 07:59:52.054534 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-5wgxf" Nov 22 07:59:52 crc kubenswrapper[4789]: I1122 07:59:52.093716 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-5wgxf" Nov 22 07:59:52 crc kubenswrapper[4789]: I1122 07:59:52.148276 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-fnbtg"] Nov 22 07:59:52 crc kubenswrapper[4789]: I1122 07:59:52.148554 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-fnbtg" podUID="ef38f088-0b6a-49a8-811f-3973b8207308" containerName="registry-server" containerID="cri-o://c29f9a75de198f73c7a05dee3c5a3986b7172607646e3ace489843f50846e465" gracePeriod=2 Nov 22 07:59:52 crc kubenswrapper[4789]: I1122 07:59:52.475224 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-xxw9z" Nov 22 07:59:52 crc kubenswrapper[4789]: I1122 07:59:52.519111 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-xxw9z" Nov 22 07:59:52 crc kubenswrapper[4789]: I1122 07:59:52.547189 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fnbtg" Nov 22 07:59:52 crc kubenswrapper[4789]: I1122 07:59:52.688661 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ef38f088-0b6a-49a8-811f-3973b8207308-catalog-content\") pod \"ef38f088-0b6a-49a8-811f-3973b8207308\" (UID: \"ef38f088-0b6a-49a8-811f-3973b8207308\") " Nov 22 07:59:52 crc kubenswrapper[4789]: I1122 07:59:52.688857 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ef38f088-0b6a-49a8-811f-3973b8207308-utilities\") pod \"ef38f088-0b6a-49a8-811f-3973b8207308\" (UID: \"ef38f088-0b6a-49a8-811f-3973b8207308\") " Nov 22 07:59:52 crc kubenswrapper[4789]: I1122 07:59:52.688904 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfnx7\" (UniqueName: \"kubernetes.io/projected/ef38f088-0b6a-49a8-811f-3973b8207308-kube-api-access-cfnx7\") pod \"ef38f088-0b6a-49a8-811f-3973b8207308\" (UID: \"ef38f088-0b6a-49a8-811f-3973b8207308\") " Nov 22 07:59:52 crc kubenswrapper[4789]: I1122 07:59:52.689849 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ef38f088-0b6a-49a8-811f-3973b8207308-utilities" (OuterVolumeSpecName: "utilities") pod "ef38f088-0b6a-49a8-811f-3973b8207308" (UID: "ef38f088-0b6a-49a8-811f-3973b8207308"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 07:59:52 crc kubenswrapper[4789]: I1122 07:59:52.694361 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef38f088-0b6a-49a8-811f-3973b8207308-kube-api-access-cfnx7" (OuterVolumeSpecName: "kube-api-access-cfnx7") pod "ef38f088-0b6a-49a8-811f-3973b8207308" (UID: "ef38f088-0b6a-49a8-811f-3973b8207308"). InnerVolumeSpecName "kube-api-access-cfnx7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 07:59:52 crc kubenswrapper[4789]: I1122 07:59:52.732717 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ef38f088-0b6a-49a8-811f-3973b8207308-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ef38f088-0b6a-49a8-811f-3973b8207308" (UID: "ef38f088-0b6a-49a8-811f-3973b8207308"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 07:59:52 crc kubenswrapper[4789]: I1122 07:59:52.749194 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rnkxn"] Nov 22 07:59:52 crc kubenswrapper[4789]: I1122 07:59:52.749510 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-rnkxn" podUID="6a078eea-384c-41cc-ae2e-30e6696422a2" containerName="registry-server" containerID="cri-o://64c3ed797114f69d42914a3534dc34a4cd2eb8c25c8dd25aa7dd991df65aa3a2" gracePeriod=2 Nov 22 07:59:52 crc kubenswrapper[4789]: I1122 07:59:52.790574 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ef38f088-0b6a-49a8-811f-3973b8207308-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 07:59:52 crc kubenswrapper[4789]: I1122 07:59:52.790608 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfnx7\" (UniqueName: \"kubernetes.io/projected/ef38f088-0b6a-49a8-811f-3973b8207308-kube-api-access-cfnx7\") on node \"crc\" DevicePath \"\"" Nov 22 07:59:52 crc kubenswrapper[4789]: I1122 07:59:52.790621 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ef38f088-0b6a-49a8-811f-3973b8207308-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 07:59:52 crc kubenswrapper[4789]: I1122 07:59:52.902131 4789 generic.go:334] "Generic (PLEG): container finished" podID="ef38f088-0b6a-49a8-811f-3973b8207308" containerID="c29f9a75de198f73c7a05dee3c5a3986b7172607646e3ace489843f50846e465" exitCode=0 Nov 22 07:59:52 crc kubenswrapper[4789]: I1122 07:59:52.902203 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fnbtg" event={"ID":"ef38f088-0b6a-49a8-811f-3973b8207308","Type":"ContainerDied","Data":"c29f9a75de198f73c7a05dee3c5a3986b7172607646e3ace489843f50846e465"} Nov 22 07:59:52 crc kubenswrapper[4789]: I1122 07:59:52.902296 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fnbtg" event={"ID":"ef38f088-0b6a-49a8-811f-3973b8207308","Type":"ContainerDied","Data":"95d4319f107342c47d17dbed8a879bb0c84f8a395182231e2c5a3b2ba25c6668"} Nov 22 07:59:52 crc kubenswrapper[4789]: I1122 07:59:52.902350 4789 scope.go:117] "RemoveContainer" containerID="c29f9a75de198f73c7a05dee3c5a3986b7172607646e3ace489843f50846e465" Nov 22 07:59:52 crc kubenswrapper[4789]: I1122 07:59:52.902212 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fnbtg" Nov 22 07:59:52 crc kubenswrapper[4789]: I1122 07:59:52.904898 4789 generic.go:334] "Generic (PLEG): container finished" podID="6a078eea-384c-41cc-ae2e-30e6696422a2" containerID="64c3ed797114f69d42914a3534dc34a4cd2eb8c25c8dd25aa7dd991df65aa3a2" exitCode=0 Nov 22 07:59:52 crc kubenswrapper[4789]: I1122 07:59:52.905316 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rnkxn" event={"ID":"6a078eea-384c-41cc-ae2e-30e6696422a2","Type":"ContainerDied","Data":"64c3ed797114f69d42914a3534dc34a4cd2eb8c25c8dd25aa7dd991df65aa3a2"} Nov 22 07:59:52 crc kubenswrapper[4789]: I1122 07:59:52.922033 4789 scope.go:117] "RemoveContainer" containerID="1f4aa4596a863189049dd757318390335eafae227666162ee8b2bacd048ac07d" Nov 22 07:59:52 crc kubenswrapper[4789]: I1122 07:59:52.932567 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-fnbtg"] Nov 22 07:59:52 crc kubenswrapper[4789]: I1122 07:59:52.945067 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-fnbtg"] Nov 22 07:59:52 crc kubenswrapper[4789]: I1122 07:59:52.955673 4789 scope.go:117] "RemoveContainer" containerID="909ffa9c112c40bf527b814f367d47a9af890485b374aefe90d7f37a1613c140" Nov 22 07:59:52 crc kubenswrapper[4789]: I1122 07:59:52.981108 4789 scope.go:117] "RemoveContainer" containerID="c29f9a75de198f73c7a05dee3c5a3986b7172607646e3ace489843f50846e465" Nov 22 07:59:52 crc kubenswrapper[4789]: E1122 07:59:52.981605 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c29f9a75de198f73c7a05dee3c5a3986b7172607646e3ace489843f50846e465\": container with ID starting with c29f9a75de198f73c7a05dee3c5a3986b7172607646e3ace489843f50846e465 not found: ID does not exist" containerID="c29f9a75de198f73c7a05dee3c5a3986b7172607646e3ace489843f50846e465" Nov 22 07:59:52 crc kubenswrapper[4789]: I1122 07:59:52.981641 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c29f9a75de198f73c7a05dee3c5a3986b7172607646e3ace489843f50846e465"} err="failed to get container status \"c29f9a75de198f73c7a05dee3c5a3986b7172607646e3ace489843f50846e465\": rpc error: code = NotFound desc = could not find container \"c29f9a75de198f73c7a05dee3c5a3986b7172607646e3ace489843f50846e465\": container with ID starting with c29f9a75de198f73c7a05dee3c5a3986b7172607646e3ace489843f50846e465 not found: ID does not exist" Nov 22 07:59:52 crc kubenswrapper[4789]: I1122 07:59:52.981671 4789 scope.go:117] "RemoveContainer" containerID="1f4aa4596a863189049dd757318390335eafae227666162ee8b2bacd048ac07d" Nov 22 07:59:52 crc kubenswrapper[4789]: E1122 07:59:52.982355 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1f4aa4596a863189049dd757318390335eafae227666162ee8b2bacd048ac07d\": container with ID starting with 1f4aa4596a863189049dd757318390335eafae227666162ee8b2bacd048ac07d not found: ID does not exist" containerID="1f4aa4596a863189049dd757318390335eafae227666162ee8b2bacd048ac07d" Nov 22 07:59:52 crc kubenswrapper[4789]: I1122 07:59:52.982385 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f4aa4596a863189049dd757318390335eafae227666162ee8b2bacd048ac07d"} err="failed to get container status \"1f4aa4596a863189049dd757318390335eafae227666162ee8b2bacd048ac07d\": rpc error: code = NotFound desc = could not find container \"1f4aa4596a863189049dd757318390335eafae227666162ee8b2bacd048ac07d\": container with ID starting with 1f4aa4596a863189049dd757318390335eafae227666162ee8b2bacd048ac07d not found: ID does not exist" Nov 22 07:59:52 crc kubenswrapper[4789]: I1122 07:59:52.982427 4789 scope.go:117] "RemoveContainer" containerID="909ffa9c112c40bf527b814f367d47a9af890485b374aefe90d7f37a1613c140" Nov 22 07:59:52 crc kubenswrapper[4789]: E1122 07:59:52.982936 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"909ffa9c112c40bf527b814f367d47a9af890485b374aefe90d7f37a1613c140\": container with ID starting with 909ffa9c112c40bf527b814f367d47a9af890485b374aefe90d7f37a1613c140 not found: ID does not exist" containerID="909ffa9c112c40bf527b814f367d47a9af890485b374aefe90d7f37a1613c140" Nov 22 07:59:52 crc kubenswrapper[4789]: I1122 07:59:52.982962 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"909ffa9c112c40bf527b814f367d47a9af890485b374aefe90d7f37a1613c140"} err="failed to get container status \"909ffa9c112c40bf527b814f367d47a9af890485b374aefe90d7f37a1613c140\": rpc error: code = NotFound desc = could not find container \"909ffa9c112c40bf527b814f367d47a9af890485b374aefe90d7f37a1613c140\": container with ID starting with 909ffa9c112c40bf527b814f367d47a9af890485b374aefe90d7f37a1613c140 not found: ID does not exist" Nov 22 07:59:53 crc kubenswrapper[4789]: I1122 07:59:53.036237 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rnkxn" Nov 22 07:59:53 crc kubenswrapper[4789]: I1122 07:59:53.195298 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a078eea-384c-41cc-ae2e-30e6696422a2-catalog-content\") pod \"6a078eea-384c-41cc-ae2e-30e6696422a2\" (UID: \"6a078eea-384c-41cc-ae2e-30e6696422a2\") " Nov 22 07:59:53 crc kubenswrapper[4789]: I1122 07:59:53.195437 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5hrkf\" (UniqueName: \"kubernetes.io/projected/6a078eea-384c-41cc-ae2e-30e6696422a2-kube-api-access-5hrkf\") pod \"6a078eea-384c-41cc-ae2e-30e6696422a2\" (UID: \"6a078eea-384c-41cc-ae2e-30e6696422a2\") " Nov 22 07:59:53 crc kubenswrapper[4789]: I1122 07:59:53.195498 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a078eea-384c-41cc-ae2e-30e6696422a2-utilities\") pod \"6a078eea-384c-41cc-ae2e-30e6696422a2\" (UID: \"6a078eea-384c-41cc-ae2e-30e6696422a2\") " Nov 22 07:59:53 crc kubenswrapper[4789]: I1122 07:59:53.196276 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6a078eea-384c-41cc-ae2e-30e6696422a2-utilities" (OuterVolumeSpecName: "utilities") pod "6a078eea-384c-41cc-ae2e-30e6696422a2" (UID: "6a078eea-384c-41cc-ae2e-30e6696422a2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 07:59:53 crc kubenswrapper[4789]: I1122 07:59:53.199673 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a078eea-384c-41cc-ae2e-30e6696422a2-kube-api-access-5hrkf" (OuterVolumeSpecName: "kube-api-access-5hrkf") pod "6a078eea-384c-41cc-ae2e-30e6696422a2" (UID: "6a078eea-384c-41cc-ae2e-30e6696422a2"). InnerVolumeSpecName "kube-api-access-5hrkf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 07:59:53 crc kubenswrapper[4789]: I1122 07:59:53.243047 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6a078eea-384c-41cc-ae2e-30e6696422a2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6a078eea-384c-41cc-ae2e-30e6696422a2" (UID: "6a078eea-384c-41cc-ae2e-30e6696422a2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 07:59:53 crc kubenswrapper[4789]: I1122 07:59:53.295226 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-jdz2m" Nov 22 07:59:53 crc kubenswrapper[4789]: I1122 07:59:53.296787 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5hrkf\" (UniqueName: \"kubernetes.io/projected/6a078eea-384c-41cc-ae2e-30e6696422a2-kube-api-access-5hrkf\") on node \"crc\" DevicePath \"\"" Nov 22 07:59:53 crc kubenswrapper[4789]: I1122 07:59:53.296882 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a078eea-384c-41cc-ae2e-30e6696422a2-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 07:59:53 crc kubenswrapper[4789]: I1122 07:59:53.296895 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a078eea-384c-41cc-ae2e-30e6696422a2-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 07:59:53 crc kubenswrapper[4789]: I1122 07:59:53.335357 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-jdz2m" Nov 22 07:59:53 crc kubenswrapper[4789]: I1122 07:59:53.661769 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-4gz55" Nov 22 07:59:53 crc kubenswrapper[4789]: I1122 07:59:53.699938 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-4gz55" Nov 22 07:59:53 crc kubenswrapper[4789]: I1122 07:59:53.910479 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rnkxn" event={"ID":"6a078eea-384c-41cc-ae2e-30e6696422a2","Type":"ContainerDied","Data":"8cfc825de34d8dc7d7bfd849784ebe1f9e7a88b1a7382c7bdd503d546ec40b41"} Nov 22 07:59:53 crc kubenswrapper[4789]: I1122 07:59:53.910524 4789 scope.go:117] "RemoveContainer" containerID="64c3ed797114f69d42914a3534dc34a4cd2eb8c25c8dd25aa7dd991df65aa3a2" Nov 22 07:59:53 crc kubenswrapper[4789]: I1122 07:59:53.910524 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rnkxn" Nov 22 07:59:53 crc kubenswrapper[4789]: I1122 07:59:53.924743 4789 scope.go:117] "RemoveContainer" containerID="913cc1afb00fa11c8a3f08822b5d01bc71a938a2fa63d2c35c15260cd886f3f8" Nov 22 07:59:53 crc kubenswrapper[4789]: I1122 07:59:53.938638 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rnkxn"] Nov 22 07:59:53 crc kubenswrapper[4789]: I1122 07:59:53.940286 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-rnkxn"] Nov 22 07:59:53 crc kubenswrapper[4789]: I1122 07:59:53.943674 4789 scope.go:117] "RemoveContainer" containerID="85b63870e53a2e3ec8be518bf3b3f93585cac0bf20b5e9cb6305b0fed81c9a69" Nov 22 07:59:53 crc kubenswrapper[4789]: I1122 07:59:53.977150 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a078eea-384c-41cc-ae2e-30e6696422a2" path="/var/lib/kubelet/pods/6a078eea-384c-41cc-ae2e-30e6696422a2/volumes" Nov 22 07:59:53 crc kubenswrapper[4789]: I1122 07:59:53.977735 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ef38f088-0b6a-49a8-811f-3973b8207308" path="/var/lib/kubelet/pods/ef38f088-0b6a-49a8-811f-3973b8207308/volumes" Nov 22 07:59:54 crc kubenswrapper[4789]: I1122 07:59:54.547498 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xxw9z"] Nov 22 07:59:54 crc kubenswrapper[4789]: I1122 07:59:54.548055 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-xxw9z" podUID="0d93195d-1cef-4b5e-bc10-c696e82d5e38" containerName="registry-server" containerID="cri-o://992bea44011eaed84ef0b6fc54e09b5a9ac97d9e6b70ca932e455a4a945f953f" gracePeriod=2 Nov 22 07:59:54 crc kubenswrapper[4789]: I1122 07:59:54.873365 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xxw9z" Nov 22 07:59:54 crc kubenswrapper[4789]: I1122 07:59:54.918264 4789 generic.go:334] "Generic (PLEG): container finished" podID="0d93195d-1cef-4b5e-bc10-c696e82d5e38" containerID="992bea44011eaed84ef0b6fc54e09b5a9ac97d9e6b70ca932e455a4a945f953f" exitCode=0 Nov 22 07:59:54 crc kubenswrapper[4789]: I1122 07:59:54.918347 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xxw9z" Nov 22 07:59:54 crc kubenswrapper[4789]: I1122 07:59:54.918352 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xxw9z" event={"ID":"0d93195d-1cef-4b5e-bc10-c696e82d5e38","Type":"ContainerDied","Data":"992bea44011eaed84ef0b6fc54e09b5a9ac97d9e6b70ca932e455a4a945f953f"} Nov 22 07:59:54 crc kubenswrapper[4789]: I1122 07:59:54.918395 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xxw9z" event={"ID":"0d93195d-1cef-4b5e-bc10-c696e82d5e38","Type":"ContainerDied","Data":"d6fa1d063398f944333d2ff2272485c3a3cd2da792824f624d724404191a1754"} Nov 22 07:59:54 crc kubenswrapper[4789]: I1122 07:59:54.918413 4789 scope.go:117] "RemoveContainer" containerID="992bea44011eaed84ef0b6fc54e09b5a9ac97d9e6b70ca932e455a4a945f953f" Nov 22 07:59:54 crc kubenswrapper[4789]: I1122 07:59:54.932665 4789 scope.go:117] "RemoveContainer" containerID="5778e8d078417c11639db2cd0a95c4ef9b482c2a914543d40321c591fe25049f" Nov 22 07:59:54 crc kubenswrapper[4789]: I1122 07:59:54.946101 4789 scope.go:117] "RemoveContainer" containerID="d1740965270fe0eb6a052fefdfa72dfede028461d06e159fef3321ae477efd04" Nov 22 07:59:54 crc kubenswrapper[4789]: I1122 07:59:54.982476 4789 scope.go:117] "RemoveContainer" containerID="992bea44011eaed84ef0b6fc54e09b5a9ac97d9e6b70ca932e455a4a945f953f" Nov 22 07:59:54 crc kubenswrapper[4789]: E1122 07:59:54.982987 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"992bea44011eaed84ef0b6fc54e09b5a9ac97d9e6b70ca932e455a4a945f953f\": container with ID starting with 992bea44011eaed84ef0b6fc54e09b5a9ac97d9e6b70ca932e455a4a945f953f not found: ID does not exist" containerID="992bea44011eaed84ef0b6fc54e09b5a9ac97d9e6b70ca932e455a4a945f953f" Nov 22 07:59:54 crc kubenswrapper[4789]: I1122 07:59:54.983021 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"992bea44011eaed84ef0b6fc54e09b5a9ac97d9e6b70ca932e455a4a945f953f"} err="failed to get container status \"992bea44011eaed84ef0b6fc54e09b5a9ac97d9e6b70ca932e455a4a945f953f\": rpc error: code = NotFound desc = could not find container \"992bea44011eaed84ef0b6fc54e09b5a9ac97d9e6b70ca932e455a4a945f953f\": container with ID starting with 992bea44011eaed84ef0b6fc54e09b5a9ac97d9e6b70ca932e455a4a945f953f not found: ID does not exist" Nov 22 07:59:54 crc kubenswrapper[4789]: I1122 07:59:54.983048 4789 scope.go:117] "RemoveContainer" containerID="5778e8d078417c11639db2cd0a95c4ef9b482c2a914543d40321c591fe25049f" Nov 22 07:59:54 crc kubenswrapper[4789]: E1122 07:59:54.983416 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5778e8d078417c11639db2cd0a95c4ef9b482c2a914543d40321c591fe25049f\": container with ID starting with 5778e8d078417c11639db2cd0a95c4ef9b482c2a914543d40321c591fe25049f not found: ID does not exist" containerID="5778e8d078417c11639db2cd0a95c4ef9b482c2a914543d40321c591fe25049f" Nov 22 07:59:54 crc kubenswrapper[4789]: I1122 07:59:54.983437 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5778e8d078417c11639db2cd0a95c4ef9b482c2a914543d40321c591fe25049f"} err="failed to get container status \"5778e8d078417c11639db2cd0a95c4ef9b482c2a914543d40321c591fe25049f\": rpc error: code = NotFound desc = could not find container \"5778e8d078417c11639db2cd0a95c4ef9b482c2a914543d40321c591fe25049f\": container with ID starting with 5778e8d078417c11639db2cd0a95c4ef9b482c2a914543d40321c591fe25049f not found: ID does not exist" Nov 22 07:59:54 crc kubenswrapper[4789]: I1122 07:59:54.983453 4789 scope.go:117] "RemoveContainer" containerID="d1740965270fe0eb6a052fefdfa72dfede028461d06e159fef3321ae477efd04" Nov 22 07:59:54 crc kubenswrapper[4789]: E1122 07:59:54.983620 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d1740965270fe0eb6a052fefdfa72dfede028461d06e159fef3321ae477efd04\": container with ID starting with d1740965270fe0eb6a052fefdfa72dfede028461d06e159fef3321ae477efd04 not found: ID does not exist" containerID="d1740965270fe0eb6a052fefdfa72dfede028461d06e159fef3321ae477efd04" Nov 22 07:59:54 crc kubenswrapper[4789]: I1122 07:59:54.983642 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d1740965270fe0eb6a052fefdfa72dfede028461d06e159fef3321ae477efd04"} err="failed to get container status \"d1740965270fe0eb6a052fefdfa72dfede028461d06e159fef3321ae477efd04\": rpc error: code = NotFound desc = could not find container \"d1740965270fe0eb6a052fefdfa72dfede028461d06e159fef3321ae477efd04\": container with ID starting with d1740965270fe0eb6a052fefdfa72dfede028461d06e159fef3321ae477efd04 not found: ID does not exist" Nov 22 07:59:55 crc kubenswrapper[4789]: I1122 07:59:55.014624 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0d93195d-1cef-4b5e-bc10-c696e82d5e38-utilities\") pod \"0d93195d-1cef-4b5e-bc10-c696e82d5e38\" (UID: \"0d93195d-1cef-4b5e-bc10-c696e82d5e38\") " Nov 22 07:59:55 crc kubenswrapper[4789]: I1122 07:59:55.015052 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xb8h9\" (UniqueName: \"kubernetes.io/projected/0d93195d-1cef-4b5e-bc10-c696e82d5e38-kube-api-access-xb8h9\") pod \"0d93195d-1cef-4b5e-bc10-c696e82d5e38\" (UID: \"0d93195d-1cef-4b5e-bc10-c696e82d5e38\") " Nov 22 07:59:55 crc kubenswrapper[4789]: I1122 07:59:55.015110 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0d93195d-1cef-4b5e-bc10-c696e82d5e38-catalog-content\") pod \"0d93195d-1cef-4b5e-bc10-c696e82d5e38\" (UID: \"0d93195d-1cef-4b5e-bc10-c696e82d5e38\") " Nov 22 07:59:55 crc kubenswrapper[4789]: I1122 07:59:55.015980 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0d93195d-1cef-4b5e-bc10-c696e82d5e38-utilities" (OuterVolumeSpecName: "utilities") pod "0d93195d-1cef-4b5e-bc10-c696e82d5e38" (UID: "0d93195d-1cef-4b5e-bc10-c696e82d5e38"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 07:59:55 crc kubenswrapper[4789]: I1122 07:59:55.021892 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d93195d-1cef-4b5e-bc10-c696e82d5e38-kube-api-access-xb8h9" (OuterVolumeSpecName: "kube-api-access-xb8h9") pod "0d93195d-1cef-4b5e-bc10-c696e82d5e38" (UID: "0d93195d-1cef-4b5e-bc10-c696e82d5e38"). InnerVolumeSpecName "kube-api-access-xb8h9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 07:59:55 crc kubenswrapper[4789]: I1122 07:59:55.035991 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0d93195d-1cef-4b5e-bc10-c696e82d5e38-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0d93195d-1cef-4b5e-bc10-c696e82d5e38" (UID: "0d93195d-1cef-4b5e-bc10-c696e82d5e38"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 07:59:55 crc kubenswrapper[4789]: I1122 07:59:55.116919 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0d93195d-1cef-4b5e-bc10-c696e82d5e38-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 07:59:55 crc kubenswrapper[4789]: I1122 07:59:55.116957 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xb8h9\" (UniqueName: \"kubernetes.io/projected/0d93195d-1cef-4b5e-bc10-c696e82d5e38-kube-api-access-xb8h9\") on node \"crc\" DevicePath \"\"" Nov 22 07:59:55 crc kubenswrapper[4789]: I1122 07:59:55.116969 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0d93195d-1cef-4b5e-bc10-c696e82d5e38-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 07:59:55 crc kubenswrapper[4789]: I1122 07:59:55.248885 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xxw9z"] Nov 22 07:59:55 crc kubenswrapper[4789]: I1122 07:59:55.251791 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-xxw9z"] Nov 22 07:59:55 crc kubenswrapper[4789]: I1122 07:59:55.972044 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0d93195d-1cef-4b5e-bc10-c696e82d5e38" path="/var/lib/kubelet/pods/0d93195d-1cef-4b5e-bc10-c696e82d5e38/volumes" Nov 22 07:59:57 crc kubenswrapper[4789]: I1122 07:59:57.148013 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4gz55"] Nov 22 07:59:57 crc kubenswrapper[4789]: I1122 07:59:57.148908 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-4gz55" podUID="92287f13-f1d5-422a-9107-f0495287e5ff" containerName="registry-server" containerID="cri-o://4e54a6def062beeba8345cbb1c4b937411a8d5c71f85714d7c6ff651d1b81637" gracePeriod=2 Nov 22 07:59:57 crc kubenswrapper[4789]: I1122 07:59:57.506319 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4gz55" Nov 22 07:59:57 crc kubenswrapper[4789]: I1122 07:59:57.646369 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92287f13-f1d5-422a-9107-f0495287e5ff-utilities\") pod \"92287f13-f1d5-422a-9107-f0495287e5ff\" (UID: \"92287f13-f1d5-422a-9107-f0495287e5ff\") " Nov 22 07:59:57 crc kubenswrapper[4789]: I1122 07:59:57.646452 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lnhh8\" (UniqueName: \"kubernetes.io/projected/92287f13-f1d5-422a-9107-f0495287e5ff-kube-api-access-lnhh8\") pod \"92287f13-f1d5-422a-9107-f0495287e5ff\" (UID: \"92287f13-f1d5-422a-9107-f0495287e5ff\") " Nov 22 07:59:57 crc kubenswrapper[4789]: I1122 07:59:57.646485 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92287f13-f1d5-422a-9107-f0495287e5ff-catalog-content\") pod \"92287f13-f1d5-422a-9107-f0495287e5ff\" (UID: \"92287f13-f1d5-422a-9107-f0495287e5ff\") " Nov 22 07:59:57 crc kubenswrapper[4789]: I1122 07:59:57.647420 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/92287f13-f1d5-422a-9107-f0495287e5ff-utilities" (OuterVolumeSpecName: "utilities") pod "92287f13-f1d5-422a-9107-f0495287e5ff" (UID: "92287f13-f1d5-422a-9107-f0495287e5ff"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 07:59:57 crc kubenswrapper[4789]: I1122 07:59:57.654030 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92287f13-f1d5-422a-9107-f0495287e5ff-kube-api-access-lnhh8" (OuterVolumeSpecName: "kube-api-access-lnhh8") pod "92287f13-f1d5-422a-9107-f0495287e5ff" (UID: "92287f13-f1d5-422a-9107-f0495287e5ff"). InnerVolumeSpecName "kube-api-access-lnhh8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 07:59:57 crc kubenswrapper[4789]: I1122 07:59:57.728932 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/92287f13-f1d5-422a-9107-f0495287e5ff-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "92287f13-f1d5-422a-9107-f0495287e5ff" (UID: "92287f13-f1d5-422a-9107-f0495287e5ff"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 07:59:57 crc kubenswrapper[4789]: I1122 07:59:57.747680 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92287f13-f1d5-422a-9107-f0495287e5ff-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 07:59:57 crc kubenswrapper[4789]: I1122 07:59:57.747724 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lnhh8\" (UniqueName: \"kubernetes.io/projected/92287f13-f1d5-422a-9107-f0495287e5ff-kube-api-access-lnhh8\") on node \"crc\" DevicePath \"\"" Nov 22 07:59:57 crc kubenswrapper[4789]: I1122 07:59:57.747736 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92287f13-f1d5-422a-9107-f0495287e5ff-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 07:59:57 crc kubenswrapper[4789]: I1122 07:59:57.938216 4789 generic.go:334] "Generic (PLEG): container finished" podID="92287f13-f1d5-422a-9107-f0495287e5ff" containerID="4e54a6def062beeba8345cbb1c4b937411a8d5c71f85714d7c6ff651d1b81637" exitCode=0 Nov 22 07:59:57 crc kubenswrapper[4789]: I1122 07:59:57.938255 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4gz55" event={"ID":"92287f13-f1d5-422a-9107-f0495287e5ff","Type":"ContainerDied","Data":"4e54a6def062beeba8345cbb1c4b937411a8d5c71f85714d7c6ff651d1b81637"} Nov 22 07:59:57 crc kubenswrapper[4789]: I1122 07:59:57.938280 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4gz55" event={"ID":"92287f13-f1d5-422a-9107-f0495287e5ff","Type":"ContainerDied","Data":"9a48770e409dd0e1f2acf609d1dfa08cfa32d424f329edbbd6e56a4c3f89aa53"} Nov 22 07:59:57 crc kubenswrapper[4789]: I1122 07:59:57.938299 4789 scope.go:117] "RemoveContainer" containerID="4e54a6def062beeba8345cbb1c4b937411a8d5c71f85714d7c6ff651d1b81637" Nov 22 07:59:57 crc kubenswrapper[4789]: I1122 07:59:57.938307 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4gz55" Nov 22 07:59:57 crc kubenswrapper[4789]: I1122 07:59:57.961797 4789 scope.go:117] "RemoveContainer" containerID="d7685033894cc0e7cdb85fe9c8867032adf67e369e118eec787f6a246d2da3fc" Nov 22 07:59:57 crc kubenswrapper[4789]: I1122 07:59:57.977918 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4gz55"] Nov 22 07:59:57 crc kubenswrapper[4789]: I1122 07:59:57.979616 4789 scope.go:117] "RemoveContainer" containerID="69ad89253c458146e9be3bc545098e967c7ac62b48b82435e9ca7ebab6e2e7d6" Nov 22 07:59:57 crc kubenswrapper[4789]: I1122 07:59:57.980382 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-4gz55"] Nov 22 07:59:57 crc kubenswrapper[4789]: I1122 07:59:57.999266 4789 scope.go:117] "RemoveContainer" containerID="4e54a6def062beeba8345cbb1c4b937411a8d5c71f85714d7c6ff651d1b81637" Nov 22 07:59:57 crc kubenswrapper[4789]: E1122 07:59:57.999739 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4e54a6def062beeba8345cbb1c4b937411a8d5c71f85714d7c6ff651d1b81637\": container with ID starting with 4e54a6def062beeba8345cbb1c4b937411a8d5c71f85714d7c6ff651d1b81637 not found: ID does not exist" containerID="4e54a6def062beeba8345cbb1c4b937411a8d5c71f85714d7c6ff651d1b81637" Nov 22 07:59:57 crc kubenswrapper[4789]: I1122 07:59:57.999788 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e54a6def062beeba8345cbb1c4b937411a8d5c71f85714d7c6ff651d1b81637"} err="failed to get container status \"4e54a6def062beeba8345cbb1c4b937411a8d5c71f85714d7c6ff651d1b81637\": rpc error: code = NotFound desc = could not find container \"4e54a6def062beeba8345cbb1c4b937411a8d5c71f85714d7c6ff651d1b81637\": container with ID starting with 4e54a6def062beeba8345cbb1c4b937411a8d5c71f85714d7c6ff651d1b81637 not found: ID does not exist" Nov 22 07:59:57 crc kubenswrapper[4789]: I1122 07:59:57.999814 4789 scope.go:117] "RemoveContainer" containerID="d7685033894cc0e7cdb85fe9c8867032adf67e369e118eec787f6a246d2da3fc" Nov 22 07:59:58 crc kubenswrapper[4789]: E1122 07:59:58.000421 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d7685033894cc0e7cdb85fe9c8867032adf67e369e118eec787f6a246d2da3fc\": container with ID starting with d7685033894cc0e7cdb85fe9c8867032adf67e369e118eec787f6a246d2da3fc not found: ID does not exist" containerID="d7685033894cc0e7cdb85fe9c8867032adf67e369e118eec787f6a246d2da3fc" Nov 22 07:59:58 crc kubenswrapper[4789]: I1122 07:59:58.000448 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d7685033894cc0e7cdb85fe9c8867032adf67e369e118eec787f6a246d2da3fc"} err="failed to get container status \"d7685033894cc0e7cdb85fe9c8867032adf67e369e118eec787f6a246d2da3fc\": rpc error: code = NotFound desc = could not find container \"d7685033894cc0e7cdb85fe9c8867032adf67e369e118eec787f6a246d2da3fc\": container with ID starting with d7685033894cc0e7cdb85fe9c8867032adf67e369e118eec787f6a246d2da3fc not found: ID does not exist" Nov 22 07:59:58 crc kubenswrapper[4789]: I1122 07:59:58.000471 4789 scope.go:117] "RemoveContainer" containerID="69ad89253c458146e9be3bc545098e967c7ac62b48b82435e9ca7ebab6e2e7d6" Nov 22 07:59:58 crc kubenswrapper[4789]: E1122 07:59:58.000795 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"69ad89253c458146e9be3bc545098e967c7ac62b48b82435e9ca7ebab6e2e7d6\": container with ID starting with 69ad89253c458146e9be3bc545098e967c7ac62b48b82435e9ca7ebab6e2e7d6 not found: ID does not exist" containerID="69ad89253c458146e9be3bc545098e967c7ac62b48b82435e9ca7ebab6e2e7d6" Nov 22 07:59:58 crc kubenswrapper[4789]: I1122 07:59:58.000826 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"69ad89253c458146e9be3bc545098e967c7ac62b48b82435e9ca7ebab6e2e7d6"} err="failed to get container status \"69ad89253c458146e9be3bc545098e967c7ac62b48b82435e9ca7ebab6e2e7d6\": rpc error: code = NotFound desc = could not find container \"69ad89253c458146e9be3bc545098e967c7ac62b48b82435e9ca7ebab6e2e7d6\": container with ID starting with 69ad89253c458146e9be3bc545098e967c7ac62b48b82435e9ca7ebab6e2e7d6 not found: ID does not exist" Nov 22 07:59:59 crc kubenswrapper[4789]: I1122 07:59:59.974148 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="92287f13-f1d5-422a-9107-f0495287e5ff" path="/var/lib/kubelet/pods/92287f13-f1d5-422a-9107-f0495287e5ff/volumes" Nov 22 08:00:00 crc kubenswrapper[4789]: I1122 08:00:00.135975 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396640-mhslx"] Nov 22 08:00:00 crc kubenswrapper[4789]: E1122 08:00:00.136422 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92287f13-f1d5-422a-9107-f0495287e5ff" containerName="extract-content" Nov 22 08:00:00 crc kubenswrapper[4789]: I1122 08:00:00.136548 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="92287f13-f1d5-422a-9107-f0495287e5ff" containerName="extract-content" Nov 22 08:00:00 crc kubenswrapper[4789]: E1122 08:00:00.136654 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a078eea-384c-41cc-ae2e-30e6696422a2" containerName="registry-server" Nov 22 08:00:00 crc kubenswrapper[4789]: I1122 08:00:00.136731 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a078eea-384c-41cc-ae2e-30e6696422a2" containerName="registry-server" Nov 22 08:00:00 crc kubenswrapper[4789]: E1122 08:00:00.136862 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75fa6283-acd8-41bc-9803-2f55119f6828" containerName="kube-multus-additional-cni-plugins" Nov 22 08:00:00 crc kubenswrapper[4789]: I1122 08:00:00.136934 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="75fa6283-acd8-41bc-9803-2f55119f6828" containerName="kube-multus-additional-cni-plugins" Nov 22 08:00:00 crc kubenswrapper[4789]: E1122 08:00:00.137018 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d93195d-1cef-4b5e-bc10-c696e82d5e38" containerName="extract-content" Nov 22 08:00:00 crc kubenswrapper[4789]: I1122 08:00:00.137121 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d93195d-1cef-4b5e-bc10-c696e82d5e38" containerName="extract-content" Nov 22 08:00:00 crc kubenswrapper[4789]: E1122 08:00:00.137224 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef38f088-0b6a-49a8-811f-3973b8207308" containerName="registry-server" Nov 22 08:00:00 crc kubenswrapper[4789]: I1122 08:00:00.137299 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef38f088-0b6a-49a8-811f-3973b8207308" containerName="registry-server" Nov 22 08:00:00 crc kubenswrapper[4789]: E1122 08:00:00.137377 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d93195d-1cef-4b5e-bc10-c696e82d5e38" containerName="registry-server" Nov 22 08:00:00 crc kubenswrapper[4789]: I1122 08:00:00.137461 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d93195d-1cef-4b5e-bc10-c696e82d5e38" containerName="registry-server" Nov 22 08:00:00 crc kubenswrapper[4789]: E1122 08:00:00.137540 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f42980d5-344a-4cf7-b937-431ea412e9ab" containerName="pruner" Nov 22 08:00:00 crc kubenswrapper[4789]: I1122 08:00:00.137610 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="f42980d5-344a-4cf7-b937-431ea412e9ab" containerName="pruner" Nov 22 08:00:00 crc kubenswrapper[4789]: E1122 08:00:00.137689 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a078eea-384c-41cc-ae2e-30e6696422a2" containerName="extract-utilities" Nov 22 08:00:00 crc kubenswrapper[4789]: I1122 08:00:00.137791 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a078eea-384c-41cc-ae2e-30e6696422a2" containerName="extract-utilities" Nov 22 08:00:00 crc kubenswrapper[4789]: E1122 08:00:00.137883 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92287f13-f1d5-422a-9107-f0495287e5ff" containerName="extract-utilities" Nov 22 08:00:00 crc kubenswrapper[4789]: I1122 08:00:00.137965 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="92287f13-f1d5-422a-9107-f0495287e5ff" containerName="extract-utilities" Nov 22 08:00:00 crc kubenswrapper[4789]: E1122 08:00:00.138052 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d93195d-1cef-4b5e-bc10-c696e82d5e38" containerName="extract-utilities" Nov 22 08:00:00 crc kubenswrapper[4789]: I1122 08:00:00.138134 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d93195d-1cef-4b5e-bc10-c696e82d5e38" containerName="extract-utilities" Nov 22 08:00:00 crc kubenswrapper[4789]: E1122 08:00:00.138225 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef38f088-0b6a-49a8-811f-3973b8207308" containerName="extract-utilities" Nov 22 08:00:00 crc kubenswrapper[4789]: I1122 08:00:00.138306 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef38f088-0b6a-49a8-811f-3973b8207308" containerName="extract-utilities" Nov 22 08:00:00 crc kubenswrapper[4789]: E1122 08:00:00.138384 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef38f088-0b6a-49a8-811f-3973b8207308" containerName="extract-content" Nov 22 08:00:00 crc kubenswrapper[4789]: I1122 08:00:00.138457 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef38f088-0b6a-49a8-811f-3973b8207308" containerName="extract-content" Nov 22 08:00:00 crc kubenswrapper[4789]: E1122 08:00:00.138633 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92287f13-f1d5-422a-9107-f0495287e5ff" containerName="registry-server" Nov 22 08:00:00 crc kubenswrapper[4789]: I1122 08:00:00.138734 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="92287f13-f1d5-422a-9107-f0495287e5ff" containerName="registry-server" Nov 22 08:00:00 crc kubenswrapper[4789]: E1122 08:00:00.138835 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a078eea-384c-41cc-ae2e-30e6696422a2" containerName="extract-content" Nov 22 08:00:00 crc kubenswrapper[4789]: I1122 08:00:00.138923 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a078eea-384c-41cc-ae2e-30e6696422a2" containerName="extract-content" Nov 22 08:00:00 crc kubenswrapper[4789]: I1122 08:00:00.139151 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a078eea-384c-41cc-ae2e-30e6696422a2" containerName="registry-server" Nov 22 08:00:00 crc kubenswrapper[4789]: I1122 08:00:00.139236 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="75fa6283-acd8-41bc-9803-2f55119f6828" containerName="kube-multus-additional-cni-plugins" Nov 22 08:00:00 crc kubenswrapper[4789]: I1122 08:00:00.139310 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d93195d-1cef-4b5e-bc10-c696e82d5e38" containerName="registry-server" Nov 22 08:00:00 crc kubenswrapper[4789]: I1122 08:00:00.139389 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="92287f13-f1d5-422a-9107-f0495287e5ff" containerName="registry-server" Nov 22 08:00:00 crc kubenswrapper[4789]: I1122 08:00:00.139466 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="f42980d5-344a-4cf7-b937-431ea412e9ab" containerName="pruner" Nov 22 08:00:00 crc kubenswrapper[4789]: I1122 08:00:00.139548 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef38f088-0b6a-49a8-811f-3973b8207308" containerName="registry-server" Nov 22 08:00:00 crc kubenswrapper[4789]: I1122 08:00:00.140121 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396640-mhslx" Nov 22 08:00:00 crc kubenswrapper[4789]: I1122 08:00:00.143581 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396640-mhslx"] Nov 22 08:00:00 crc kubenswrapper[4789]: I1122 08:00:00.178538 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/032913e2-a38e-4c15-b11b-5e47cb60c458-secret-volume\") pod \"collect-profiles-29396640-mhslx\" (UID: \"032913e2-a38e-4c15-b11b-5e47cb60c458\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396640-mhslx" Nov 22 08:00:00 crc kubenswrapper[4789]: I1122 08:00:00.178596 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dhhgw\" (UniqueName: \"kubernetes.io/projected/032913e2-a38e-4c15-b11b-5e47cb60c458-kube-api-access-dhhgw\") pod \"collect-profiles-29396640-mhslx\" (UID: \"032913e2-a38e-4c15-b11b-5e47cb60c458\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396640-mhslx" Nov 22 08:00:00 crc kubenswrapper[4789]: I1122 08:00:00.178651 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/032913e2-a38e-4c15-b11b-5e47cb60c458-config-volume\") pod \"collect-profiles-29396640-mhslx\" (UID: \"032913e2-a38e-4c15-b11b-5e47cb60c458\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396640-mhslx" Nov 22 08:00:00 crc kubenswrapper[4789]: I1122 08:00:00.188788 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 22 08:00:00 crc kubenswrapper[4789]: I1122 08:00:00.188835 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 22 08:00:00 crc kubenswrapper[4789]: I1122 08:00:00.279469 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/032913e2-a38e-4c15-b11b-5e47cb60c458-secret-volume\") pod \"collect-profiles-29396640-mhslx\" (UID: \"032913e2-a38e-4c15-b11b-5e47cb60c458\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396640-mhslx" Nov 22 08:00:00 crc kubenswrapper[4789]: I1122 08:00:00.279514 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dhhgw\" (UniqueName: \"kubernetes.io/projected/032913e2-a38e-4c15-b11b-5e47cb60c458-kube-api-access-dhhgw\") pod \"collect-profiles-29396640-mhslx\" (UID: \"032913e2-a38e-4c15-b11b-5e47cb60c458\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396640-mhslx" Nov 22 08:00:00 crc kubenswrapper[4789]: I1122 08:00:00.279867 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/032913e2-a38e-4c15-b11b-5e47cb60c458-config-volume\") pod \"collect-profiles-29396640-mhslx\" (UID: \"032913e2-a38e-4c15-b11b-5e47cb60c458\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396640-mhslx" Nov 22 08:00:00 crc kubenswrapper[4789]: I1122 08:00:00.280665 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/032913e2-a38e-4c15-b11b-5e47cb60c458-config-volume\") pod \"collect-profiles-29396640-mhslx\" (UID: \"032913e2-a38e-4c15-b11b-5e47cb60c458\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396640-mhslx" Nov 22 08:00:00 crc kubenswrapper[4789]: I1122 08:00:00.284562 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/032913e2-a38e-4c15-b11b-5e47cb60c458-secret-volume\") pod \"collect-profiles-29396640-mhslx\" (UID: \"032913e2-a38e-4c15-b11b-5e47cb60c458\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396640-mhslx" Nov 22 08:00:00 crc kubenswrapper[4789]: I1122 08:00:00.294272 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dhhgw\" (UniqueName: \"kubernetes.io/projected/032913e2-a38e-4c15-b11b-5e47cb60c458-kube-api-access-dhhgw\") pod \"collect-profiles-29396640-mhslx\" (UID: \"032913e2-a38e-4c15-b11b-5e47cb60c458\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396640-mhslx" Nov 22 08:00:00 crc kubenswrapper[4789]: I1122 08:00:00.506278 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396640-mhslx" Nov 22 08:00:00 crc kubenswrapper[4789]: I1122 08:00:00.678806 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396640-mhslx"] Nov 22 08:00:00 crc kubenswrapper[4789]: I1122 08:00:00.954423 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29396640-mhslx" event={"ID":"032913e2-a38e-4c15-b11b-5e47cb60c458","Type":"ContainerStarted","Data":"820e8a8bd1a43d8843e05cf917d843b4574a9c927cf518d8137935cd9321f470"} Nov 22 08:00:01 crc kubenswrapper[4789]: I1122 08:00:01.962121 4789 generic.go:334] "Generic (PLEG): container finished" podID="032913e2-a38e-4c15-b11b-5e47cb60c458" containerID="71f0fe50a32dfe9aeda16a4708c8a0cf7f62844adc58aa09c8a8444050c78770" exitCode=0 Nov 22 08:00:01 crc kubenswrapper[4789]: I1122 08:00:01.962286 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29396640-mhslx" event={"ID":"032913e2-a38e-4c15-b11b-5e47cb60c458","Type":"ContainerDied","Data":"71f0fe50a32dfe9aeda16a4708c8a0cf7f62844adc58aa09c8a8444050c78770"} Nov 22 08:00:03 crc kubenswrapper[4789]: I1122 08:00:03.171890 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396640-mhslx" Nov 22 08:00:03 crc kubenswrapper[4789]: I1122 08:00:03.316923 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/032913e2-a38e-4c15-b11b-5e47cb60c458-config-volume\") pod \"032913e2-a38e-4c15-b11b-5e47cb60c458\" (UID: \"032913e2-a38e-4c15-b11b-5e47cb60c458\") " Nov 22 08:00:03 crc kubenswrapper[4789]: I1122 08:00:03.316984 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/032913e2-a38e-4c15-b11b-5e47cb60c458-secret-volume\") pod \"032913e2-a38e-4c15-b11b-5e47cb60c458\" (UID: \"032913e2-a38e-4c15-b11b-5e47cb60c458\") " Nov 22 08:00:03 crc kubenswrapper[4789]: I1122 08:00:03.317058 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dhhgw\" (UniqueName: \"kubernetes.io/projected/032913e2-a38e-4c15-b11b-5e47cb60c458-kube-api-access-dhhgw\") pod \"032913e2-a38e-4c15-b11b-5e47cb60c458\" (UID: \"032913e2-a38e-4c15-b11b-5e47cb60c458\") " Nov 22 08:00:03 crc kubenswrapper[4789]: I1122 08:00:03.317991 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/032913e2-a38e-4c15-b11b-5e47cb60c458-config-volume" (OuterVolumeSpecName: "config-volume") pod "032913e2-a38e-4c15-b11b-5e47cb60c458" (UID: "032913e2-a38e-4c15-b11b-5e47cb60c458"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:00:03 crc kubenswrapper[4789]: I1122 08:00:03.325013 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/032913e2-a38e-4c15-b11b-5e47cb60c458-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "032913e2-a38e-4c15-b11b-5e47cb60c458" (UID: "032913e2-a38e-4c15-b11b-5e47cb60c458"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:00:03 crc kubenswrapper[4789]: I1122 08:00:03.325943 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/032913e2-a38e-4c15-b11b-5e47cb60c458-kube-api-access-dhhgw" (OuterVolumeSpecName: "kube-api-access-dhhgw") pod "032913e2-a38e-4c15-b11b-5e47cb60c458" (UID: "032913e2-a38e-4c15-b11b-5e47cb60c458"). InnerVolumeSpecName "kube-api-access-dhhgw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:00:03 crc kubenswrapper[4789]: I1122 08:00:03.418511 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dhhgw\" (UniqueName: \"kubernetes.io/projected/032913e2-a38e-4c15-b11b-5e47cb60c458-kube-api-access-dhhgw\") on node \"crc\" DevicePath \"\"" Nov 22 08:00:03 crc kubenswrapper[4789]: I1122 08:00:03.418555 4789 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/032913e2-a38e-4c15-b11b-5e47cb60c458-config-volume\") on node \"crc\" DevicePath \"\"" Nov 22 08:00:03 crc kubenswrapper[4789]: I1122 08:00:03.418567 4789 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/032913e2-a38e-4c15-b11b-5e47cb60c458-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 22 08:00:03 crc kubenswrapper[4789]: I1122 08:00:03.974517 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29396640-mhslx" event={"ID":"032913e2-a38e-4c15-b11b-5e47cb60c458","Type":"ContainerDied","Data":"820e8a8bd1a43d8843e05cf917d843b4574a9c927cf518d8137935cd9321f470"} Nov 22 08:00:03 crc kubenswrapper[4789]: I1122 08:00:03.974561 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="820e8a8bd1a43d8843e05cf917d843b4574a9c927cf518d8137935cd9321f470" Nov 22 08:00:03 crc kubenswrapper[4789]: I1122 08:00:03.974623 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396640-mhslx" Nov 22 08:00:05 crc kubenswrapper[4789]: I1122 08:00:05.372331 4789 patch_prober.go:28] interesting pod/machine-config-daemon-pxvrm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 08:00:05 crc kubenswrapper[4789]: I1122 08:00:05.372386 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 08:00:35 crc kubenswrapper[4789]: I1122 08:00:35.372668 4789 patch_prober.go:28] interesting pod/machine-config-daemon-pxvrm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 08:00:35 crc kubenswrapper[4789]: I1122 08:00:35.373286 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 08:00:35 crc kubenswrapper[4789]: I1122 08:00:35.373335 4789 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" Nov 22 08:00:35 crc kubenswrapper[4789]: I1122 08:00:35.373934 4789 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8f2ef4eceebb3af5fedfd9bd9b5e59a01bbba909dc1d449d76acf701bd82428a"} pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 22 08:00:35 crc kubenswrapper[4789]: I1122 08:00:35.373988 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" containerID="cri-o://8f2ef4eceebb3af5fedfd9bd9b5e59a01bbba909dc1d449d76acf701bd82428a" gracePeriod=600 Nov 22 08:00:36 crc kubenswrapper[4789]: I1122 08:00:36.146126 4789 generic.go:334] "Generic (PLEG): container finished" podID="58aa8071-7ecc-4692-8789-2db04ec70647" containerID="8f2ef4eceebb3af5fedfd9bd9b5e59a01bbba909dc1d449d76acf701bd82428a" exitCode=0 Nov 22 08:00:36 crc kubenswrapper[4789]: I1122 08:00:36.146212 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" event={"ID":"58aa8071-7ecc-4692-8789-2db04ec70647","Type":"ContainerDied","Data":"8f2ef4eceebb3af5fedfd9bd9b5e59a01bbba909dc1d449d76acf701bd82428a"} Nov 22 08:00:36 crc kubenswrapper[4789]: I1122 08:00:36.146715 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" event={"ID":"58aa8071-7ecc-4692-8789-2db04ec70647","Type":"ContainerStarted","Data":"1297ba3dc6bf83e1607e5d32fcba3c36d955882d84a195c466995f25a0f928f8"} Nov 22 08:00:37 crc kubenswrapper[4789]: I1122 08:00:37.003341 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-gbjc4"] Nov 22 08:00:37 crc kubenswrapper[4789]: I1122 08:00:37.006954 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-gbjc4" podUID="d4a4f530-1afc-402b-8026-584bba4f8f1a" containerName="registry-server" containerID="cri-o://61a0c521d576645357310609f30f970de20cb2843d225885d81fc416c2fa2ea0" gracePeriod=30 Nov 22 08:00:37 crc kubenswrapper[4789]: I1122 08:00:37.012706 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7t8sl"] Nov 22 08:00:37 crc kubenswrapper[4789]: I1122 08:00:37.013060 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-7t8sl" podUID="7426acc3-0cbe-43f6-a327-886c9de10a94" containerName="registry-server" containerID="cri-o://a8182ffc687b14ba917c20bb199af1ad661d56966207b973881b0f5668cbddf3" gracePeriod=30 Nov 22 08:00:37 crc kubenswrapper[4789]: I1122 08:00:37.033499 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-77svl"] Nov 22 08:00:37 crc kubenswrapper[4789]: I1122 08:00:37.033726 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-77svl" podUID="9eb8d5d8-3ad7-41d1-9bb2-2d6865619967" containerName="marketplace-operator" containerID="cri-o://fb23172fe6847c647dc6b91554782ea07d2b1b68ed3374ee83e720a42b9a3f8a" gracePeriod=30 Nov 22 08:00:37 crc kubenswrapper[4789]: I1122 08:00:37.046579 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-2tqfp"] Nov 22 08:00:37 crc kubenswrapper[4789]: E1122 08:00:37.047090 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="032913e2-a38e-4c15-b11b-5e47cb60c458" containerName="collect-profiles" Nov 22 08:00:37 crc kubenswrapper[4789]: I1122 08:00:37.047170 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="032913e2-a38e-4c15-b11b-5e47cb60c458" containerName="collect-profiles" Nov 22 08:00:37 crc kubenswrapper[4789]: I1122 08:00:37.047327 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="032913e2-a38e-4c15-b11b-5e47cb60c458" containerName="collect-profiles" Nov 22 08:00:37 crc kubenswrapper[4789]: I1122 08:00:37.047853 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-2tqfp" Nov 22 08:00:37 crc kubenswrapper[4789]: I1122 08:00:37.056331 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5wgxf"] Nov 22 08:00:37 crc kubenswrapper[4789]: I1122 08:00:37.056627 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-5wgxf" podUID="8235ef4a-0670-4b55-bfe6-4cc3cb1b24de" containerName="registry-server" containerID="cri-o://9ac489479b5b330e2d26f3a51496cab2b8298caf3f196d7ef43c98283d7c272b" gracePeriod=30 Nov 22 08:00:37 crc kubenswrapper[4789]: I1122 08:00:37.061246 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-2tqfp"] Nov 22 08:00:37 crc kubenswrapper[4789]: I1122 08:00:37.066308 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jdz2m"] Nov 22 08:00:37 crc kubenswrapper[4789]: I1122 08:00:37.066560 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-jdz2m" podUID="2bbcd7cf-ccb1-4b91-9e71-2aa87677be19" containerName="registry-server" containerID="cri-o://98103cc7b23e40c25b0d4918ebc1f0b6f2f9d1abd1bac6bd361eb6b477cf1715" gracePeriod=30 Nov 22 08:00:37 crc kubenswrapper[4789]: I1122 08:00:37.234710 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tt4wg\" (UniqueName: \"kubernetes.io/projected/279eae2d-e07b-4a6e-984f-7348f2a7641a-kube-api-access-tt4wg\") pod \"marketplace-operator-79b997595-2tqfp\" (UID: \"279eae2d-e07b-4a6e-984f-7348f2a7641a\") " pod="openshift-marketplace/marketplace-operator-79b997595-2tqfp" Nov 22 08:00:37 crc kubenswrapper[4789]: I1122 08:00:37.234792 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/279eae2d-e07b-4a6e-984f-7348f2a7641a-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-2tqfp\" (UID: \"279eae2d-e07b-4a6e-984f-7348f2a7641a\") " pod="openshift-marketplace/marketplace-operator-79b997595-2tqfp" Nov 22 08:00:37 crc kubenswrapper[4789]: I1122 08:00:37.234911 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/279eae2d-e07b-4a6e-984f-7348f2a7641a-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-2tqfp\" (UID: \"279eae2d-e07b-4a6e-984f-7348f2a7641a\") " pod="openshift-marketplace/marketplace-operator-79b997595-2tqfp" Nov 22 08:00:37 crc kubenswrapper[4789]: I1122 08:00:37.336174 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/279eae2d-e07b-4a6e-984f-7348f2a7641a-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-2tqfp\" (UID: \"279eae2d-e07b-4a6e-984f-7348f2a7641a\") " pod="openshift-marketplace/marketplace-operator-79b997595-2tqfp" Nov 22 08:00:37 crc kubenswrapper[4789]: I1122 08:00:37.336273 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tt4wg\" (UniqueName: \"kubernetes.io/projected/279eae2d-e07b-4a6e-984f-7348f2a7641a-kube-api-access-tt4wg\") pod \"marketplace-operator-79b997595-2tqfp\" (UID: \"279eae2d-e07b-4a6e-984f-7348f2a7641a\") " pod="openshift-marketplace/marketplace-operator-79b997595-2tqfp" Nov 22 08:00:37 crc kubenswrapper[4789]: I1122 08:00:37.336306 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/279eae2d-e07b-4a6e-984f-7348f2a7641a-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-2tqfp\" (UID: \"279eae2d-e07b-4a6e-984f-7348f2a7641a\") " pod="openshift-marketplace/marketplace-operator-79b997595-2tqfp" Nov 22 08:00:37 crc kubenswrapper[4789]: I1122 08:00:37.338044 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/279eae2d-e07b-4a6e-984f-7348f2a7641a-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-2tqfp\" (UID: \"279eae2d-e07b-4a6e-984f-7348f2a7641a\") " pod="openshift-marketplace/marketplace-operator-79b997595-2tqfp" Nov 22 08:00:37 crc kubenswrapper[4789]: I1122 08:00:37.342391 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/279eae2d-e07b-4a6e-984f-7348f2a7641a-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-2tqfp\" (UID: \"279eae2d-e07b-4a6e-984f-7348f2a7641a\") " pod="openshift-marketplace/marketplace-operator-79b997595-2tqfp" Nov 22 08:00:37 crc kubenswrapper[4789]: I1122 08:00:37.358448 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tt4wg\" (UniqueName: \"kubernetes.io/projected/279eae2d-e07b-4a6e-984f-7348f2a7641a-kube-api-access-tt4wg\") pod \"marketplace-operator-79b997595-2tqfp\" (UID: \"279eae2d-e07b-4a6e-984f-7348f2a7641a\") " pod="openshift-marketplace/marketplace-operator-79b997595-2tqfp" Nov 22 08:00:37 crc kubenswrapper[4789]: I1122 08:00:37.370696 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-2tqfp" Nov 22 08:00:37 crc kubenswrapper[4789]: I1122 08:00:37.539796 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-2tqfp"] Nov 22 08:00:37 crc kubenswrapper[4789]: W1122 08:00:37.559201 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod279eae2d_e07b_4a6e_984f_7348f2a7641a.slice/crio-02463b17331076007e4b2791cbadb78b33b77c20828b0b64a4b7c50048edf1c2 WatchSource:0}: Error finding container 02463b17331076007e4b2791cbadb78b33b77c20828b0b64a4b7c50048edf1c2: Status 404 returned error can't find the container with id 02463b17331076007e4b2791cbadb78b33b77c20828b0b64a4b7c50048edf1c2 Nov 22 08:00:37 crc kubenswrapper[4789]: I1122 08:00:37.927961 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7t8sl" Nov 22 08:00:37 crc kubenswrapper[4789]: I1122 08:00:37.975051 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gbjc4" Nov 22 08:00:37 crc kubenswrapper[4789]: I1122 08:00:37.998143 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-77svl" Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.032056 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5wgxf" Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.045381 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-745s6\" (UniqueName: \"kubernetes.io/projected/7426acc3-0cbe-43f6-a327-886c9de10a94-kube-api-access-745s6\") pod \"7426acc3-0cbe-43f6-a327-886c9de10a94\" (UID: \"7426acc3-0cbe-43f6-a327-886c9de10a94\") " Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.045561 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7426acc3-0cbe-43f6-a327-886c9de10a94-utilities\") pod \"7426acc3-0cbe-43f6-a327-886c9de10a94\" (UID: \"7426acc3-0cbe-43f6-a327-886c9de10a94\") " Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.045593 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7426acc3-0cbe-43f6-a327-886c9de10a94-catalog-content\") pod \"7426acc3-0cbe-43f6-a327-886c9de10a94\" (UID: \"7426acc3-0cbe-43f6-a327-886c9de10a94\") " Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.047111 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7426acc3-0cbe-43f6-a327-886c9de10a94-utilities" (OuterVolumeSpecName: "utilities") pod "7426acc3-0cbe-43f6-a327-886c9de10a94" (UID: "7426acc3-0cbe-43f6-a327-886c9de10a94"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.057609 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7426acc3-0cbe-43f6-a327-886c9de10a94-kube-api-access-745s6" (OuterVolumeSpecName: "kube-api-access-745s6") pod "7426acc3-0cbe-43f6-a327-886c9de10a94" (UID: "7426acc3-0cbe-43f6-a327-886c9de10a94"). InnerVolumeSpecName "kube-api-access-745s6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.070021 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jdz2m" Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.112021 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7426acc3-0cbe-43f6-a327-886c9de10a94-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7426acc3-0cbe-43f6-a327-886c9de10a94" (UID: "7426acc3-0cbe-43f6-a327-886c9de10a94"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.149597 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8235ef4a-0670-4b55-bfe6-4cc3cb1b24de-catalog-content\") pod \"8235ef4a-0670-4b55-bfe6-4cc3cb1b24de\" (UID: \"8235ef4a-0670-4b55-bfe6-4cc3cb1b24de\") " Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.150340 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vlv48\" (UniqueName: \"kubernetes.io/projected/8235ef4a-0670-4b55-bfe6-4cc3cb1b24de-kube-api-access-vlv48\") pod \"8235ef4a-0670-4b55-bfe6-4cc3cb1b24de\" (UID: \"8235ef4a-0670-4b55-bfe6-4cc3cb1b24de\") " Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.150385 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9eb8d5d8-3ad7-41d1-9bb2-2d6865619967-marketplace-trusted-ca\") pod \"9eb8d5d8-3ad7-41d1-9bb2-2d6865619967\" (UID: \"9eb8d5d8-3ad7-41d1-9bb2-2d6865619967\") " Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.150402 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4a4f530-1afc-402b-8026-584bba4f8f1a-utilities\") pod \"d4a4f530-1afc-402b-8026-584bba4f8f1a\" (UID: \"d4a4f530-1afc-402b-8026-584bba4f8f1a\") " Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.150436 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/9eb8d5d8-3ad7-41d1-9bb2-2d6865619967-marketplace-operator-metrics\") pod \"9eb8d5d8-3ad7-41d1-9bb2-2d6865619967\" (UID: \"9eb8d5d8-3ad7-41d1-9bb2-2d6865619967\") " Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.150477 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xvst4\" (UniqueName: \"kubernetes.io/projected/9eb8d5d8-3ad7-41d1-9bb2-2d6865619967-kube-api-access-xvst4\") pod \"9eb8d5d8-3ad7-41d1-9bb2-2d6865619967\" (UID: \"9eb8d5d8-3ad7-41d1-9bb2-2d6865619967\") " Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.150538 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8235ef4a-0670-4b55-bfe6-4cc3cb1b24de-utilities\") pod \"8235ef4a-0670-4b55-bfe6-4cc3cb1b24de\" (UID: \"8235ef4a-0670-4b55-bfe6-4cc3cb1b24de\") " Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.150568 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pf4qs\" (UniqueName: \"kubernetes.io/projected/d4a4f530-1afc-402b-8026-584bba4f8f1a-kube-api-access-pf4qs\") pod \"d4a4f530-1afc-402b-8026-584bba4f8f1a\" (UID: \"d4a4f530-1afc-402b-8026-584bba4f8f1a\") " Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.150612 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4a4f530-1afc-402b-8026-584bba4f8f1a-catalog-content\") pod \"d4a4f530-1afc-402b-8026-584bba4f8f1a\" (UID: \"d4a4f530-1afc-402b-8026-584bba4f8f1a\") " Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.150940 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-745s6\" (UniqueName: \"kubernetes.io/projected/7426acc3-0cbe-43f6-a327-886c9de10a94-kube-api-access-745s6\") on node \"crc\" DevicePath \"\"" Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.150957 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7426acc3-0cbe-43f6-a327-886c9de10a94-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.150967 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7426acc3-0cbe-43f6-a327-886c9de10a94-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.151658 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8235ef4a-0670-4b55-bfe6-4cc3cb1b24de-utilities" (OuterVolumeSpecName: "utilities") pod "8235ef4a-0670-4b55-bfe6-4cc3cb1b24de" (UID: "8235ef4a-0670-4b55-bfe6-4cc3cb1b24de"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.151881 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9eb8d5d8-3ad7-41d1-9bb2-2d6865619967-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "9eb8d5d8-3ad7-41d1-9bb2-2d6865619967" (UID: "9eb8d5d8-3ad7-41d1-9bb2-2d6865619967"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.152798 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d4a4f530-1afc-402b-8026-584bba4f8f1a-utilities" (OuterVolumeSpecName: "utilities") pod "d4a4f530-1afc-402b-8026-584bba4f8f1a" (UID: "d4a4f530-1afc-402b-8026-584bba4f8f1a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.154133 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4a4f530-1afc-402b-8026-584bba4f8f1a-kube-api-access-pf4qs" (OuterVolumeSpecName: "kube-api-access-pf4qs") pod "d4a4f530-1afc-402b-8026-584bba4f8f1a" (UID: "d4a4f530-1afc-402b-8026-584bba4f8f1a"). InnerVolumeSpecName "kube-api-access-pf4qs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.154375 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9eb8d5d8-3ad7-41d1-9bb2-2d6865619967-kube-api-access-xvst4" (OuterVolumeSpecName: "kube-api-access-xvst4") pod "9eb8d5d8-3ad7-41d1-9bb2-2d6865619967" (UID: "9eb8d5d8-3ad7-41d1-9bb2-2d6865619967"). InnerVolumeSpecName "kube-api-access-xvst4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.156040 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8235ef4a-0670-4b55-bfe6-4cc3cb1b24de-kube-api-access-vlv48" (OuterVolumeSpecName: "kube-api-access-vlv48") pod "8235ef4a-0670-4b55-bfe6-4cc3cb1b24de" (UID: "8235ef4a-0670-4b55-bfe6-4cc3cb1b24de"). InnerVolumeSpecName "kube-api-access-vlv48". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.156501 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9eb8d5d8-3ad7-41d1-9bb2-2d6865619967-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "9eb8d5d8-3ad7-41d1-9bb2-2d6865619967" (UID: "9eb8d5d8-3ad7-41d1-9bb2-2d6865619967"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.160608 4789 generic.go:334] "Generic (PLEG): container finished" podID="2bbcd7cf-ccb1-4b91-9e71-2aa87677be19" containerID="98103cc7b23e40c25b0d4918ebc1f0b6f2f9d1abd1bac6bd361eb6b477cf1715" exitCode=0 Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.160655 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jdz2m" Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.160692 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jdz2m" event={"ID":"2bbcd7cf-ccb1-4b91-9e71-2aa87677be19","Type":"ContainerDied","Data":"98103cc7b23e40c25b0d4918ebc1f0b6f2f9d1abd1bac6bd361eb6b477cf1715"} Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.160721 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jdz2m" event={"ID":"2bbcd7cf-ccb1-4b91-9e71-2aa87677be19","Type":"ContainerDied","Data":"2f5b97af05a9223958e3057a9f1b53f9d711571f91f0d1669d11c99f24c9e9d5"} Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.160740 4789 scope.go:117] "RemoveContainer" containerID="98103cc7b23e40c25b0d4918ebc1f0b6f2f9d1abd1bac6bd361eb6b477cf1715" Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.164302 4789 generic.go:334] "Generic (PLEG): container finished" podID="9eb8d5d8-3ad7-41d1-9bb2-2d6865619967" containerID="fb23172fe6847c647dc6b91554782ea07d2b1b68ed3374ee83e720a42b9a3f8a" exitCode=0 Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.164347 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-77svl" Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.164409 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-77svl" event={"ID":"9eb8d5d8-3ad7-41d1-9bb2-2d6865619967","Type":"ContainerDied","Data":"fb23172fe6847c647dc6b91554782ea07d2b1b68ed3374ee83e720a42b9a3f8a"} Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.164445 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-77svl" event={"ID":"9eb8d5d8-3ad7-41d1-9bb2-2d6865619967","Type":"ContainerDied","Data":"1db435158b0366b3afe5909488c9c5c268a79c71daf279844c5c6598b7a1dc7c"} Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.166296 4789 generic.go:334] "Generic (PLEG): container finished" podID="d4a4f530-1afc-402b-8026-584bba4f8f1a" containerID="61a0c521d576645357310609f30f970de20cb2843d225885d81fc416c2fa2ea0" exitCode=0 Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.166375 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gbjc4" event={"ID":"d4a4f530-1afc-402b-8026-584bba4f8f1a","Type":"ContainerDied","Data":"61a0c521d576645357310609f30f970de20cb2843d225885d81fc416c2fa2ea0"} Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.166402 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gbjc4" event={"ID":"d4a4f530-1afc-402b-8026-584bba4f8f1a","Type":"ContainerDied","Data":"d8a0191b9e09a1bbbe54a395bc17206a8c48b10df45cb728e0da0366f87ec72c"} Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.166436 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gbjc4" Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.168905 4789 generic.go:334] "Generic (PLEG): container finished" podID="8235ef4a-0670-4b55-bfe6-4cc3cb1b24de" containerID="9ac489479b5b330e2d26f3a51496cab2b8298caf3f196d7ef43c98283d7c272b" exitCode=0 Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.168938 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5wgxf" event={"ID":"8235ef4a-0670-4b55-bfe6-4cc3cb1b24de","Type":"ContainerDied","Data":"9ac489479b5b330e2d26f3a51496cab2b8298caf3f196d7ef43c98283d7c272b"} Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.168976 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5wgxf" event={"ID":"8235ef4a-0670-4b55-bfe6-4cc3cb1b24de","Type":"ContainerDied","Data":"51126406b32d5dfb23c27551f52a44b5452197883de71ef092d207dad0aa8053"} Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.169038 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5wgxf" Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.170819 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-2tqfp" event={"ID":"279eae2d-e07b-4a6e-984f-7348f2a7641a","Type":"ContainerStarted","Data":"1dbc81b3356193957d615b5ff390e1a73ddb4185205551f4794b44d844ca665b"} Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.170860 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-2tqfp" event={"ID":"279eae2d-e07b-4a6e-984f-7348f2a7641a","Type":"ContainerStarted","Data":"02463b17331076007e4b2791cbadb78b33b77c20828b0b64a4b7c50048edf1c2"} Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.171043 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-2tqfp" Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.171996 4789 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-2tqfp container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.55:8080/healthz\": dial tcp 10.217.0.55:8080: connect: connection refused" start-of-body= Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.172042 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-2tqfp" podUID="279eae2d-e07b-4a6e-984f-7348f2a7641a" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.55:8080/healthz\": dial tcp 10.217.0.55:8080: connect: connection refused" Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.172610 4789 generic.go:334] "Generic (PLEG): container finished" podID="7426acc3-0cbe-43f6-a327-886c9de10a94" containerID="a8182ffc687b14ba917c20bb199af1ad661d56966207b973881b0f5668cbddf3" exitCode=0 Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.172639 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7t8sl" event={"ID":"7426acc3-0cbe-43f6-a327-886c9de10a94","Type":"ContainerDied","Data":"a8182ffc687b14ba917c20bb199af1ad661d56966207b973881b0f5668cbddf3"} Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.172654 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7t8sl" event={"ID":"7426acc3-0cbe-43f6-a327-886c9de10a94","Type":"ContainerDied","Data":"c0ae897dc3391c4856d5ad1d55fe30b19df432432d8e7eef5135d1cf4e1f7504"} Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.172875 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7t8sl" Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.181672 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8235ef4a-0670-4b55-bfe6-4cc3cb1b24de-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8235ef4a-0670-4b55-bfe6-4cc3cb1b24de" (UID: "8235ef4a-0670-4b55-bfe6-4cc3cb1b24de"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.188221 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-2tqfp" podStartSLOduration=1.188194641 podStartE2EDuration="1.188194641s" podCreationTimestamp="2025-11-22 08:00:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:00:38.185967369 +0000 UTC m=+252.420367642" watchObservedRunningTime="2025-11-22 08:00:38.188194641 +0000 UTC m=+252.422594914" Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.223298 4789 scope.go:117] "RemoveContainer" containerID="9ee1547395e4e03e0ee50d99b4764acdf0a3df06d3f8b323334099913b9c8224" Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.232429 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-77svl"] Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.235245 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-77svl"] Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.240111 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d4a4f530-1afc-402b-8026-584bba4f8f1a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d4a4f530-1afc-402b-8026-584bba4f8f1a" (UID: "d4a4f530-1afc-402b-8026-584bba4f8f1a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.244672 4789 scope.go:117] "RemoveContainer" containerID="210d915204b25ed0acd603a467395328ce394aa17e753337ad9d6f5889e0bcde" Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.247004 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7t8sl"] Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.252486 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2bbcd7cf-ccb1-4b91-9e71-2aa87677be19-catalog-content\") pod \"2bbcd7cf-ccb1-4b91-9e71-2aa87677be19\" (UID: \"2bbcd7cf-ccb1-4b91-9e71-2aa87677be19\") " Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.252549 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vs8fh\" (UniqueName: \"kubernetes.io/projected/2bbcd7cf-ccb1-4b91-9e71-2aa87677be19-kube-api-access-vs8fh\") pod \"2bbcd7cf-ccb1-4b91-9e71-2aa87677be19\" (UID: \"2bbcd7cf-ccb1-4b91-9e71-2aa87677be19\") " Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.252631 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2bbcd7cf-ccb1-4b91-9e71-2aa87677be19-utilities\") pod \"2bbcd7cf-ccb1-4b91-9e71-2aa87677be19\" (UID: \"2bbcd7cf-ccb1-4b91-9e71-2aa87677be19\") " Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.252878 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vlv48\" (UniqueName: \"kubernetes.io/projected/8235ef4a-0670-4b55-bfe6-4cc3cb1b24de-kube-api-access-vlv48\") on node \"crc\" DevicePath \"\"" Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.252898 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4a4f530-1afc-402b-8026-584bba4f8f1a-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.252907 4789 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9eb8d5d8-3ad7-41d1-9bb2-2d6865619967-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.252915 4789 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/9eb8d5d8-3ad7-41d1-9bb2-2d6865619967-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.252924 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xvst4\" (UniqueName: \"kubernetes.io/projected/9eb8d5d8-3ad7-41d1-9bb2-2d6865619967-kube-api-access-xvst4\") on node \"crc\" DevicePath \"\"" Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.252933 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8235ef4a-0670-4b55-bfe6-4cc3cb1b24de-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.252941 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pf4qs\" (UniqueName: \"kubernetes.io/projected/d4a4f530-1afc-402b-8026-584bba4f8f1a-kube-api-access-pf4qs\") on node \"crc\" DevicePath \"\"" Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.252949 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4a4f530-1afc-402b-8026-584bba4f8f1a-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.252959 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8235ef4a-0670-4b55-bfe6-4cc3cb1b24de-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.253570 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2bbcd7cf-ccb1-4b91-9e71-2aa87677be19-utilities" (OuterVolumeSpecName: "utilities") pod "2bbcd7cf-ccb1-4b91-9e71-2aa87677be19" (UID: "2bbcd7cf-ccb1-4b91-9e71-2aa87677be19"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.255163 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-7t8sl"] Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.255836 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2bbcd7cf-ccb1-4b91-9e71-2aa87677be19-kube-api-access-vs8fh" (OuterVolumeSpecName: "kube-api-access-vs8fh") pod "2bbcd7cf-ccb1-4b91-9e71-2aa87677be19" (UID: "2bbcd7cf-ccb1-4b91-9e71-2aa87677be19"). InnerVolumeSpecName "kube-api-access-vs8fh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.306172 4789 scope.go:117] "RemoveContainer" containerID="98103cc7b23e40c25b0d4918ebc1f0b6f2f9d1abd1bac6bd361eb6b477cf1715" Nov 22 08:00:38 crc kubenswrapper[4789]: E1122 08:00:38.307967 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"98103cc7b23e40c25b0d4918ebc1f0b6f2f9d1abd1bac6bd361eb6b477cf1715\": container with ID starting with 98103cc7b23e40c25b0d4918ebc1f0b6f2f9d1abd1bac6bd361eb6b477cf1715 not found: ID does not exist" containerID="98103cc7b23e40c25b0d4918ebc1f0b6f2f9d1abd1bac6bd361eb6b477cf1715" Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.308012 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"98103cc7b23e40c25b0d4918ebc1f0b6f2f9d1abd1bac6bd361eb6b477cf1715"} err="failed to get container status \"98103cc7b23e40c25b0d4918ebc1f0b6f2f9d1abd1bac6bd361eb6b477cf1715\": rpc error: code = NotFound desc = could not find container \"98103cc7b23e40c25b0d4918ebc1f0b6f2f9d1abd1bac6bd361eb6b477cf1715\": container with ID starting with 98103cc7b23e40c25b0d4918ebc1f0b6f2f9d1abd1bac6bd361eb6b477cf1715 not found: ID does not exist" Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.308044 4789 scope.go:117] "RemoveContainer" containerID="9ee1547395e4e03e0ee50d99b4764acdf0a3df06d3f8b323334099913b9c8224" Nov 22 08:00:38 crc kubenswrapper[4789]: E1122 08:00:38.308391 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9ee1547395e4e03e0ee50d99b4764acdf0a3df06d3f8b323334099913b9c8224\": container with ID starting with 9ee1547395e4e03e0ee50d99b4764acdf0a3df06d3f8b323334099913b9c8224 not found: ID does not exist" containerID="9ee1547395e4e03e0ee50d99b4764acdf0a3df06d3f8b323334099913b9c8224" Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.308431 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ee1547395e4e03e0ee50d99b4764acdf0a3df06d3f8b323334099913b9c8224"} err="failed to get container status \"9ee1547395e4e03e0ee50d99b4764acdf0a3df06d3f8b323334099913b9c8224\": rpc error: code = NotFound desc = could not find container \"9ee1547395e4e03e0ee50d99b4764acdf0a3df06d3f8b323334099913b9c8224\": container with ID starting with 9ee1547395e4e03e0ee50d99b4764acdf0a3df06d3f8b323334099913b9c8224 not found: ID does not exist" Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.308459 4789 scope.go:117] "RemoveContainer" containerID="210d915204b25ed0acd603a467395328ce394aa17e753337ad9d6f5889e0bcde" Nov 22 08:00:38 crc kubenswrapper[4789]: E1122 08:00:38.308912 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"210d915204b25ed0acd603a467395328ce394aa17e753337ad9d6f5889e0bcde\": container with ID starting with 210d915204b25ed0acd603a467395328ce394aa17e753337ad9d6f5889e0bcde not found: ID does not exist" containerID="210d915204b25ed0acd603a467395328ce394aa17e753337ad9d6f5889e0bcde" Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.308942 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"210d915204b25ed0acd603a467395328ce394aa17e753337ad9d6f5889e0bcde"} err="failed to get container status \"210d915204b25ed0acd603a467395328ce394aa17e753337ad9d6f5889e0bcde\": rpc error: code = NotFound desc = could not find container \"210d915204b25ed0acd603a467395328ce394aa17e753337ad9d6f5889e0bcde\": container with ID starting with 210d915204b25ed0acd603a467395328ce394aa17e753337ad9d6f5889e0bcde not found: ID does not exist" Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.308961 4789 scope.go:117] "RemoveContainer" containerID="fb23172fe6847c647dc6b91554782ea07d2b1b68ed3374ee83e720a42b9a3f8a" Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.326736 4789 scope.go:117] "RemoveContainer" containerID="fb23172fe6847c647dc6b91554782ea07d2b1b68ed3374ee83e720a42b9a3f8a" Nov 22 08:00:38 crc kubenswrapper[4789]: E1122 08:00:38.328046 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fb23172fe6847c647dc6b91554782ea07d2b1b68ed3374ee83e720a42b9a3f8a\": container with ID starting with fb23172fe6847c647dc6b91554782ea07d2b1b68ed3374ee83e720a42b9a3f8a not found: ID does not exist" containerID="fb23172fe6847c647dc6b91554782ea07d2b1b68ed3374ee83e720a42b9a3f8a" Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.328095 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fb23172fe6847c647dc6b91554782ea07d2b1b68ed3374ee83e720a42b9a3f8a"} err="failed to get container status \"fb23172fe6847c647dc6b91554782ea07d2b1b68ed3374ee83e720a42b9a3f8a\": rpc error: code = NotFound desc = could not find container \"fb23172fe6847c647dc6b91554782ea07d2b1b68ed3374ee83e720a42b9a3f8a\": container with ID starting with fb23172fe6847c647dc6b91554782ea07d2b1b68ed3374ee83e720a42b9a3f8a not found: ID does not exist" Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.328131 4789 scope.go:117] "RemoveContainer" containerID="61a0c521d576645357310609f30f970de20cb2843d225885d81fc416c2fa2ea0" Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.340624 4789 scope.go:117] "RemoveContainer" containerID="76501fa438782b1c9700d1539ccbad88ac50a96f05bc101608894bf1711b6f9d" Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.344575 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2bbcd7cf-ccb1-4b91-9e71-2aa87677be19-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2bbcd7cf-ccb1-4b91-9e71-2aa87677be19" (UID: "2bbcd7cf-ccb1-4b91-9e71-2aa87677be19"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.353629 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2bbcd7cf-ccb1-4b91-9e71-2aa87677be19-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.353887 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vs8fh\" (UniqueName: \"kubernetes.io/projected/2bbcd7cf-ccb1-4b91-9e71-2aa87677be19-kube-api-access-vs8fh\") on node \"crc\" DevicePath \"\"" Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.353928 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2bbcd7cf-ccb1-4b91-9e71-2aa87677be19-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.358115 4789 scope.go:117] "RemoveContainer" containerID="13fa65cbf77f7199881e2a4fa541ef485009f272481faf8e2f0f5f882f228268" Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.370246 4789 scope.go:117] "RemoveContainer" containerID="61a0c521d576645357310609f30f970de20cb2843d225885d81fc416c2fa2ea0" Nov 22 08:00:38 crc kubenswrapper[4789]: E1122 08:00:38.370603 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"61a0c521d576645357310609f30f970de20cb2843d225885d81fc416c2fa2ea0\": container with ID starting with 61a0c521d576645357310609f30f970de20cb2843d225885d81fc416c2fa2ea0 not found: ID does not exist" containerID="61a0c521d576645357310609f30f970de20cb2843d225885d81fc416c2fa2ea0" Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.370632 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61a0c521d576645357310609f30f970de20cb2843d225885d81fc416c2fa2ea0"} err="failed to get container status \"61a0c521d576645357310609f30f970de20cb2843d225885d81fc416c2fa2ea0\": rpc error: code = NotFound desc = could not find container \"61a0c521d576645357310609f30f970de20cb2843d225885d81fc416c2fa2ea0\": container with ID starting with 61a0c521d576645357310609f30f970de20cb2843d225885d81fc416c2fa2ea0 not found: ID does not exist" Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.370654 4789 scope.go:117] "RemoveContainer" containerID="76501fa438782b1c9700d1539ccbad88ac50a96f05bc101608894bf1711b6f9d" Nov 22 08:00:38 crc kubenswrapper[4789]: E1122 08:00:38.370913 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"76501fa438782b1c9700d1539ccbad88ac50a96f05bc101608894bf1711b6f9d\": container with ID starting with 76501fa438782b1c9700d1539ccbad88ac50a96f05bc101608894bf1711b6f9d not found: ID does not exist" containerID="76501fa438782b1c9700d1539ccbad88ac50a96f05bc101608894bf1711b6f9d" Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.370935 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"76501fa438782b1c9700d1539ccbad88ac50a96f05bc101608894bf1711b6f9d"} err="failed to get container status \"76501fa438782b1c9700d1539ccbad88ac50a96f05bc101608894bf1711b6f9d\": rpc error: code = NotFound desc = could not find container \"76501fa438782b1c9700d1539ccbad88ac50a96f05bc101608894bf1711b6f9d\": container with ID starting with 76501fa438782b1c9700d1539ccbad88ac50a96f05bc101608894bf1711b6f9d not found: ID does not exist" Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.370947 4789 scope.go:117] "RemoveContainer" containerID="13fa65cbf77f7199881e2a4fa541ef485009f272481faf8e2f0f5f882f228268" Nov 22 08:00:38 crc kubenswrapper[4789]: E1122 08:00:38.371245 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"13fa65cbf77f7199881e2a4fa541ef485009f272481faf8e2f0f5f882f228268\": container with ID starting with 13fa65cbf77f7199881e2a4fa541ef485009f272481faf8e2f0f5f882f228268 not found: ID does not exist" containerID="13fa65cbf77f7199881e2a4fa541ef485009f272481faf8e2f0f5f882f228268" Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.371282 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"13fa65cbf77f7199881e2a4fa541ef485009f272481faf8e2f0f5f882f228268"} err="failed to get container status \"13fa65cbf77f7199881e2a4fa541ef485009f272481faf8e2f0f5f882f228268\": rpc error: code = NotFound desc = could not find container \"13fa65cbf77f7199881e2a4fa541ef485009f272481faf8e2f0f5f882f228268\": container with ID starting with 13fa65cbf77f7199881e2a4fa541ef485009f272481faf8e2f0f5f882f228268 not found: ID does not exist" Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.371305 4789 scope.go:117] "RemoveContainer" containerID="9ac489479b5b330e2d26f3a51496cab2b8298caf3f196d7ef43c98283d7c272b" Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.384015 4789 scope.go:117] "RemoveContainer" containerID="6b1f2ba6174a685908041aeae46a98db241deac5d93d783eabcf5df90333b251" Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.398324 4789 scope.go:117] "RemoveContainer" containerID="63766f140e768be82aa548a637ff889922d5555ed2edbee551dcc16cf013a7ba" Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.416206 4789 scope.go:117] "RemoveContainer" containerID="9ac489479b5b330e2d26f3a51496cab2b8298caf3f196d7ef43c98283d7c272b" Nov 22 08:00:38 crc kubenswrapper[4789]: E1122 08:00:38.416654 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9ac489479b5b330e2d26f3a51496cab2b8298caf3f196d7ef43c98283d7c272b\": container with ID starting with 9ac489479b5b330e2d26f3a51496cab2b8298caf3f196d7ef43c98283d7c272b not found: ID does not exist" containerID="9ac489479b5b330e2d26f3a51496cab2b8298caf3f196d7ef43c98283d7c272b" Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.416682 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ac489479b5b330e2d26f3a51496cab2b8298caf3f196d7ef43c98283d7c272b"} err="failed to get container status \"9ac489479b5b330e2d26f3a51496cab2b8298caf3f196d7ef43c98283d7c272b\": rpc error: code = NotFound desc = could not find container \"9ac489479b5b330e2d26f3a51496cab2b8298caf3f196d7ef43c98283d7c272b\": container with ID starting with 9ac489479b5b330e2d26f3a51496cab2b8298caf3f196d7ef43c98283d7c272b not found: ID does not exist" Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.416712 4789 scope.go:117] "RemoveContainer" containerID="6b1f2ba6174a685908041aeae46a98db241deac5d93d783eabcf5df90333b251" Nov 22 08:00:38 crc kubenswrapper[4789]: E1122 08:00:38.417143 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b1f2ba6174a685908041aeae46a98db241deac5d93d783eabcf5df90333b251\": container with ID starting with 6b1f2ba6174a685908041aeae46a98db241deac5d93d783eabcf5df90333b251 not found: ID does not exist" containerID="6b1f2ba6174a685908041aeae46a98db241deac5d93d783eabcf5df90333b251" Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.417164 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b1f2ba6174a685908041aeae46a98db241deac5d93d783eabcf5df90333b251"} err="failed to get container status \"6b1f2ba6174a685908041aeae46a98db241deac5d93d783eabcf5df90333b251\": rpc error: code = NotFound desc = could not find container \"6b1f2ba6174a685908041aeae46a98db241deac5d93d783eabcf5df90333b251\": container with ID starting with 6b1f2ba6174a685908041aeae46a98db241deac5d93d783eabcf5df90333b251 not found: ID does not exist" Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.417177 4789 scope.go:117] "RemoveContainer" containerID="63766f140e768be82aa548a637ff889922d5555ed2edbee551dcc16cf013a7ba" Nov 22 08:00:38 crc kubenswrapper[4789]: E1122 08:00:38.417422 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"63766f140e768be82aa548a637ff889922d5555ed2edbee551dcc16cf013a7ba\": container with ID starting with 63766f140e768be82aa548a637ff889922d5555ed2edbee551dcc16cf013a7ba not found: ID does not exist" containerID="63766f140e768be82aa548a637ff889922d5555ed2edbee551dcc16cf013a7ba" Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.417443 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"63766f140e768be82aa548a637ff889922d5555ed2edbee551dcc16cf013a7ba"} err="failed to get container status \"63766f140e768be82aa548a637ff889922d5555ed2edbee551dcc16cf013a7ba\": rpc error: code = NotFound desc = could not find container \"63766f140e768be82aa548a637ff889922d5555ed2edbee551dcc16cf013a7ba\": container with ID starting with 63766f140e768be82aa548a637ff889922d5555ed2edbee551dcc16cf013a7ba not found: ID does not exist" Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.417466 4789 scope.go:117] "RemoveContainer" containerID="a8182ffc687b14ba917c20bb199af1ad661d56966207b973881b0f5668cbddf3" Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.431527 4789 scope.go:117] "RemoveContainer" containerID="f95d16bef21f99bf0cd2f9e7b4cf410f5d2751c9c8328eeb1981f21907a7e1bf" Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.443862 4789 scope.go:117] "RemoveContainer" containerID="3db8b3ea616e93792b38156af198afd4a8b8c1d8ce0cc030b03c35ad16d449c8" Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.457167 4789 scope.go:117] "RemoveContainer" containerID="a8182ffc687b14ba917c20bb199af1ad661d56966207b973881b0f5668cbddf3" Nov 22 08:00:38 crc kubenswrapper[4789]: E1122 08:00:38.457730 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a8182ffc687b14ba917c20bb199af1ad661d56966207b973881b0f5668cbddf3\": container with ID starting with a8182ffc687b14ba917c20bb199af1ad661d56966207b973881b0f5668cbddf3 not found: ID does not exist" containerID="a8182ffc687b14ba917c20bb199af1ad661d56966207b973881b0f5668cbddf3" Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.457793 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8182ffc687b14ba917c20bb199af1ad661d56966207b973881b0f5668cbddf3"} err="failed to get container status \"a8182ffc687b14ba917c20bb199af1ad661d56966207b973881b0f5668cbddf3\": rpc error: code = NotFound desc = could not find container \"a8182ffc687b14ba917c20bb199af1ad661d56966207b973881b0f5668cbddf3\": container with ID starting with a8182ffc687b14ba917c20bb199af1ad661d56966207b973881b0f5668cbddf3 not found: ID does not exist" Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.457825 4789 scope.go:117] "RemoveContainer" containerID="f95d16bef21f99bf0cd2f9e7b4cf410f5d2751c9c8328eeb1981f21907a7e1bf" Nov 22 08:00:38 crc kubenswrapper[4789]: E1122 08:00:38.458262 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f95d16bef21f99bf0cd2f9e7b4cf410f5d2751c9c8328eeb1981f21907a7e1bf\": container with ID starting with f95d16bef21f99bf0cd2f9e7b4cf410f5d2751c9c8328eeb1981f21907a7e1bf not found: ID does not exist" containerID="f95d16bef21f99bf0cd2f9e7b4cf410f5d2751c9c8328eeb1981f21907a7e1bf" Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.458291 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f95d16bef21f99bf0cd2f9e7b4cf410f5d2751c9c8328eeb1981f21907a7e1bf"} err="failed to get container status \"f95d16bef21f99bf0cd2f9e7b4cf410f5d2751c9c8328eeb1981f21907a7e1bf\": rpc error: code = NotFound desc = could not find container \"f95d16bef21f99bf0cd2f9e7b4cf410f5d2751c9c8328eeb1981f21907a7e1bf\": container with ID starting with f95d16bef21f99bf0cd2f9e7b4cf410f5d2751c9c8328eeb1981f21907a7e1bf not found: ID does not exist" Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.458314 4789 scope.go:117] "RemoveContainer" containerID="3db8b3ea616e93792b38156af198afd4a8b8c1d8ce0cc030b03c35ad16d449c8" Nov 22 08:00:38 crc kubenswrapper[4789]: E1122 08:00:38.458610 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3db8b3ea616e93792b38156af198afd4a8b8c1d8ce0cc030b03c35ad16d449c8\": container with ID starting with 3db8b3ea616e93792b38156af198afd4a8b8c1d8ce0cc030b03c35ad16d449c8 not found: ID does not exist" containerID="3db8b3ea616e93792b38156af198afd4a8b8c1d8ce0cc030b03c35ad16d449c8" Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.458803 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3db8b3ea616e93792b38156af198afd4a8b8c1d8ce0cc030b03c35ad16d449c8"} err="failed to get container status \"3db8b3ea616e93792b38156af198afd4a8b8c1d8ce0cc030b03c35ad16d449c8\": rpc error: code = NotFound desc = could not find container \"3db8b3ea616e93792b38156af198afd4a8b8c1d8ce0cc030b03c35ad16d449c8\": container with ID starting with 3db8b3ea616e93792b38156af198afd4a8b8c1d8ce0cc030b03c35ad16d449c8 not found: ID does not exist" Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.493176 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jdz2m"] Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.502391 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-jdz2m"] Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.517885 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-gbjc4"] Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.534959 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-gbjc4"] Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.537343 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5wgxf"] Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.539927 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-5wgxf"] Nov 22 08:00:38 crc kubenswrapper[4789]: I1122 08:00:38.725566 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-wnfzj"] Nov 22 08:00:39 crc kubenswrapper[4789]: I1122 08:00:39.019738 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-442k7"] Nov 22 08:00:39 crc kubenswrapper[4789]: E1122 08:00:39.019986 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4a4f530-1afc-402b-8026-584bba4f8f1a" containerName="extract-content" Nov 22 08:00:39 crc kubenswrapper[4789]: I1122 08:00:39.020002 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4a4f530-1afc-402b-8026-584bba4f8f1a" containerName="extract-content" Nov 22 08:00:39 crc kubenswrapper[4789]: E1122 08:00:39.020015 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7426acc3-0cbe-43f6-a327-886c9de10a94" containerName="extract-content" Nov 22 08:00:39 crc kubenswrapper[4789]: I1122 08:00:39.020023 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="7426acc3-0cbe-43f6-a327-886c9de10a94" containerName="extract-content" Nov 22 08:00:39 crc kubenswrapper[4789]: E1122 08:00:39.020037 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7426acc3-0cbe-43f6-a327-886c9de10a94" containerName="extract-utilities" Nov 22 08:00:39 crc kubenswrapper[4789]: I1122 08:00:39.020048 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="7426acc3-0cbe-43f6-a327-886c9de10a94" containerName="extract-utilities" Nov 22 08:00:39 crc kubenswrapper[4789]: E1122 08:00:39.020060 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9eb8d5d8-3ad7-41d1-9bb2-2d6865619967" containerName="marketplace-operator" Nov 22 08:00:39 crc kubenswrapper[4789]: I1122 08:00:39.020067 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="9eb8d5d8-3ad7-41d1-9bb2-2d6865619967" containerName="marketplace-operator" Nov 22 08:00:39 crc kubenswrapper[4789]: E1122 08:00:39.020078 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8235ef4a-0670-4b55-bfe6-4cc3cb1b24de" containerName="registry-server" Nov 22 08:00:39 crc kubenswrapper[4789]: I1122 08:00:39.020085 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="8235ef4a-0670-4b55-bfe6-4cc3cb1b24de" containerName="registry-server" Nov 22 08:00:39 crc kubenswrapper[4789]: E1122 08:00:39.020098 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8235ef4a-0670-4b55-bfe6-4cc3cb1b24de" containerName="extract-utilities" Nov 22 08:00:39 crc kubenswrapper[4789]: I1122 08:00:39.020106 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="8235ef4a-0670-4b55-bfe6-4cc3cb1b24de" containerName="extract-utilities" Nov 22 08:00:39 crc kubenswrapper[4789]: E1122 08:00:39.020116 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bbcd7cf-ccb1-4b91-9e71-2aa87677be19" containerName="extract-content" Nov 22 08:00:39 crc kubenswrapper[4789]: I1122 08:00:39.020123 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bbcd7cf-ccb1-4b91-9e71-2aa87677be19" containerName="extract-content" Nov 22 08:00:39 crc kubenswrapper[4789]: E1122 08:00:39.020132 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8235ef4a-0670-4b55-bfe6-4cc3cb1b24de" containerName="extract-content" Nov 22 08:00:39 crc kubenswrapper[4789]: I1122 08:00:39.020140 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="8235ef4a-0670-4b55-bfe6-4cc3cb1b24de" containerName="extract-content" Nov 22 08:00:39 crc kubenswrapper[4789]: E1122 08:00:39.020151 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7426acc3-0cbe-43f6-a327-886c9de10a94" containerName="registry-server" Nov 22 08:00:39 crc kubenswrapper[4789]: I1122 08:00:39.020158 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="7426acc3-0cbe-43f6-a327-886c9de10a94" containerName="registry-server" Nov 22 08:00:39 crc kubenswrapper[4789]: E1122 08:00:39.020168 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4a4f530-1afc-402b-8026-584bba4f8f1a" containerName="registry-server" Nov 22 08:00:39 crc kubenswrapper[4789]: I1122 08:00:39.020175 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4a4f530-1afc-402b-8026-584bba4f8f1a" containerName="registry-server" Nov 22 08:00:39 crc kubenswrapper[4789]: E1122 08:00:39.020184 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4a4f530-1afc-402b-8026-584bba4f8f1a" containerName="extract-utilities" Nov 22 08:00:39 crc kubenswrapper[4789]: I1122 08:00:39.020193 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4a4f530-1afc-402b-8026-584bba4f8f1a" containerName="extract-utilities" Nov 22 08:00:39 crc kubenswrapper[4789]: E1122 08:00:39.020203 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bbcd7cf-ccb1-4b91-9e71-2aa87677be19" containerName="extract-utilities" Nov 22 08:00:39 crc kubenswrapper[4789]: I1122 08:00:39.020211 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bbcd7cf-ccb1-4b91-9e71-2aa87677be19" containerName="extract-utilities" Nov 22 08:00:39 crc kubenswrapper[4789]: E1122 08:00:39.020226 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bbcd7cf-ccb1-4b91-9e71-2aa87677be19" containerName="registry-server" Nov 22 08:00:39 crc kubenswrapper[4789]: I1122 08:00:39.020234 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bbcd7cf-ccb1-4b91-9e71-2aa87677be19" containerName="registry-server" Nov 22 08:00:39 crc kubenswrapper[4789]: I1122 08:00:39.020355 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="8235ef4a-0670-4b55-bfe6-4cc3cb1b24de" containerName="registry-server" Nov 22 08:00:39 crc kubenswrapper[4789]: I1122 08:00:39.020376 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4a4f530-1afc-402b-8026-584bba4f8f1a" containerName="registry-server" Nov 22 08:00:39 crc kubenswrapper[4789]: I1122 08:00:39.020386 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="2bbcd7cf-ccb1-4b91-9e71-2aa87677be19" containerName="registry-server" Nov 22 08:00:39 crc kubenswrapper[4789]: I1122 08:00:39.020401 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="7426acc3-0cbe-43f6-a327-886c9de10a94" containerName="registry-server" Nov 22 08:00:39 crc kubenswrapper[4789]: I1122 08:00:39.020411 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="9eb8d5d8-3ad7-41d1-9bb2-2d6865619967" containerName="marketplace-operator" Nov 22 08:00:39 crc kubenswrapper[4789]: I1122 08:00:39.021199 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-442k7" Nov 22 08:00:39 crc kubenswrapper[4789]: I1122 08:00:39.023957 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 22 08:00:39 crc kubenswrapper[4789]: I1122 08:00:39.027827 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-442k7"] Nov 22 08:00:39 crc kubenswrapper[4789]: I1122 08:00:39.163176 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nh28m\" (UniqueName: \"kubernetes.io/projected/a74f8c2f-6d6f-4529-ae51-d72146b30b10-kube-api-access-nh28m\") pod \"certified-operators-442k7\" (UID: \"a74f8c2f-6d6f-4529-ae51-d72146b30b10\") " pod="openshift-marketplace/certified-operators-442k7" Nov 22 08:00:39 crc kubenswrapper[4789]: I1122 08:00:39.163524 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a74f8c2f-6d6f-4529-ae51-d72146b30b10-catalog-content\") pod \"certified-operators-442k7\" (UID: \"a74f8c2f-6d6f-4529-ae51-d72146b30b10\") " pod="openshift-marketplace/certified-operators-442k7" Nov 22 08:00:39 crc kubenswrapper[4789]: I1122 08:00:39.163556 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a74f8c2f-6d6f-4529-ae51-d72146b30b10-utilities\") pod \"certified-operators-442k7\" (UID: \"a74f8c2f-6d6f-4529-ae51-d72146b30b10\") " pod="openshift-marketplace/certified-operators-442k7" Nov 22 08:00:39 crc kubenswrapper[4789]: I1122 08:00:39.187680 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-2tqfp" Nov 22 08:00:39 crc kubenswrapper[4789]: I1122 08:00:39.264365 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nh28m\" (UniqueName: \"kubernetes.io/projected/a74f8c2f-6d6f-4529-ae51-d72146b30b10-kube-api-access-nh28m\") pod \"certified-operators-442k7\" (UID: \"a74f8c2f-6d6f-4529-ae51-d72146b30b10\") " pod="openshift-marketplace/certified-operators-442k7" Nov 22 08:00:39 crc kubenswrapper[4789]: I1122 08:00:39.264425 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a74f8c2f-6d6f-4529-ae51-d72146b30b10-catalog-content\") pod \"certified-operators-442k7\" (UID: \"a74f8c2f-6d6f-4529-ae51-d72146b30b10\") " pod="openshift-marketplace/certified-operators-442k7" Nov 22 08:00:39 crc kubenswrapper[4789]: I1122 08:00:39.264469 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a74f8c2f-6d6f-4529-ae51-d72146b30b10-utilities\") pod \"certified-operators-442k7\" (UID: \"a74f8c2f-6d6f-4529-ae51-d72146b30b10\") " pod="openshift-marketplace/certified-operators-442k7" Nov 22 08:00:39 crc kubenswrapper[4789]: I1122 08:00:39.265096 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a74f8c2f-6d6f-4529-ae51-d72146b30b10-utilities\") pod \"certified-operators-442k7\" (UID: \"a74f8c2f-6d6f-4529-ae51-d72146b30b10\") " pod="openshift-marketplace/certified-operators-442k7" Nov 22 08:00:39 crc kubenswrapper[4789]: I1122 08:00:39.265213 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a74f8c2f-6d6f-4529-ae51-d72146b30b10-catalog-content\") pod \"certified-operators-442k7\" (UID: \"a74f8c2f-6d6f-4529-ae51-d72146b30b10\") " pod="openshift-marketplace/certified-operators-442k7" Nov 22 08:00:39 crc kubenswrapper[4789]: I1122 08:00:39.284324 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nh28m\" (UniqueName: \"kubernetes.io/projected/a74f8c2f-6d6f-4529-ae51-d72146b30b10-kube-api-access-nh28m\") pod \"certified-operators-442k7\" (UID: \"a74f8c2f-6d6f-4529-ae51-d72146b30b10\") " pod="openshift-marketplace/certified-operators-442k7" Nov 22 08:00:39 crc kubenswrapper[4789]: I1122 08:00:39.337444 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-442k7" Nov 22 08:00:39 crc kubenswrapper[4789]: I1122 08:00:39.618662 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-b96v6"] Nov 22 08:00:39 crc kubenswrapper[4789]: I1122 08:00:39.620325 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b96v6" Nov 22 08:00:39 crc kubenswrapper[4789]: I1122 08:00:39.624965 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 22 08:00:39 crc kubenswrapper[4789]: I1122 08:00:39.629387 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-b96v6"] Nov 22 08:00:39 crc kubenswrapper[4789]: I1122 08:00:39.721508 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-442k7"] Nov 22 08:00:39 crc kubenswrapper[4789]: W1122 08:00:39.731862 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda74f8c2f_6d6f_4529_ae51_d72146b30b10.slice/crio-19e480a5cdf95cef82a537dcf30899d2e685e01e4476f0fa45285a98691ff8e7 WatchSource:0}: Error finding container 19e480a5cdf95cef82a537dcf30899d2e685e01e4476f0fa45285a98691ff8e7: Status 404 returned error can't find the container with id 19e480a5cdf95cef82a537dcf30899d2e685e01e4476f0fa45285a98691ff8e7 Nov 22 08:00:39 crc kubenswrapper[4789]: I1122 08:00:39.769585 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c6da9b7c-c7b1-4a6c-afac-525ffaa3b204-utilities\") pod \"redhat-marketplace-b96v6\" (UID: \"c6da9b7c-c7b1-4a6c-afac-525ffaa3b204\") " pod="openshift-marketplace/redhat-marketplace-b96v6" Nov 22 08:00:39 crc kubenswrapper[4789]: I1122 08:00:39.769631 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lxp6b\" (UniqueName: \"kubernetes.io/projected/c6da9b7c-c7b1-4a6c-afac-525ffaa3b204-kube-api-access-lxp6b\") pod \"redhat-marketplace-b96v6\" (UID: \"c6da9b7c-c7b1-4a6c-afac-525ffaa3b204\") " pod="openshift-marketplace/redhat-marketplace-b96v6" Nov 22 08:00:39 crc kubenswrapper[4789]: I1122 08:00:39.769808 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c6da9b7c-c7b1-4a6c-afac-525ffaa3b204-catalog-content\") pod \"redhat-marketplace-b96v6\" (UID: \"c6da9b7c-c7b1-4a6c-afac-525ffaa3b204\") " pod="openshift-marketplace/redhat-marketplace-b96v6" Nov 22 08:00:39 crc kubenswrapper[4789]: I1122 08:00:39.870716 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c6da9b7c-c7b1-4a6c-afac-525ffaa3b204-utilities\") pod \"redhat-marketplace-b96v6\" (UID: \"c6da9b7c-c7b1-4a6c-afac-525ffaa3b204\") " pod="openshift-marketplace/redhat-marketplace-b96v6" Nov 22 08:00:39 crc kubenswrapper[4789]: I1122 08:00:39.870782 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lxp6b\" (UniqueName: \"kubernetes.io/projected/c6da9b7c-c7b1-4a6c-afac-525ffaa3b204-kube-api-access-lxp6b\") pod \"redhat-marketplace-b96v6\" (UID: \"c6da9b7c-c7b1-4a6c-afac-525ffaa3b204\") " pod="openshift-marketplace/redhat-marketplace-b96v6" Nov 22 08:00:39 crc kubenswrapper[4789]: I1122 08:00:39.870822 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c6da9b7c-c7b1-4a6c-afac-525ffaa3b204-catalog-content\") pod \"redhat-marketplace-b96v6\" (UID: \"c6da9b7c-c7b1-4a6c-afac-525ffaa3b204\") " pod="openshift-marketplace/redhat-marketplace-b96v6" Nov 22 08:00:39 crc kubenswrapper[4789]: I1122 08:00:39.871241 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c6da9b7c-c7b1-4a6c-afac-525ffaa3b204-utilities\") pod \"redhat-marketplace-b96v6\" (UID: \"c6da9b7c-c7b1-4a6c-afac-525ffaa3b204\") " pod="openshift-marketplace/redhat-marketplace-b96v6" Nov 22 08:00:39 crc kubenswrapper[4789]: I1122 08:00:39.871305 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c6da9b7c-c7b1-4a6c-afac-525ffaa3b204-catalog-content\") pod \"redhat-marketplace-b96v6\" (UID: \"c6da9b7c-c7b1-4a6c-afac-525ffaa3b204\") " pod="openshift-marketplace/redhat-marketplace-b96v6" Nov 22 08:00:39 crc kubenswrapper[4789]: I1122 08:00:39.889972 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lxp6b\" (UniqueName: \"kubernetes.io/projected/c6da9b7c-c7b1-4a6c-afac-525ffaa3b204-kube-api-access-lxp6b\") pod \"redhat-marketplace-b96v6\" (UID: \"c6da9b7c-c7b1-4a6c-afac-525ffaa3b204\") " pod="openshift-marketplace/redhat-marketplace-b96v6" Nov 22 08:00:39 crc kubenswrapper[4789]: I1122 08:00:39.948523 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b96v6" Nov 22 08:00:39 crc kubenswrapper[4789]: I1122 08:00:39.971542 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2bbcd7cf-ccb1-4b91-9e71-2aa87677be19" path="/var/lib/kubelet/pods/2bbcd7cf-ccb1-4b91-9e71-2aa87677be19/volumes" Nov 22 08:00:39 crc kubenswrapper[4789]: I1122 08:00:39.972791 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7426acc3-0cbe-43f6-a327-886c9de10a94" path="/var/lib/kubelet/pods/7426acc3-0cbe-43f6-a327-886c9de10a94/volumes" Nov 22 08:00:39 crc kubenswrapper[4789]: I1122 08:00:39.973331 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8235ef4a-0670-4b55-bfe6-4cc3cb1b24de" path="/var/lib/kubelet/pods/8235ef4a-0670-4b55-bfe6-4cc3cb1b24de/volumes" Nov 22 08:00:39 crc kubenswrapper[4789]: I1122 08:00:39.974411 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9eb8d5d8-3ad7-41d1-9bb2-2d6865619967" path="/var/lib/kubelet/pods/9eb8d5d8-3ad7-41d1-9bb2-2d6865619967/volumes" Nov 22 08:00:39 crc kubenswrapper[4789]: I1122 08:00:39.974868 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d4a4f530-1afc-402b-8026-584bba4f8f1a" path="/var/lib/kubelet/pods/d4a4f530-1afc-402b-8026-584bba4f8f1a/volumes" Nov 22 08:00:40 crc kubenswrapper[4789]: I1122 08:00:40.145684 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-b96v6"] Nov 22 08:00:40 crc kubenswrapper[4789]: I1122 08:00:40.189153 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-442k7" event={"ID":"a74f8c2f-6d6f-4529-ae51-d72146b30b10","Type":"ContainerStarted","Data":"19e480a5cdf95cef82a537dcf30899d2e685e01e4476f0fa45285a98691ff8e7"} Nov 22 08:00:40 crc kubenswrapper[4789]: I1122 08:00:40.190099 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b96v6" event={"ID":"c6da9b7c-c7b1-4a6c-afac-525ffaa3b204","Type":"ContainerStarted","Data":"45c10631c83ae0b94db3a87511d9374f50a7762d37149ce5d68a20dc4d309cd8"} Nov 22 08:00:41 crc kubenswrapper[4789]: I1122 08:00:41.195066 4789 generic.go:334] "Generic (PLEG): container finished" podID="a74f8c2f-6d6f-4529-ae51-d72146b30b10" containerID="4406455f81a28b8f41d647dc7d07aec5d696acd3b1156a573db5fe693e770d37" exitCode=0 Nov 22 08:00:41 crc kubenswrapper[4789]: I1122 08:00:41.195161 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-442k7" event={"ID":"a74f8c2f-6d6f-4529-ae51-d72146b30b10","Type":"ContainerDied","Data":"4406455f81a28b8f41d647dc7d07aec5d696acd3b1156a573db5fe693e770d37"} Nov 22 08:00:41 crc kubenswrapper[4789]: I1122 08:00:41.196927 4789 generic.go:334] "Generic (PLEG): container finished" podID="c6da9b7c-c7b1-4a6c-afac-525ffaa3b204" containerID="08becad61d1a552f48b587abf8d1468fa505f6d2a27167f7d00ee6cd820cba7f" exitCode=0 Nov 22 08:00:41 crc kubenswrapper[4789]: I1122 08:00:41.196957 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b96v6" event={"ID":"c6da9b7c-c7b1-4a6c-afac-525ffaa3b204","Type":"ContainerDied","Data":"08becad61d1a552f48b587abf8d1468fa505f6d2a27167f7d00ee6cd820cba7f"} Nov 22 08:00:41 crc kubenswrapper[4789]: I1122 08:00:41.413882 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-z9cvt"] Nov 22 08:00:41 crc kubenswrapper[4789]: I1122 08:00:41.415056 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z9cvt" Nov 22 08:00:41 crc kubenswrapper[4789]: I1122 08:00:41.419863 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 22 08:00:41 crc kubenswrapper[4789]: I1122 08:00:41.425508 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-z9cvt"] Nov 22 08:00:41 crc kubenswrapper[4789]: I1122 08:00:41.589832 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a253f042-14c5-4345-8072-55ab86201242-utilities\") pod \"redhat-operators-z9cvt\" (UID: \"a253f042-14c5-4345-8072-55ab86201242\") " pod="openshift-marketplace/redhat-operators-z9cvt" Nov 22 08:00:41 crc kubenswrapper[4789]: I1122 08:00:41.589874 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a253f042-14c5-4345-8072-55ab86201242-catalog-content\") pod \"redhat-operators-z9cvt\" (UID: \"a253f042-14c5-4345-8072-55ab86201242\") " pod="openshift-marketplace/redhat-operators-z9cvt" Nov 22 08:00:41 crc kubenswrapper[4789]: I1122 08:00:41.589905 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-47lhc\" (UniqueName: \"kubernetes.io/projected/a253f042-14c5-4345-8072-55ab86201242-kube-api-access-47lhc\") pod \"redhat-operators-z9cvt\" (UID: \"a253f042-14c5-4345-8072-55ab86201242\") " pod="openshift-marketplace/redhat-operators-z9cvt" Nov 22 08:00:41 crc kubenswrapper[4789]: I1122 08:00:41.691334 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a253f042-14c5-4345-8072-55ab86201242-utilities\") pod \"redhat-operators-z9cvt\" (UID: \"a253f042-14c5-4345-8072-55ab86201242\") " pod="openshift-marketplace/redhat-operators-z9cvt" Nov 22 08:00:41 crc kubenswrapper[4789]: I1122 08:00:41.691379 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a253f042-14c5-4345-8072-55ab86201242-catalog-content\") pod \"redhat-operators-z9cvt\" (UID: \"a253f042-14c5-4345-8072-55ab86201242\") " pod="openshift-marketplace/redhat-operators-z9cvt" Nov 22 08:00:41 crc kubenswrapper[4789]: I1122 08:00:41.691422 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-47lhc\" (UniqueName: \"kubernetes.io/projected/a253f042-14c5-4345-8072-55ab86201242-kube-api-access-47lhc\") pod \"redhat-operators-z9cvt\" (UID: \"a253f042-14c5-4345-8072-55ab86201242\") " pod="openshift-marketplace/redhat-operators-z9cvt" Nov 22 08:00:41 crc kubenswrapper[4789]: I1122 08:00:41.691837 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a253f042-14c5-4345-8072-55ab86201242-utilities\") pod \"redhat-operators-z9cvt\" (UID: \"a253f042-14c5-4345-8072-55ab86201242\") " pod="openshift-marketplace/redhat-operators-z9cvt" Nov 22 08:00:41 crc kubenswrapper[4789]: I1122 08:00:41.692033 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a253f042-14c5-4345-8072-55ab86201242-catalog-content\") pod \"redhat-operators-z9cvt\" (UID: \"a253f042-14c5-4345-8072-55ab86201242\") " pod="openshift-marketplace/redhat-operators-z9cvt" Nov 22 08:00:41 crc kubenswrapper[4789]: I1122 08:00:41.712724 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-47lhc\" (UniqueName: \"kubernetes.io/projected/a253f042-14c5-4345-8072-55ab86201242-kube-api-access-47lhc\") pod \"redhat-operators-z9cvt\" (UID: \"a253f042-14c5-4345-8072-55ab86201242\") " pod="openshift-marketplace/redhat-operators-z9cvt" Nov 22 08:00:41 crc kubenswrapper[4789]: I1122 08:00:41.733401 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z9cvt" Nov 22 08:00:42 crc kubenswrapper[4789]: I1122 08:00:42.018336 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-rpzdx"] Nov 22 08:00:42 crc kubenswrapper[4789]: I1122 08:00:42.021523 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rpzdx" Nov 22 08:00:42 crc kubenswrapper[4789]: I1122 08:00:42.023441 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rpzdx"] Nov 22 08:00:42 crc kubenswrapper[4789]: I1122 08:00:42.024729 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 22 08:00:42 crc kubenswrapper[4789]: I1122 08:00:42.197461 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8rg5s\" (UniqueName: \"kubernetes.io/projected/60fde5c9-9cd4-4966-8183-01e3c24970b0-kube-api-access-8rg5s\") pod \"community-operators-rpzdx\" (UID: \"60fde5c9-9cd4-4966-8183-01e3c24970b0\") " pod="openshift-marketplace/community-operators-rpzdx" Nov 22 08:00:42 crc kubenswrapper[4789]: I1122 08:00:42.197908 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/60fde5c9-9cd4-4966-8183-01e3c24970b0-utilities\") pod \"community-operators-rpzdx\" (UID: \"60fde5c9-9cd4-4966-8183-01e3c24970b0\") " pod="openshift-marketplace/community-operators-rpzdx" Nov 22 08:00:42 crc kubenswrapper[4789]: I1122 08:00:42.198016 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/60fde5c9-9cd4-4966-8183-01e3c24970b0-catalog-content\") pod \"community-operators-rpzdx\" (UID: \"60fde5c9-9cd4-4966-8183-01e3c24970b0\") " pod="openshift-marketplace/community-operators-rpzdx" Nov 22 08:00:42 crc kubenswrapper[4789]: I1122 08:00:42.221531 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-z9cvt"] Nov 22 08:00:42 crc kubenswrapper[4789]: W1122 08:00:42.228314 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda253f042_14c5_4345_8072_55ab86201242.slice/crio-f3fd1f19a3cf47d8992d8a8fbb930a8c60bad2a6031388771e867741c0b2cf9f WatchSource:0}: Error finding container f3fd1f19a3cf47d8992d8a8fbb930a8c60bad2a6031388771e867741c0b2cf9f: Status 404 returned error can't find the container with id f3fd1f19a3cf47d8992d8a8fbb930a8c60bad2a6031388771e867741c0b2cf9f Nov 22 08:00:42 crc kubenswrapper[4789]: I1122 08:00:42.299360 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8rg5s\" (UniqueName: \"kubernetes.io/projected/60fde5c9-9cd4-4966-8183-01e3c24970b0-kube-api-access-8rg5s\") pod \"community-operators-rpzdx\" (UID: \"60fde5c9-9cd4-4966-8183-01e3c24970b0\") " pod="openshift-marketplace/community-operators-rpzdx" Nov 22 08:00:42 crc kubenswrapper[4789]: I1122 08:00:42.299415 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/60fde5c9-9cd4-4966-8183-01e3c24970b0-utilities\") pod \"community-operators-rpzdx\" (UID: \"60fde5c9-9cd4-4966-8183-01e3c24970b0\") " pod="openshift-marketplace/community-operators-rpzdx" Nov 22 08:00:42 crc kubenswrapper[4789]: I1122 08:00:42.299462 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/60fde5c9-9cd4-4966-8183-01e3c24970b0-catalog-content\") pod \"community-operators-rpzdx\" (UID: \"60fde5c9-9cd4-4966-8183-01e3c24970b0\") " pod="openshift-marketplace/community-operators-rpzdx" Nov 22 08:00:42 crc kubenswrapper[4789]: I1122 08:00:42.299938 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/60fde5c9-9cd4-4966-8183-01e3c24970b0-catalog-content\") pod \"community-operators-rpzdx\" (UID: \"60fde5c9-9cd4-4966-8183-01e3c24970b0\") " pod="openshift-marketplace/community-operators-rpzdx" Nov 22 08:00:42 crc kubenswrapper[4789]: I1122 08:00:42.300882 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/60fde5c9-9cd4-4966-8183-01e3c24970b0-utilities\") pod \"community-operators-rpzdx\" (UID: \"60fde5c9-9cd4-4966-8183-01e3c24970b0\") " pod="openshift-marketplace/community-operators-rpzdx" Nov 22 08:00:42 crc kubenswrapper[4789]: I1122 08:00:42.318781 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8rg5s\" (UniqueName: \"kubernetes.io/projected/60fde5c9-9cd4-4966-8183-01e3c24970b0-kube-api-access-8rg5s\") pod \"community-operators-rpzdx\" (UID: \"60fde5c9-9cd4-4966-8183-01e3c24970b0\") " pod="openshift-marketplace/community-operators-rpzdx" Nov 22 08:00:42 crc kubenswrapper[4789]: I1122 08:00:42.340331 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rpzdx" Nov 22 08:00:42 crc kubenswrapper[4789]: I1122 08:00:42.510997 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rpzdx"] Nov 22 08:00:42 crc kubenswrapper[4789]: W1122 08:00:42.515917 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod60fde5c9_9cd4_4966_8183_01e3c24970b0.slice/crio-caafba200eab213619cf714f97ba8f0ddd65092472e7353feaf4e4f515b49fac WatchSource:0}: Error finding container caafba200eab213619cf714f97ba8f0ddd65092472e7353feaf4e4f515b49fac: Status 404 returned error can't find the container with id caafba200eab213619cf714f97ba8f0ddd65092472e7353feaf4e4f515b49fac Nov 22 08:00:43 crc kubenswrapper[4789]: I1122 08:00:43.215438 4789 generic.go:334] "Generic (PLEG): container finished" podID="60fde5c9-9cd4-4966-8183-01e3c24970b0" containerID="0cf3b3d6989f960b6e34ef32263a3c128260ca83ddbd04dc79cdff70925bc719" exitCode=0 Nov 22 08:00:43 crc kubenswrapper[4789]: I1122 08:00:43.215846 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rpzdx" event={"ID":"60fde5c9-9cd4-4966-8183-01e3c24970b0","Type":"ContainerDied","Data":"0cf3b3d6989f960b6e34ef32263a3c128260ca83ddbd04dc79cdff70925bc719"} Nov 22 08:00:43 crc kubenswrapper[4789]: I1122 08:00:43.215880 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rpzdx" event={"ID":"60fde5c9-9cd4-4966-8183-01e3c24970b0","Type":"ContainerStarted","Data":"caafba200eab213619cf714f97ba8f0ddd65092472e7353feaf4e4f515b49fac"} Nov 22 08:00:43 crc kubenswrapper[4789]: I1122 08:00:43.218969 4789 generic.go:334] "Generic (PLEG): container finished" podID="a74f8c2f-6d6f-4529-ae51-d72146b30b10" containerID="6524b297afd436ac15dc96bf32082b7819d329ba1936d77b245b9e351227d3f1" exitCode=0 Nov 22 08:00:43 crc kubenswrapper[4789]: I1122 08:00:43.219051 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-442k7" event={"ID":"a74f8c2f-6d6f-4529-ae51-d72146b30b10","Type":"ContainerDied","Data":"6524b297afd436ac15dc96bf32082b7819d329ba1936d77b245b9e351227d3f1"} Nov 22 08:00:43 crc kubenswrapper[4789]: I1122 08:00:43.222534 4789 generic.go:334] "Generic (PLEG): container finished" podID="c6da9b7c-c7b1-4a6c-afac-525ffaa3b204" containerID="6e0845efc1a832c6b69a98d68aed4409d9e48363b2443468eaaa3b910156ed65" exitCode=0 Nov 22 08:00:43 crc kubenswrapper[4789]: I1122 08:00:43.222607 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b96v6" event={"ID":"c6da9b7c-c7b1-4a6c-afac-525ffaa3b204","Type":"ContainerDied","Data":"6e0845efc1a832c6b69a98d68aed4409d9e48363b2443468eaaa3b910156ed65"} Nov 22 08:00:43 crc kubenswrapper[4789]: I1122 08:00:43.229301 4789 generic.go:334] "Generic (PLEG): container finished" podID="a253f042-14c5-4345-8072-55ab86201242" containerID="8b4fdbb66bde8db83de6e979354de3f3a198d201f0c3a3c2c5b654d5d8ad5816" exitCode=0 Nov 22 08:00:43 crc kubenswrapper[4789]: I1122 08:00:43.229341 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z9cvt" event={"ID":"a253f042-14c5-4345-8072-55ab86201242","Type":"ContainerDied","Data":"8b4fdbb66bde8db83de6e979354de3f3a198d201f0c3a3c2c5b654d5d8ad5816"} Nov 22 08:00:43 crc kubenswrapper[4789]: I1122 08:00:43.229362 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z9cvt" event={"ID":"a253f042-14c5-4345-8072-55ab86201242","Type":"ContainerStarted","Data":"f3fd1f19a3cf47d8992d8a8fbb930a8c60bad2a6031388771e867741c0b2cf9f"} Nov 22 08:00:46 crc kubenswrapper[4789]: I1122 08:00:46.248119 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rpzdx" event={"ID":"60fde5c9-9cd4-4966-8183-01e3c24970b0","Type":"ContainerStarted","Data":"c340535caeae2e75981daf316f37e3a80a3b28d759f0d4cd4c8451a2adb9d38b"} Nov 22 08:00:46 crc kubenswrapper[4789]: I1122 08:00:46.251384 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-442k7" event={"ID":"a74f8c2f-6d6f-4529-ae51-d72146b30b10","Type":"ContainerStarted","Data":"f53e0014460f83a6cef2a02b6487ab66357e7bab6e8013d0d65081ae3f77da33"} Nov 22 08:00:46 crc kubenswrapper[4789]: I1122 08:00:46.252930 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b96v6" event={"ID":"c6da9b7c-c7b1-4a6c-afac-525ffaa3b204","Type":"ContainerStarted","Data":"9603a360138c80927d8293de7173b2a4514a3b8ff5e1bd154143b3f42aceb51f"} Nov 22 08:00:46 crc kubenswrapper[4789]: I1122 08:00:46.254169 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z9cvt" event={"ID":"a253f042-14c5-4345-8072-55ab86201242","Type":"ContainerStarted","Data":"ce27c7c483e16422b61e64aa3521cb6b02acf48825b1450cd2ce0389b72bb514"} Nov 22 08:00:46 crc kubenswrapper[4789]: I1122 08:00:46.295951 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-442k7" podStartSLOduration=3.325272833 podStartE2EDuration="7.295928826s" podCreationTimestamp="2025-11-22 08:00:39 +0000 UTC" firstStartedPulling="2025-11-22 08:00:41.196825458 +0000 UTC m=+255.431225731" lastFinishedPulling="2025-11-22 08:00:45.167481451 +0000 UTC m=+259.401881724" observedRunningTime="2025-11-22 08:00:46.289576559 +0000 UTC m=+260.523976852" watchObservedRunningTime="2025-11-22 08:00:46.295928826 +0000 UTC m=+260.530329099" Nov 22 08:00:46 crc kubenswrapper[4789]: I1122 08:00:46.329685 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-b96v6" podStartSLOduration=3.20493133 podStartE2EDuration="7.329654192s" podCreationTimestamp="2025-11-22 08:00:39 +0000 UTC" firstStartedPulling="2025-11-22 08:00:41.197913479 +0000 UTC m=+255.432313752" lastFinishedPulling="2025-11-22 08:00:45.322636331 +0000 UTC m=+259.557036614" observedRunningTime="2025-11-22 08:00:46.32813963 +0000 UTC m=+260.562539903" watchObservedRunningTime="2025-11-22 08:00:46.329654192 +0000 UTC m=+260.564054485" Nov 22 08:00:47 crc kubenswrapper[4789]: I1122 08:00:47.267199 4789 generic.go:334] "Generic (PLEG): container finished" podID="a253f042-14c5-4345-8072-55ab86201242" containerID="ce27c7c483e16422b61e64aa3521cb6b02acf48825b1450cd2ce0389b72bb514" exitCode=0 Nov 22 08:00:47 crc kubenswrapper[4789]: I1122 08:00:47.267394 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z9cvt" event={"ID":"a253f042-14c5-4345-8072-55ab86201242","Type":"ContainerDied","Data":"ce27c7c483e16422b61e64aa3521cb6b02acf48825b1450cd2ce0389b72bb514"} Nov 22 08:00:47 crc kubenswrapper[4789]: I1122 08:00:47.270302 4789 generic.go:334] "Generic (PLEG): container finished" podID="60fde5c9-9cd4-4966-8183-01e3c24970b0" containerID="c340535caeae2e75981daf316f37e3a80a3b28d759f0d4cd4c8451a2adb9d38b" exitCode=0 Nov 22 08:00:47 crc kubenswrapper[4789]: I1122 08:00:47.270410 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rpzdx" event={"ID":"60fde5c9-9cd4-4966-8183-01e3c24970b0","Type":"ContainerDied","Data":"c340535caeae2e75981daf316f37e3a80a3b28d759f0d4cd4c8451a2adb9d38b"} Nov 22 08:00:49 crc kubenswrapper[4789]: I1122 08:00:49.281641 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rpzdx" event={"ID":"60fde5c9-9cd4-4966-8183-01e3c24970b0","Type":"ContainerStarted","Data":"44050f61da68741145f56433dfcedb45fbeed06c82a5c9ad57a32fd546225b7d"} Nov 22 08:00:49 crc kubenswrapper[4789]: I1122 08:00:49.284020 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z9cvt" event={"ID":"a253f042-14c5-4345-8072-55ab86201242","Type":"ContainerStarted","Data":"f4bf4629ad595f15b319b20e900e9c01f67265c2a336caf256054ddfa2bd5031"} Nov 22 08:00:49 crc kubenswrapper[4789]: I1122 08:00:49.302530 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-rpzdx" podStartSLOduration=1.961541833 podStartE2EDuration="7.302513419s" podCreationTimestamp="2025-11-22 08:00:42 +0000 UTC" firstStartedPulling="2025-11-22 08:00:43.218307268 +0000 UTC m=+257.452707541" lastFinishedPulling="2025-11-22 08:00:48.559278854 +0000 UTC m=+262.793679127" observedRunningTime="2025-11-22 08:00:49.300411851 +0000 UTC m=+263.534812134" watchObservedRunningTime="2025-11-22 08:00:49.302513419 +0000 UTC m=+263.536913692" Nov 22 08:00:49 crc kubenswrapper[4789]: I1122 08:00:49.328090 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-z9cvt" podStartSLOduration=2.977835 podStartE2EDuration="8.328074698s" podCreationTimestamp="2025-11-22 08:00:41 +0000 UTC" firstStartedPulling="2025-11-22 08:00:43.233577898 +0000 UTC m=+257.467978171" lastFinishedPulling="2025-11-22 08:00:48.583817576 +0000 UTC m=+262.818217869" observedRunningTime="2025-11-22 08:00:49.325391164 +0000 UTC m=+263.559791447" watchObservedRunningTime="2025-11-22 08:00:49.328074698 +0000 UTC m=+263.562474971" Nov 22 08:00:49 crc kubenswrapper[4789]: I1122 08:00:49.338516 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-442k7" Nov 22 08:00:49 crc kubenswrapper[4789]: I1122 08:00:49.338600 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-442k7" Nov 22 08:00:49 crc kubenswrapper[4789]: I1122 08:00:49.376650 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-442k7" Nov 22 08:00:49 crc kubenswrapper[4789]: I1122 08:00:49.949427 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-b96v6" Nov 22 08:00:49 crc kubenswrapper[4789]: I1122 08:00:49.949818 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-b96v6" Nov 22 08:00:49 crc kubenswrapper[4789]: I1122 08:00:49.990263 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-b96v6" Nov 22 08:00:51 crc kubenswrapper[4789]: I1122 08:00:51.734612 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-z9cvt" Nov 22 08:00:51 crc kubenswrapper[4789]: I1122 08:00:51.734966 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-z9cvt" Nov 22 08:00:52 crc kubenswrapper[4789]: I1122 08:00:52.341470 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-rpzdx" Nov 22 08:00:52 crc kubenswrapper[4789]: I1122 08:00:52.341524 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-rpzdx" Nov 22 08:00:52 crc kubenswrapper[4789]: I1122 08:00:52.379762 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-rpzdx" Nov 22 08:00:52 crc kubenswrapper[4789]: I1122 08:00:52.776935 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-z9cvt" podUID="a253f042-14c5-4345-8072-55ab86201242" containerName="registry-server" probeResult="failure" output=< Nov 22 08:00:52 crc kubenswrapper[4789]: timeout: failed to connect service ":50051" within 1s Nov 22 08:00:52 crc kubenswrapper[4789]: > Nov 22 08:00:53 crc kubenswrapper[4789]: I1122 08:00:53.343190 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-rpzdx" Nov 22 08:00:59 crc kubenswrapper[4789]: I1122 08:00:59.375171 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-442k7" Nov 22 08:01:00 crc kubenswrapper[4789]: I1122 08:01:00.019634 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-b96v6" Nov 22 08:01:01 crc kubenswrapper[4789]: I1122 08:01:01.772521 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-z9cvt" Nov 22 08:01:01 crc kubenswrapper[4789]: I1122 08:01:01.820019 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-z9cvt" Nov 22 08:01:03 crc kubenswrapper[4789]: I1122 08:01:03.747405 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-wnfzj" podUID="55cac199-8203-40b3-8a4f-c571017d998b" containerName="oauth-openshift" containerID="cri-o://a9caf00327bdcea8f671019a3e4ffeb6fb58a2fb6bc5f3641a2a64cdfadfb78d" gracePeriod=15 Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.084937 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-wnfzj" Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.116899 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-8445cf6b-czcxm"] Nov 22 08:01:04 crc kubenswrapper[4789]: E1122 08:01:04.117211 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55cac199-8203-40b3-8a4f-c571017d998b" containerName="oauth-openshift" Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.117228 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="55cac199-8203-40b3-8a4f-c571017d998b" containerName="oauth-openshift" Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.117329 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="55cac199-8203-40b3-8a4f-c571017d998b" containerName="oauth-openshift" Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.117714 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-8445cf6b-czcxm" Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.122200 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-8445cf6b-czcxm"] Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.131631 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/55cac199-8203-40b3-8a4f-c571017d998b-v4-0-config-system-session\") pod \"55cac199-8203-40b3-8a4f-c571017d998b\" (UID: \"55cac199-8203-40b3-8a4f-c571017d998b\") " Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.131667 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/55cac199-8203-40b3-8a4f-c571017d998b-audit-policies\") pod \"55cac199-8203-40b3-8a4f-c571017d998b\" (UID: \"55cac199-8203-40b3-8a4f-c571017d998b\") " Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.131687 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/55cac199-8203-40b3-8a4f-c571017d998b-v4-0-config-user-template-login\") pod \"55cac199-8203-40b3-8a4f-c571017d998b\" (UID: \"55cac199-8203-40b3-8a4f-c571017d998b\") " Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.131836 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-49kt8\" (UniqueName: \"kubernetes.io/projected/f5a3a04f-033f-45ca-b676-e72936d5d92c-kube-api-access-49kt8\") pod \"oauth-openshift-8445cf6b-czcxm\" (UID: \"f5a3a04f-033f-45ca-b676-e72936d5d92c\") " pod="openshift-authentication/oauth-openshift-8445cf6b-czcxm" Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.131862 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/f5a3a04f-033f-45ca-b676-e72936d5d92c-v4-0-config-system-serving-cert\") pod \"oauth-openshift-8445cf6b-czcxm\" (UID: \"f5a3a04f-033f-45ca-b676-e72936d5d92c\") " pod="openshift-authentication/oauth-openshift-8445cf6b-czcxm" Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.131881 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/f5a3a04f-033f-45ca-b676-e72936d5d92c-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-8445cf6b-czcxm\" (UID: \"f5a3a04f-033f-45ca-b676-e72936d5d92c\") " pod="openshift-authentication/oauth-openshift-8445cf6b-czcxm" Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.131913 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/f5a3a04f-033f-45ca-b676-e72936d5d92c-v4-0-config-system-cliconfig\") pod \"oauth-openshift-8445cf6b-czcxm\" (UID: \"f5a3a04f-033f-45ca-b676-e72936d5d92c\") " pod="openshift-authentication/oauth-openshift-8445cf6b-czcxm" Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.131931 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/f5a3a04f-033f-45ca-b676-e72936d5d92c-v4-0-config-user-template-login\") pod \"oauth-openshift-8445cf6b-czcxm\" (UID: \"f5a3a04f-033f-45ca-b676-e72936d5d92c\") " pod="openshift-authentication/oauth-openshift-8445cf6b-czcxm" Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.131951 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f5a3a04f-033f-45ca-b676-e72936d5d92c-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-8445cf6b-czcxm\" (UID: \"f5a3a04f-033f-45ca-b676-e72936d5d92c\") " pod="openshift-authentication/oauth-openshift-8445cf6b-czcxm" Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.131970 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/f5a3a04f-033f-45ca-b676-e72936d5d92c-v4-0-config-user-template-error\") pod \"oauth-openshift-8445cf6b-czcxm\" (UID: \"f5a3a04f-033f-45ca-b676-e72936d5d92c\") " pod="openshift-authentication/oauth-openshift-8445cf6b-czcxm" Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.131988 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/f5a3a04f-033f-45ca-b676-e72936d5d92c-v4-0-config-system-session\") pod \"oauth-openshift-8445cf6b-czcxm\" (UID: \"f5a3a04f-033f-45ca-b676-e72936d5d92c\") " pod="openshift-authentication/oauth-openshift-8445cf6b-czcxm" Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.132010 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/f5a3a04f-033f-45ca-b676-e72936d5d92c-v4-0-config-system-router-certs\") pod \"oauth-openshift-8445cf6b-czcxm\" (UID: \"f5a3a04f-033f-45ca-b676-e72936d5d92c\") " pod="openshift-authentication/oauth-openshift-8445cf6b-czcxm" Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.132033 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f5a3a04f-033f-45ca-b676-e72936d5d92c-audit-policies\") pod \"oauth-openshift-8445cf6b-czcxm\" (UID: \"f5a3a04f-033f-45ca-b676-e72936d5d92c\") " pod="openshift-authentication/oauth-openshift-8445cf6b-czcxm" Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.132050 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f5a3a04f-033f-45ca-b676-e72936d5d92c-audit-dir\") pod \"oauth-openshift-8445cf6b-czcxm\" (UID: \"f5a3a04f-033f-45ca-b676-e72936d5d92c\") " pod="openshift-authentication/oauth-openshift-8445cf6b-czcxm" Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.132066 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/f5a3a04f-033f-45ca-b676-e72936d5d92c-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-8445cf6b-czcxm\" (UID: \"f5a3a04f-033f-45ca-b676-e72936d5d92c\") " pod="openshift-authentication/oauth-openshift-8445cf6b-czcxm" Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.132084 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/f5a3a04f-033f-45ca-b676-e72936d5d92c-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-8445cf6b-czcxm\" (UID: \"f5a3a04f-033f-45ca-b676-e72936d5d92c\") " pod="openshift-authentication/oauth-openshift-8445cf6b-czcxm" Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.132100 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/f5a3a04f-033f-45ca-b676-e72936d5d92c-v4-0-config-system-service-ca\") pod \"oauth-openshift-8445cf6b-czcxm\" (UID: \"f5a3a04f-033f-45ca-b676-e72936d5d92c\") " pod="openshift-authentication/oauth-openshift-8445cf6b-czcxm" Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.132569 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/55cac199-8203-40b3-8a4f-c571017d998b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "55cac199-8203-40b3-8a4f-c571017d998b" (UID: "55cac199-8203-40b3-8a4f-c571017d998b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.141704 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55cac199-8203-40b3-8a4f-c571017d998b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "55cac199-8203-40b3-8a4f-c571017d998b" (UID: "55cac199-8203-40b3-8a4f-c571017d998b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.142684 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55cac199-8203-40b3-8a4f-c571017d998b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "55cac199-8203-40b3-8a4f-c571017d998b" (UID: "55cac199-8203-40b3-8a4f-c571017d998b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.232448 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/55cac199-8203-40b3-8a4f-c571017d998b-v4-0-config-system-trusted-ca-bundle\") pod \"55cac199-8203-40b3-8a4f-c571017d998b\" (UID: \"55cac199-8203-40b3-8a4f-c571017d998b\") " Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.232492 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/55cac199-8203-40b3-8a4f-c571017d998b-v4-0-config-system-serving-cert\") pod \"55cac199-8203-40b3-8a4f-c571017d998b\" (UID: \"55cac199-8203-40b3-8a4f-c571017d998b\") " Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.232521 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/55cac199-8203-40b3-8a4f-c571017d998b-v4-0-config-user-idp-0-file-data\") pod \"55cac199-8203-40b3-8a4f-c571017d998b\" (UID: \"55cac199-8203-40b3-8a4f-c571017d998b\") " Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.232883 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/55cac199-8203-40b3-8a4f-c571017d998b-v4-0-config-system-router-certs\") pod \"55cac199-8203-40b3-8a4f-c571017d998b\" (UID: \"55cac199-8203-40b3-8a4f-c571017d998b\") " Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.232905 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/55cac199-8203-40b3-8a4f-c571017d998b-audit-dir\") pod \"55cac199-8203-40b3-8a4f-c571017d998b\" (UID: \"55cac199-8203-40b3-8a4f-c571017d998b\") " Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.232924 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/55cac199-8203-40b3-8a4f-c571017d998b-v4-0-config-system-service-ca\") pod \"55cac199-8203-40b3-8a4f-c571017d998b\" (UID: \"55cac199-8203-40b3-8a4f-c571017d998b\") " Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.232940 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wdcz2\" (UniqueName: \"kubernetes.io/projected/55cac199-8203-40b3-8a4f-c571017d998b-kube-api-access-wdcz2\") pod \"55cac199-8203-40b3-8a4f-c571017d998b\" (UID: \"55cac199-8203-40b3-8a4f-c571017d998b\") " Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.232955 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/55cac199-8203-40b3-8a4f-c571017d998b-v4-0-config-user-template-error\") pod \"55cac199-8203-40b3-8a4f-c571017d998b\" (UID: \"55cac199-8203-40b3-8a4f-c571017d998b\") " Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.232973 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/55cac199-8203-40b3-8a4f-c571017d998b-v4-0-config-system-ocp-branding-template\") pod \"55cac199-8203-40b3-8a4f-c571017d998b\" (UID: \"55cac199-8203-40b3-8a4f-c571017d998b\") " Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.232991 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/55cac199-8203-40b3-8a4f-c571017d998b-v4-0-config-system-cliconfig\") pod \"55cac199-8203-40b3-8a4f-c571017d998b\" (UID: \"55cac199-8203-40b3-8a4f-c571017d998b\") " Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.233008 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/55cac199-8203-40b3-8a4f-c571017d998b-v4-0-config-user-template-provider-selection\") pod \"55cac199-8203-40b3-8a4f-c571017d998b\" (UID: \"55cac199-8203-40b3-8a4f-c571017d998b\") " Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.233080 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/f5a3a04f-033f-45ca-b676-e72936d5d92c-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-8445cf6b-czcxm\" (UID: \"f5a3a04f-033f-45ca-b676-e72936d5d92c\") " pod="openshift-authentication/oauth-openshift-8445cf6b-czcxm" Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.233101 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/f5a3a04f-033f-45ca-b676-e72936d5d92c-v4-0-config-system-service-ca\") pod \"oauth-openshift-8445cf6b-czcxm\" (UID: \"f5a3a04f-033f-45ca-b676-e72936d5d92c\") " pod="openshift-authentication/oauth-openshift-8445cf6b-czcxm" Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.233129 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-49kt8\" (UniqueName: \"kubernetes.io/projected/f5a3a04f-033f-45ca-b676-e72936d5d92c-kube-api-access-49kt8\") pod \"oauth-openshift-8445cf6b-czcxm\" (UID: \"f5a3a04f-033f-45ca-b676-e72936d5d92c\") " pod="openshift-authentication/oauth-openshift-8445cf6b-czcxm" Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.233147 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/f5a3a04f-033f-45ca-b676-e72936d5d92c-v4-0-config-system-serving-cert\") pod \"oauth-openshift-8445cf6b-czcxm\" (UID: \"f5a3a04f-033f-45ca-b676-e72936d5d92c\") " pod="openshift-authentication/oauth-openshift-8445cf6b-czcxm" Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.233164 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/f5a3a04f-033f-45ca-b676-e72936d5d92c-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-8445cf6b-czcxm\" (UID: \"f5a3a04f-033f-45ca-b676-e72936d5d92c\") " pod="openshift-authentication/oauth-openshift-8445cf6b-czcxm" Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.233215 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/f5a3a04f-033f-45ca-b676-e72936d5d92c-v4-0-config-system-cliconfig\") pod \"oauth-openshift-8445cf6b-czcxm\" (UID: \"f5a3a04f-033f-45ca-b676-e72936d5d92c\") " pod="openshift-authentication/oauth-openshift-8445cf6b-czcxm" Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.233216 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/55cac199-8203-40b3-8a4f-c571017d998b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "55cac199-8203-40b3-8a4f-c571017d998b" (UID: "55cac199-8203-40b3-8a4f-c571017d998b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.233244 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/f5a3a04f-033f-45ca-b676-e72936d5d92c-v4-0-config-user-template-login\") pod \"oauth-openshift-8445cf6b-czcxm\" (UID: \"f5a3a04f-033f-45ca-b676-e72936d5d92c\") " pod="openshift-authentication/oauth-openshift-8445cf6b-czcxm" Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.233269 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f5a3a04f-033f-45ca-b676-e72936d5d92c-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-8445cf6b-czcxm\" (UID: \"f5a3a04f-033f-45ca-b676-e72936d5d92c\") " pod="openshift-authentication/oauth-openshift-8445cf6b-czcxm" Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.233295 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/f5a3a04f-033f-45ca-b676-e72936d5d92c-v4-0-config-user-template-error\") pod \"oauth-openshift-8445cf6b-czcxm\" (UID: \"f5a3a04f-033f-45ca-b676-e72936d5d92c\") " pod="openshift-authentication/oauth-openshift-8445cf6b-czcxm" Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.233320 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/f5a3a04f-033f-45ca-b676-e72936d5d92c-v4-0-config-system-session\") pod \"oauth-openshift-8445cf6b-czcxm\" (UID: \"f5a3a04f-033f-45ca-b676-e72936d5d92c\") " pod="openshift-authentication/oauth-openshift-8445cf6b-czcxm" Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.233347 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/f5a3a04f-033f-45ca-b676-e72936d5d92c-v4-0-config-system-router-certs\") pod \"oauth-openshift-8445cf6b-czcxm\" (UID: \"f5a3a04f-033f-45ca-b676-e72936d5d92c\") " pod="openshift-authentication/oauth-openshift-8445cf6b-czcxm" Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.233377 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f5a3a04f-033f-45ca-b676-e72936d5d92c-audit-policies\") pod \"oauth-openshift-8445cf6b-czcxm\" (UID: \"f5a3a04f-033f-45ca-b676-e72936d5d92c\") " pod="openshift-authentication/oauth-openshift-8445cf6b-czcxm" Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.233401 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f5a3a04f-033f-45ca-b676-e72936d5d92c-audit-dir\") pod \"oauth-openshift-8445cf6b-czcxm\" (UID: \"f5a3a04f-033f-45ca-b676-e72936d5d92c\") " pod="openshift-authentication/oauth-openshift-8445cf6b-czcxm" Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.233860 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/55cac199-8203-40b3-8a4f-c571017d998b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "55cac199-8203-40b3-8a4f-c571017d998b" (UID: "55cac199-8203-40b3-8a4f-c571017d998b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.233918 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/55cac199-8203-40b3-8a4f-c571017d998b-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "55cac199-8203-40b3-8a4f-c571017d998b" (UID: "55cac199-8203-40b3-8a4f-c571017d998b"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.234132 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/55cac199-8203-40b3-8a4f-c571017d998b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "55cac199-8203-40b3-8a4f-c571017d998b" (UID: "55cac199-8203-40b3-8a4f-c571017d998b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.234410 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/f5a3a04f-033f-45ca-b676-e72936d5d92c-v4-0-config-system-cliconfig\") pod \"oauth-openshift-8445cf6b-czcxm\" (UID: \"f5a3a04f-033f-45ca-b676-e72936d5d92c\") " pod="openshift-authentication/oauth-openshift-8445cf6b-czcxm" Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.233429 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/f5a3a04f-033f-45ca-b676-e72936d5d92c-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-8445cf6b-czcxm\" (UID: \"f5a3a04f-033f-45ca-b676-e72936d5d92c\") " pod="openshift-authentication/oauth-openshift-8445cf6b-czcxm" Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.234766 4789 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/55cac199-8203-40b3-8a4f-c571017d998b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.234784 4789 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/55cac199-8203-40b3-8a4f-c571017d998b-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.234797 4789 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/55cac199-8203-40b3-8a4f-c571017d998b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.234799 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f5a3a04f-033f-45ca-b676-e72936d5d92c-audit-dir\") pod \"oauth-openshift-8445cf6b-czcxm\" (UID: \"f5a3a04f-033f-45ca-b676-e72936d5d92c\") " pod="openshift-authentication/oauth-openshift-8445cf6b-czcxm" Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.234809 4789 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/55cac199-8203-40b3-8a4f-c571017d998b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.234912 4789 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/55cac199-8203-40b3-8a4f-c571017d998b-audit-dir\") on node \"crc\" DevicePath \"\"" Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.235080 4789 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/55cac199-8203-40b3-8a4f-c571017d998b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.235114 4789 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/55cac199-8203-40b3-8a4f-c571017d998b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.235616 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f5a3a04f-033f-45ca-b676-e72936d5d92c-audit-policies\") pod \"oauth-openshift-8445cf6b-czcxm\" (UID: \"f5a3a04f-033f-45ca-b676-e72936d5d92c\") " pod="openshift-authentication/oauth-openshift-8445cf6b-czcxm" Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.236132 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f5a3a04f-033f-45ca-b676-e72936d5d92c-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-8445cf6b-czcxm\" (UID: \"f5a3a04f-033f-45ca-b676-e72936d5d92c\") " pod="openshift-authentication/oauth-openshift-8445cf6b-czcxm" Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.236261 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/f5a3a04f-033f-45ca-b676-e72936d5d92c-v4-0-config-system-service-ca\") pod \"oauth-openshift-8445cf6b-czcxm\" (UID: \"f5a3a04f-033f-45ca-b676-e72936d5d92c\") " pod="openshift-authentication/oauth-openshift-8445cf6b-czcxm" Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.237457 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55cac199-8203-40b3-8a4f-c571017d998b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "55cac199-8203-40b3-8a4f-c571017d998b" (UID: "55cac199-8203-40b3-8a4f-c571017d998b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.238079 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55cac199-8203-40b3-8a4f-c571017d998b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "55cac199-8203-40b3-8a4f-c571017d998b" (UID: "55cac199-8203-40b3-8a4f-c571017d998b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.239892 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55cac199-8203-40b3-8a4f-c571017d998b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "55cac199-8203-40b3-8a4f-c571017d998b" (UID: "55cac199-8203-40b3-8a4f-c571017d998b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.240068 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/f5a3a04f-033f-45ca-b676-e72936d5d92c-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-8445cf6b-czcxm\" (UID: \"f5a3a04f-033f-45ca-b676-e72936d5d92c\") " pod="openshift-authentication/oauth-openshift-8445cf6b-czcxm" Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.240158 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55cac199-8203-40b3-8a4f-c571017d998b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "55cac199-8203-40b3-8a4f-c571017d998b" (UID: "55cac199-8203-40b3-8a4f-c571017d998b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.240521 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/f5a3a04f-033f-45ca-b676-e72936d5d92c-v4-0-config-user-template-login\") pod \"oauth-openshift-8445cf6b-czcxm\" (UID: \"f5a3a04f-033f-45ca-b676-e72936d5d92c\") " pod="openshift-authentication/oauth-openshift-8445cf6b-czcxm" Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.240622 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/f5a3a04f-033f-45ca-b676-e72936d5d92c-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-8445cf6b-czcxm\" (UID: \"f5a3a04f-033f-45ca-b676-e72936d5d92c\") " pod="openshift-authentication/oauth-openshift-8445cf6b-czcxm" Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.240543 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55cac199-8203-40b3-8a4f-c571017d998b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "55cac199-8203-40b3-8a4f-c571017d998b" (UID: "55cac199-8203-40b3-8a4f-c571017d998b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.240666 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55cac199-8203-40b3-8a4f-c571017d998b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "55cac199-8203-40b3-8a4f-c571017d998b" (UID: "55cac199-8203-40b3-8a4f-c571017d998b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.240825 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/f5a3a04f-033f-45ca-b676-e72936d5d92c-v4-0-config-system-serving-cert\") pod \"oauth-openshift-8445cf6b-czcxm\" (UID: \"f5a3a04f-033f-45ca-b676-e72936d5d92c\") " pod="openshift-authentication/oauth-openshift-8445cf6b-czcxm" Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.240869 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/f5a3a04f-033f-45ca-b676-e72936d5d92c-v4-0-config-system-router-certs\") pod \"oauth-openshift-8445cf6b-czcxm\" (UID: \"f5a3a04f-033f-45ca-b676-e72936d5d92c\") " pod="openshift-authentication/oauth-openshift-8445cf6b-czcxm" Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.241520 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/55cac199-8203-40b3-8a4f-c571017d998b-kube-api-access-wdcz2" (OuterVolumeSpecName: "kube-api-access-wdcz2") pod "55cac199-8203-40b3-8a4f-c571017d998b" (UID: "55cac199-8203-40b3-8a4f-c571017d998b"). InnerVolumeSpecName "kube-api-access-wdcz2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.242031 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/f5a3a04f-033f-45ca-b676-e72936d5d92c-v4-0-config-user-template-error\") pod \"oauth-openshift-8445cf6b-czcxm\" (UID: \"f5a3a04f-033f-45ca-b676-e72936d5d92c\") " pod="openshift-authentication/oauth-openshift-8445cf6b-czcxm" Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.242922 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/f5a3a04f-033f-45ca-b676-e72936d5d92c-v4-0-config-system-session\") pod \"oauth-openshift-8445cf6b-czcxm\" (UID: \"f5a3a04f-033f-45ca-b676-e72936d5d92c\") " pod="openshift-authentication/oauth-openshift-8445cf6b-czcxm" Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.243683 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/f5a3a04f-033f-45ca-b676-e72936d5d92c-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-8445cf6b-czcxm\" (UID: \"f5a3a04f-033f-45ca-b676-e72936d5d92c\") " pod="openshift-authentication/oauth-openshift-8445cf6b-czcxm" Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.250882 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-49kt8\" (UniqueName: \"kubernetes.io/projected/f5a3a04f-033f-45ca-b676-e72936d5d92c-kube-api-access-49kt8\") pod \"oauth-openshift-8445cf6b-czcxm\" (UID: \"f5a3a04f-033f-45ca-b676-e72936d5d92c\") " pod="openshift-authentication/oauth-openshift-8445cf6b-czcxm" Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.336604 4789 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/55cac199-8203-40b3-8a4f-c571017d998b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.336645 4789 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/55cac199-8203-40b3-8a4f-c571017d998b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.336659 4789 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/55cac199-8203-40b3-8a4f-c571017d998b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.336675 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wdcz2\" (UniqueName: \"kubernetes.io/projected/55cac199-8203-40b3-8a4f-c571017d998b-kube-api-access-wdcz2\") on node \"crc\" DevicePath \"\"" Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.336688 4789 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/55cac199-8203-40b3-8a4f-c571017d998b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.336701 4789 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/55cac199-8203-40b3-8a4f-c571017d998b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.336716 4789 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/55cac199-8203-40b3-8a4f-c571017d998b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.436265 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-8445cf6b-czcxm" Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.612857 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-8445cf6b-czcxm"] Nov 22 08:01:04 crc kubenswrapper[4789]: W1122 08:01:04.620472 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf5a3a04f_033f_45ca_b676_e72936d5d92c.slice/crio-e3738245ac9889ccf1781ba4d862977e991c4112ce9204855b0a47e3bcfdf1de WatchSource:0}: Error finding container e3738245ac9889ccf1781ba4d862977e991c4112ce9204855b0a47e3bcfdf1de: Status 404 returned error can't find the container with id e3738245ac9889ccf1781ba4d862977e991c4112ce9204855b0a47e3bcfdf1de Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.633238 4789 generic.go:334] "Generic (PLEG): container finished" podID="55cac199-8203-40b3-8a4f-c571017d998b" containerID="a9caf00327bdcea8f671019a3e4ffeb6fb58a2fb6bc5f3641a2a64cdfadfb78d" exitCode=0 Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.633292 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-wnfzj" event={"ID":"55cac199-8203-40b3-8a4f-c571017d998b","Type":"ContainerDied","Data":"a9caf00327bdcea8f671019a3e4ffeb6fb58a2fb6bc5f3641a2a64cdfadfb78d"} Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.633314 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-wnfzj" Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.633348 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-wnfzj" event={"ID":"55cac199-8203-40b3-8a4f-c571017d998b","Type":"ContainerDied","Data":"e7832ac8d5e80153fd0aacde4e65f0d420c572e913313f1ea089032b5310a7ea"} Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.633366 4789 scope.go:117] "RemoveContainer" containerID="a9caf00327bdcea8f671019a3e4ffeb6fb58a2fb6bc5f3641a2a64cdfadfb78d" Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.634453 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-8445cf6b-czcxm" event={"ID":"f5a3a04f-033f-45ca-b676-e72936d5d92c","Type":"ContainerStarted","Data":"e3738245ac9889ccf1781ba4d862977e991c4112ce9204855b0a47e3bcfdf1de"} Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.648440 4789 scope.go:117] "RemoveContainer" containerID="a9caf00327bdcea8f671019a3e4ffeb6fb58a2fb6bc5f3641a2a64cdfadfb78d" Nov 22 08:01:04 crc kubenswrapper[4789]: E1122 08:01:04.649235 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a9caf00327bdcea8f671019a3e4ffeb6fb58a2fb6bc5f3641a2a64cdfadfb78d\": container with ID starting with a9caf00327bdcea8f671019a3e4ffeb6fb58a2fb6bc5f3641a2a64cdfadfb78d not found: ID does not exist" containerID="a9caf00327bdcea8f671019a3e4ffeb6fb58a2fb6bc5f3641a2a64cdfadfb78d" Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.649270 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a9caf00327bdcea8f671019a3e4ffeb6fb58a2fb6bc5f3641a2a64cdfadfb78d"} err="failed to get container status \"a9caf00327bdcea8f671019a3e4ffeb6fb58a2fb6bc5f3641a2a64cdfadfb78d\": rpc error: code = NotFound desc = could not find container \"a9caf00327bdcea8f671019a3e4ffeb6fb58a2fb6bc5f3641a2a64cdfadfb78d\": container with ID starting with a9caf00327bdcea8f671019a3e4ffeb6fb58a2fb6bc5f3641a2a64cdfadfb78d not found: ID does not exist" Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.660140 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-wnfzj"] Nov 22 08:01:04 crc kubenswrapper[4789]: I1122 08:01:04.663214 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-wnfzj"] Nov 22 08:01:05 crc kubenswrapper[4789]: I1122 08:01:05.641418 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-8445cf6b-czcxm" event={"ID":"f5a3a04f-033f-45ca-b676-e72936d5d92c","Type":"ContainerStarted","Data":"0caff534ce52d00900f46342864cc75f2ca8fd6f9f38cf36c4cbdac0cbdbd916"} Nov 22 08:01:05 crc kubenswrapper[4789]: I1122 08:01:05.642026 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-8445cf6b-czcxm" Nov 22 08:01:05 crc kubenswrapper[4789]: I1122 08:01:05.646309 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-8445cf6b-czcxm" Nov 22 08:01:05 crc kubenswrapper[4789]: I1122 08:01:05.662291 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-8445cf6b-czcxm" podStartSLOduration=27.662270339 podStartE2EDuration="27.662270339s" podCreationTimestamp="2025-11-22 08:00:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:01:05.660267034 +0000 UTC m=+279.894667317" watchObservedRunningTime="2025-11-22 08:01:05.662270339 +0000 UTC m=+279.896670622" Nov 22 08:01:05 crc kubenswrapper[4789]: I1122 08:01:05.973003 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="55cac199-8203-40b3-8a4f-c571017d998b" path="/var/lib/kubelet/pods/55cac199-8203-40b3-8a4f-c571017d998b/volumes" Nov 22 08:02:35 crc kubenswrapper[4789]: I1122 08:02:35.371916 4789 patch_prober.go:28] interesting pod/machine-config-daemon-pxvrm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 08:02:35 crc kubenswrapper[4789]: I1122 08:02:35.372738 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 08:03:05 crc kubenswrapper[4789]: I1122 08:03:05.372595 4789 patch_prober.go:28] interesting pod/machine-config-daemon-pxvrm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 08:03:05 crc kubenswrapper[4789]: I1122 08:03:05.373309 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 08:03:13 crc kubenswrapper[4789]: I1122 08:03:13.841074 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-dqrkt"] Nov 22 08:03:13 crc kubenswrapper[4789]: I1122 08:03:13.842455 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-dqrkt" Nov 22 08:03:13 crc kubenswrapper[4789]: I1122 08:03:13.858303 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-dqrkt"] Nov 22 08:03:13 crc kubenswrapper[4789]: I1122 08:03:13.986658 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ef72538b-7fc4-4b77-82b2-582a5487566b-bound-sa-token\") pod \"image-registry-66df7c8f76-dqrkt\" (UID: \"ef72538b-7fc4-4b77-82b2-582a5487566b\") " pod="openshift-image-registry/image-registry-66df7c8f76-dqrkt" Nov 22 08:03:13 crc kubenswrapper[4789]: I1122 08:03:13.986716 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/ef72538b-7fc4-4b77-82b2-582a5487566b-registry-certificates\") pod \"image-registry-66df7c8f76-dqrkt\" (UID: \"ef72538b-7fc4-4b77-82b2-582a5487566b\") " pod="openshift-image-registry/image-registry-66df7c8f76-dqrkt" Nov 22 08:03:13 crc kubenswrapper[4789]: I1122 08:03:13.986827 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/ef72538b-7fc4-4b77-82b2-582a5487566b-installation-pull-secrets\") pod \"image-registry-66df7c8f76-dqrkt\" (UID: \"ef72538b-7fc4-4b77-82b2-582a5487566b\") " pod="openshift-image-registry/image-registry-66df7c8f76-dqrkt" Nov 22 08:03:13 crc kubenswrapper[4789]: I1122 08:03:13.986861 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ef72538b-7fc4-4b77-82b2-582a5487566b-trusted-ca\") pod \"image-registry-66df7c8f76-dqrkt\" (UID: \"ef72538b-7fc4-4b77-82b2-582a5487566b\") " pod="openshift-image-registry/image-registry-66df7c8f76-dqrkt" Nov 22 08:03:13 crc kubenswrapper[4789]: I1122 08:03:13.986882 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/ef72538b-7fc4-4b77-82b2-582a5487566b-ca-trust-extracted\") pod \"image-registry-66df7c8f76-dqrkt\" (UID: \"ef72538b-7fc4-4b77-82b2-582a5487566b\") " pod="openshift-image-registry/image-registry-66df7c8f76-dqrkt" Nov 22 08:03:13 crc kubenswrapper[4789]: I1122 08:03:13.986914 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-prk7c\" (UniqueName: \"kubernetes.io/projected/ef72538b-7fc4-4b77-82b2-582a5487566b-kube-api-access-prk7c\") pod \"image-registry-66df7c8f76-dqrkt\" (UID: \"ef72538b-7fc4-4b77-82b2-582a5487566b\") " pod="openshift-image-registry/image-registry-66df7c8f76-dqrkt" Nov 22 08:03:13 crc kubenswrapper[4789]: I1122 08:03:13.986945 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-dqrkt\" (UID: \"ef72538b-7fc4-4b77-82b2-582a5487566b\") " pod="openshift-image-registry/image-registry-66df7c8f76-dqrkt" Nov 22 08:03:13 crc kubenswrapper[4789]: I1122 08:03:13.986986 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/ef72538b-7fc4-4b77-82b2-582a5487566b-registry-tls\") pod \"image-registry-66df7c8f76-dqrkt\" (UID: \"ef72538b-7fc4-4b77-82b2-582a5487566b\") " pod="openshift-image-registry/image-registry-66df7c8f76-dqrkt" Nov 22 08:03:14 crc kubenswrapper[4789]: I1122 08:03:14.008705 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-dqrkt\" (UID: \"ef72538b-7fc4-4b77-82b2-582a5487566b\") " pod="openshift-image-registry/image-registry-66df7c8f76-dqrkt" Nov 22 08:03:14 crc kubenswrapper[4789]: I1122 08:03:14.087789 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/ef72538b-7fc4-4b77-82b2-582a5487566b-installation-pull-secrets\") pod \"image-registry-66df7c8f76-dqrkt\" (UID: \"ef72538b-7fc4-4b77-82b2-582a5487566b\") " pod="openshift-image-registry/image-registry-66df7c8f76-dqrkt" Nov 22 08:03:14 crc kubenswrapper[4789]: I1122 08:03:14.087840 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ef72538b-7fc4-4b77-82b2-582a5487566b-trusted-ca\") pod \"image-registry-66df7c8f76-dqrkt\" (UID: \"ef72538b-7fc4-4b77-82b2-582a5487566b\") " pod="openshift-image-registry/image-registry-66df7c8f76-dqrkt" Nov 22 08:03:14 crc kubenswrapper[4789]: I1122 08:03:14.087857 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/ef72538b-7fc4-4b77-82b2-582a5487566b-ca-trust-extracted\") pod \"image-registry-66df7c8f76-dqrkt\" (UID: \"ef72538b-7fc4-4b77-82b2-582a5487566b\") " pod="openshift-image-registry/image-registry-66df7c8f76-dqrkt" Nov 22 08:03:14 crc kubenswrapper[4789]: I1122 08:03:14.087880 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-prk7c\" (UniqueName: \"kubernetes.io/projected/ef72538b-7fc4-4b77-82b2-582a5487566b-kube-api-access-prk7c\") pod \"image-registry-66df7c8f76-dqrkt\" (UID: \"ef72538b-7fc4-4b77-82b2-582a5487566b\") " pod="openshift-image-registry/image-registry-66df7c8f76-dqrkt" Nov 22 08:03:14 crc kubenswrapper[4789]: I1122 08:03:14.087906 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/ef72538b-7fc4-4b77-82b2-582a5487566b-registry-tls\") pod \"image-registry-66df7c8f76-dqrkt\" (UID: \"ef72538b-7fc4-4b77-82b2-582a5487566b\") " pod="openshift-image-registry/image-registry-66df7c8f76-dqrkt" Nov 22 08:03:14 crc kubenswrapper[4789]: I1122 08:03:14.087942 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ef72538b-7fc4-4b77-82b2-582a5487566b-bound-sa-token\") pod \"image-registry-66df7c8f76-dqrkt\" (UID: \"ef72538b-7fc4-4b77-82b2-582a5487566b\") " pod="openshift-image-registry/image-registry-66df7c8f76-dqrkt" Nov 22 08:03:14 crc kubenswrapper[4789]: I1122 08:03:14.087971 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/ef72538b-7fc4-4b77-82b2-582a5487566b-registry-certificates\") pod \"image-registry-66df7c8f76-dqrkt\" (UID: \"ef72538b-7fc4-4b77-82b2-582a5487566b\") " pod="openshift-image-registry/image-registry-66df7c8f76-dqrkt" Nov 22 08:03:14 crc kubenswrapper[4789]: I1122 08:03:14.089051 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/ef72538b-7fc4-4b77-82b2-582a5487566b-ca-trust-extracted\") pod \"image-registry-66df7c8f76-dqrkt\" (UID: \"ef72538b-7fc4-4b77-82b2-582a5487566b\") " pod="openshift-image-registry/image-registry-66df7c8f76-dqrkt" Nov 22 08:03:14 crc kubenswrapper[4789]: I1122 08:03:14.089506 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/ef72538b-7fc4-4b77-82b2-582a5487566b-registry-certificates\") pod \"image-registry-66df7c8f76-dqrkt\" (UID: \"ef72538b-7fc4-4b77-82b2-582a5487566b\") " pod="openshift-image-registry/image-registry-66df7c8f76-dqrkt" Nov 22 08:03:14 crc kubenswrapper[4789]: I1122 08:03:14.090719 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ef72538b-7fc4-4b77-82b2-582a5487566b-trusted-ca\") pod \"image-registry-66df7c8f76-dqrkt\" (UID: \"ef72538b-7fc4-4b77-82b2-582a5487566b\") " pod="openshift-image-registry/image-registry-66df7c8f76-dqrkt" Nov 22 08:03:14 crc kubenswrapper[4789]: I1122 08:03:14.097377 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/ef72538b-7fc4-4b77-82b2-582a5487566b-installation-pull-secrets\") pod \"image-registry-66df7c8f76-dqrkt\" (UID: \"ef72538b-7fc4-4b77-82b2-582a5487566b\") " pod="openshift-image-registry/image-registry-66df7c8f76-dqrkt" Nov 22 08:03:14 crc kubenswrapper[4789]: I1122 08:03:14.098034 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/ef72538b-7fc4-4b77-82b2-582a5487566b-registry-tls\") pod \"image-registry-66df7c8f76-dqrkt\" (UID: \"ef72538b-7fc4-4b77-82b2-582a5487566b\") " pod="openshift-image-registry/image-registry-66df7c8f76-dqrkt" Nov 22 08:03:14 crc kubenswrapper[4789]: I1122 08:03:14.110355 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ef72538b-7fc4-4b77-82b2-582a5487566b-bound-sa-token\") pod \"image-registry-66df7c8f76-dqrkt\" (UID: \"ef72538b-7fc4-4b77-82b2-582a5487566b\") " pod="openshift-image-registry/image-registry-66df7c8f76-dqrkt" Nov 22 08:03:14 crc kubenswrapper[4789]: I1122 08:03:14.113800 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-prk7c\" (UniqueName: \"kubernetes.io/projected/ef72538b-7fc4-4b77-82b2-582a5487566b-kube-api-access-prk7c\") pod \"image-registry-66df7c8f76-dqrkt\" (UID: \"ef72538b-7fc4-4b77-82b2-582a5487566b\") " pod="openshift-image-registry/image-registry-66df7c8f76-dqrkt" Nov 22 08:03:14 crc kubenswrapper[4789]: I1122 08:03:14.157052 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-dqrkt" Nov 22 08:03:14 crc kubenswrapper[4789]: I1122 08:03:14.376104 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-dqrkt"] Nov 22 08:03:15 crc kubenswrapper[4789]: I1122 08:03:15.386631 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-dqrkt" event={"ID":"ef72538b-7fc4-4b77-82b2-582a5487566b","Type":"ContainerStarted","Data":"92a01593ea1d61b4b88d434f63b3acae828db7ce83d92dd7936ffc1061b3a359"} Nov 22 08:03:15 crc kubenswrapper[4789]: I1122 08:03:15.387013 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-dqrkt" Nov 22 08:03:15 crc kubenswrapper[4789]: I1122 08:03:15.387025 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-dqrkt" event={"ID":"ef72538b-7fc4-4b77-82b2-582a5487566b","Type":"ContainerStarted","Data":"eb1735b302c9d154292cc59e8a71907bf7cb076e38eaaa1cc3d77586582a4c7e"} Nov 22 08:03:15 crc kubenswrapper[4789]: I1122 08:03:15.406826 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-dqrkt" podStartSLOduration=2.406805808 podStartE2EDuration="2.406805808s" podCreationTimestamp="2025-11-22 08:03:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:03:15.402961061 +0000 UTC m=+409.637361334" watchObservedRunningTime="2025-11-22 08:03:15.406805808 +0000 UTC m=+409.641206081" Nov 22 08:03:34 crc kubenswrapper[4789]: I1122 08:03:34.162600 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-dqrkt" Nov 22 08:03:34 crc kubenswrapper[4789]: I1122 08:03:34.214185 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-l4q7j"] Nov 22 08:03:35 crc kubenswrapper[4789]: I1122 08:03:35.371896 4789 patch_prober.go:28] interesting pod/machine-config-daemon-pxvrm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 08:03:35 crc kubenswrapper[4789]: I1122 08:03:35.372208 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 08:03:35 crc kubenswrapper[4789]: I1122 08:03:35.372253 4789 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" Nov 22 08:03:35 crc kubenswrapper[4789]: I1122 08:03:35.372795 4789 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1297ba3dc6bf83e1607e5d32fcba3c36d955882d84a195c466995f25a0f928f8"} pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 22 08:03:35 crc kubenswrapper[4789]: I1122 08:03:35.372867 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" containerID="cri-o://1297ba3dc6bf83e1607e5d32fcba3c36d955882d84a195c466995f25a0f928f8" gracePeriod=600 Nov 22 08:03:36 crc kubenswrapper[4789]: I1122 08:03:36.491903 4789 generic.go:334] "Generic (PLEG): container finished" podID="58aa8071-7ecc-4692-8789-2db04ec70647" containerID="1297ba3dc6bf83e1607e5d32fcba3c36d955882d84a195c466995f25a0f928f8" exitCode=0 Nov 22 08:03:36 crc kubenswrapper[4789]: I1122 08:03:36.491992 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" event={"ID":"58aa8071-7ecc-4692-8789-2db04ec70647","Type":"ContainerDied","Data":"1297ba3dc6bf83e1607e5d32fcba3c36d955882d84a195c466995f25a0f928f8"} Nov 22 08:03:36 crc kubenswrapper[4789]: I1122 08:03:36.492316 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" event={"ID":"58aa8071-7ecc-4692-8789-2db04ec70647","Type":"ContainerStarted","Data":"5afa28473bdb65dc9e867328a0ad6d3bb782e8cf253ef73d039405b797171f7a"} Nov 22 08:03:36 crc kubenswrapper[4789]: I1122 08:03:36.492346 4789 scope.go:117] "RemoveContainer" containerID="8f2ef4eceebb3af5fedfd9bd9b5e59a01bbba909dc1d449d76acf701bd82428a" Nov 22 08:03:59 crc kubenswrapper[4789]: I1122 08:03:59.255509 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" podUID="b80556f6-5b32-486d-8459-f17bf90c4e24" containerName="registry" containerID="cri-o://4035555c9cf0e724ce2aa131cf4571b2740d4a9f623cc09f130d9eef32829946" gracePeriod=30 Nov 22 08:03:59 crc kubenswrapper[4789]: I1122 08:03:59.626899 4789 generic.go:334] "Generic (PLEG): container finished" podID="b80556f6-5b32-486d-8459-f17bf90c4e24" containerID="4035555c9cf0e724ce2aa131cf4571b2740d4a9f623cc09f130d9eef32829946" exitCode=0 Nov 22 08:03:59 crc kubenswrapper[4789]: I1122 08:03:59.627037 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" event={"ID":"b80556f6-5b32-486d-8459-f17bf90c4e24","Type":"ContainerDied","Data":"4035555c9cf0e724ce2aa131cf4571b2740d4a9f623cc09f130d9eef32829946"} Nov 22 08:04:00 crc kubenswrapper[4789]: I1122 08:04:00.104575 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 08:04:00 crc kubenswrapper[4789]: I1122 08:04:00.201659 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fc9n7\" (UniqueName: \"kubernetes.io/projected/b80556f6-5b32-486d-8459-f17bf90c4e24-kube-api-access-fc9n7\") pod \"b80556f6-5b32-486d-8459-f17bf90c4e24\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " Nov 22 08:04:00 crc kubenswrapper[4789]: I1122 08:04:00.201741 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/b80556f6-5b32-486d-8459-f17bf90c4e24-ca-trust-extracted\") pod \"b80556f6-5b32-486d-8459-f17bf90c4e24\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " Nov 22 08:04:00 crc kubenswrapper[4789]: I1122 08:04:00.201987 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"b80556f6-5b32-486d-8459-f17bf90c4e24\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " Nov 22 08:04:00 crc kubenswrapper[4789]: I1122 08:04:00.202022 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b80556f6-5b32-486d-8459-f17bf90c4e24-bound-sa-token\") pod \"b80556f6-5b32-486d-8459-f17bf90c4e24\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " Nov 22 08:04:00 crc kubenswrapper[4789]: I1122 08:04:00.202054 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/b80556f6-5b32-486d-8459-f17bf90c4e24-installation-pull-secrets\") pod \"b80556f6-5b32-486d-8459-f17bf90c4e24\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " Nov 22 08:04:00 crc kubenswrapper[4789]: I1122 08:04:00.202084 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b80556f6-5b32-486d-8459-f17bf90c4e24-trusted-ca\") pod \"b80556f6-5b32-486d-8459-f17bf90c4e24\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " Nov 22 08:04:00 crc kubenswrapper[4789]: I1122 08:04:00.202109 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/b80556f6-5b32-486d-8459-f17bf90c4e24-registry-certificates\") pod \"b80556f6-5b32-486d-8459-f17bf90c4e24\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " Nov 22 08:04:00 crc kubenswrapper[4789]: I1122 08:04:00.202168 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/b80556f6-5b32-486d-8459-f17bf90c4e24-registry-tls\") pod \"b80556f6-5b32-486d-8459-f17bf90c4e24\" (UID: \"b80556f6-5b32-486d-8459-f17bf90c4e24\") " Nov 22 08:04:00 crc kubenswrapper[4789]: I1122 08:04:00.204237 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b80556f6-5b32-486d-8459-f17bf90c4e24-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "b80556f6-5b32-486d-8459-f17bf90c4e24" (UID: "b80556f6-5b32-486d-8459-f17bf90c4e24"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:04:00 crc kubenswrapper[4789]: I1122 08:04:00.204357 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b80556f6-5b32-486d-8459-f17bf90c4e24-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "b80556f6-5b32-486d-8459-f17bf90c4e24" (UID: "b80556f6-5b32-486d-8459-f17bf90c4e24"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:04:00 crc kubenswrapper[4789]: I1122 08:04:00.207915 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b80556f6-5b32-486d-8459-f17bf90c4e24-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "b80556f6-5b32-486d-8459-f17bf90c4e24" (UID: "b80556f6-5b32-486d-8459-f17bf90c4e24"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:04:00 crc kubenswrapper[4789]: I1122 08:04:00.208308 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b80556f6-5b32-486d-8459-f17bf90c4e24-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "b80556f6-5b32-486d-8459-f17bf90c4e24" (UID: "b80556f6-5b32-486d-8459-f17bf90c4e24"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:04:00 crc kubenswrapper[4789]: I1122 08:04:00.208714 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b80556f6-5b32-486d-8459-f17bf90c4e24-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "b80556f6-5b32-486d-8459-f17bf90c4e24" (UID: "b80556f6-5b32-486d-8459-f17bf90c4e24"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:04:00 crc kubenswrapper[4789]: I1122 08:04:00.210650 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b80556f6-5b32-486d-8459-f17bf90c4e24-kube-api-access-fc9n7" (OuterVolumeSpecName: "kube-api-access-fc9n7") pod "b80556f6-5b32-486d-8459-f17bf90c4e24" (UID: "b80556f6-5b32-486d-8459-f17bf90c4e24"). InnerVolumeSpecName "kube-api-access-fc9n7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:04:00 crc kubenswrapper[4789]: I1122 08:04:00.213564 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "b80556f6-5b32-486d-8459-f17bf90c4e24" (UID: "b80556f6-5b32-486d-8459-f17bf90c4e24"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 22 08:04:00 crc kubenswrapper[4789]: I1122 08:04:00.221313 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b80556f6-5b32-486d-8459-f17bf90c4e24-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "b80556f6-5b32-486d-8459-f17bf90c4e24" (UID: "b80556f6-5b32-486d-8459-f17bf90c4e24"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:04:00 crc kubenswrapper[4789]: I1122 08:04:00.304111 4789 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/b80556f6-5b32-486d-8459-f17bf90c4e24-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Nov 22 08:04:00 crc kubenswrapper[4789]: I1122 08:04:00.304143 4789 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b80556f6-5b32-486d-8459-f17bf90c4e24-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 22 08:04:00 crc kubenswrapper[4789]: I1122 08:04:00.304153 4789 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/b80556f6-5b32-486d-8459-f17bf90c4e24-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Nov 22 08:04:00 crc kubenswrapper[4789]: I1122 08:04:00.304162 4789 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b80556f6-5b32-486d-8459-f17bf90c4e24-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 22 08:04:00 crc kubenswrapper[4789]: I1122 08:04:00.304171 4789 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/b80556f6-5b32-486d-8459-f17bf90c4e24-registry-certificates\") on node \"crc\" DevicePath \"\"" Nov 22 08:04:00 crc kubenswrapper[4789]: I1122 08:04:00.304188 4789 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/b80556f6-5b32-486d-8459-f17bf90c4e24-registry-tls\") on node \"crc\" DevicePath \"\"" Nov 22 08:04:00 crc kubenswrapper[4789]: I1122 08:04:00.304197 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fc9n7\" (UniqueName: \"kubernetes.io/projected/b80556f6-5b32-486d-8459-f17bf90c4e24-kube-api-access-fc9n7\") on node \"crc\" DevicePath \"\"" Nov 22 08:04:00 crc kubenswrapper[4789]: I1122 08:04:00.632325 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" event={"ID":"b80556f6-5b32-486d-8459-f17bf90c4e24","Type":"ContainerDied","Data":"4f6b1ab90738b52b1c9f0ee01449e02798904a435fa37930c555980928ef50ee"} Nov 22 08:04:00 crc kubenswrapper[4789]: I1122 08:04:00.632375 4789 scope.go:117] "RemoveContainer" containerID="4035555c9cf0e724ce2aa131cf4571b2740d4a9f623cc09f130d9eef32829946" Nov 22 08:04:00 crc kubenswrapper[4789]: I1122 08:04:00.632386 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-l4q7j" Nov 22 08:04:00 crc kubenswrapper[4789]: I1122 08:04:00.654613 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-l4q7j"] Nov 22 08:04:00 crc kubenswrapper[4789]: I1122 08:04:00.659302 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-l4q7j"] Nov 22 08:04:01 crc kubenswrapper[4789]: I1122 08:04:01.971165 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b80556f6-5b32-486d-8459-f17bf90c4e24" path="/var/lib/kubelet/pods/b80556f6-5b32-486d-8459-f17bf90c4e24/volumes" Nov 22 08:05:35 crc kubenswrapper[4789]: I1122 08:05:35.372334 4789 patch_prober.go:28] interesting pod/machine-config-daemon-pxvrm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 08:05:35 crc kubenswrapper[4789]: I1122 08:05:35.372791 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 08:06:05 crc kubenswrapper[4789]: I1122 08:06:05.372294 4789 patch_prober.go:28] interesting pod/machine-config-daemon-pxvrm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 08:06:05 crc kubenswrapper[4789]: I1122 08:06:05.372923 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 08:06:35 crc kubenswrapper[4789]: I1122 08:06:35.372639 4789 patch_prober.go:28] interesting pod/machine-config-daemon-pxvrm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 08:06:35 crc kubenswrapper[4789]: I1122 08:06:35.373533 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 08:06:35 crc kubenswrapper[4789]: I1122 08:06:35.373592 4789 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" Nov 22 08:06:35 crc kubenswrapper[4789]: I1122 08:06:35.374242 4789 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5afa28473bdb65dc9e867328a0ad6d3bb782e8cf253ef73d039405b797171f7a"} pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 22 08:06:35 crc kubenswrapper[4789]: I1122 08:06:35.374310 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" containerID="cri-o://5afa28473bdb65dc9e867328a0ad6d3bb782e8cf253ef73d039405b797171f7a" gracePeriod=600 Nov 22 08:06:36 crc kubenswrapper[4789]: I1122 08:06:36.549531 4789 generic.go:334] "Generic (PLEG): container finished" podID="58aa8071-7ecc-4692-8789-2db04ec70647" containerID="5afa28473bdb65dc9e867328a0ad6d3bb782e8cf253ef73d039405b797171f7a" exitCode=0 Nov 22 08:06:36 crc kubenswrapper[4789]: I1122 08:06:36.549568 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" event={"ID":"58aa8071-7ecc-4692-8789-2db04ec70647","Type":"ContainerDied","Data":"5afa28473bdb65dc9e867328a0ad6d3bb782e8cf253ef73d039405b797171f7a"} Nov 22 08:06:36 crc kubenswrapper[4789]: I1122 08:06:36.549870 4789 scope.go:117] "RemoveContainer" containerID="1297ba3dc6bf83e1607e5d32fcba3c36d955882d84a195c466995f25a0f928f8" Nov 22 08:06:37 crc kubenswrapper[4789]: I1122 08:06:37.556088 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" event={"ID":"58aa8071-7ecc-4692-8789-2db04ec70647","Type":"ContainerStarted","Data":"447a15a0714201a673fc1d0cc8a8f355942fbc82479335216d2bf7c26b973475"} Nov 22 08:07:29 crc kubenswrapper[4789]: I1122 08:07:29.977136 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-4s4sh"] Nov 22 08:07:29 crc kubenswrapper[4789]: I1122 08:07:29.977917 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-4s4sh" podUID="23967110-3d55-4767-9fa9-230138c92e42" containerName="controller-manager" containerID="cri-o://a635c33cabfc04dd7ad49a8cbc51985f284a781bc5913fb9856dae4473104bce" gracePeriod=30 Nov 22 08:07:30 crc kubenswrapper[4789]: I1122 08:07:30.068146 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-fpw88"] Nov 22 08:07:30 crc kubenswrapper[4789]: I1122 08:07:30.068347 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fpw88" podUID="d4d9ab07-53b6-4aef-bcac-c0140d870370" containerName="route-controller-manager" containerID="cri-o://855070c6843b68f8678ae5705eeb15a639521a84fa045cd6e5e761646a80d863" gracePeriod=30 Nov 22 08:07:30 crc kubenswrapper[4789]: I1122 08:07:30.309017 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-4s4sh" Nov 22 08:07:30 crc kubenswrapper[4789]: I1122 08:07:30.394012 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fpw88" Nov 22 08:07:30 crc kubenswrapper[4789]: I1122 08:07:30.419486 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/23967110-3d55-4767-9fa9-230138c92e42-config\") pod \"23967110-3d55-4767-9fa9-230138c92e42\" (UID: \"23967110-3d55-4767-9fa9-230138c92e42\") " Nov 22 08:07:30 crc kubenswrapper[4789]: I1122 08:07:30.419560 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/23967110-3d55-4767-9fa9-230138c92e42-client-ca\") pod \"23967110-3d55-4767-9fa9-230138c92e42\" (UID: \"23967110-3d55-4767-9fa9-230138c92e42\") " Nov 22 08:07:30 crc kubenswrapper[4789]: I1122 08:07:30.419581 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/23967110-3d55-4767-9fa9-230138c92e42-proxy-ca-bundles\") pod \"23967110-3d55-4767-9fa9-230138c92e42\" (UID: \"23967110-3d55-4767-9fa9-230138c92e42\") " Nov 22 08:07:30 crc kubenswrapper[4789]: I1122 08:07:30.419611 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bb5dq\" (UniqueName: \"kubernetes.io/projected/23967110-3d55-4767-9fa9-230138c92e42-kube-api-access-bb5dq\") pod \"23967110-3d55-4767-9fa9-230138c92e42\" (UID: \"23967110-3d55-4767-9fa9-230138c92e42\") " Nov 22 08:07:30 crc kubenswrapper[4789]: I1122 08:07:30.419635 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/23967110-3d55-4767-9fa9-230138c92e42-serving-cert\") pod \"23967110-3d55-4767-9fa9-230138c92e42\" (UID: \"23967110-3d55-4767-9fa9-230138c92e42\") " Nov 22 08:07:30 crc kubenswrapper[4789]: I1122 08:07:30.421183 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/23967110-3d55-4767-9fa9-230138c92e42-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "23967110-3d55-4767-9fa9-230138c92e42" (UID: "23967110-3d55-4767-9fa9-230138c92e42"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:07:30 crc kubenswrapper[4789]: I1122 08:07:30.421274 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/23967110-3d55-4767-9fa9-230138c92e42-config" (OuterVolumeSpecName: "config") pod "23967110-3d55-4767-9fa9-230138c92e42" (UID: "23967110-3d55-4767-9fa9-230138c92e42"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:07:30 crc kubenswrapper[4789]: I1122 08:07:30.421785 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/23967110-3d55-4767-9fa9-230138c92e42-client-ca" (OuterVolumeSpecName: "client-ca") pod "23967110-3d55-4767-9fa9-230138c92e42" (UID: "23967110-3d55-4767-9fa9-230138c92e42"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:07:30 crc kubenswrapper[4789]: I1122 08:07:30.425178 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/23967110-3d55-4767-9fa9-230138c92e42-kube-api-access-bb5dq" (OuterVolumeSpecName: "kube-api-access-bb5dq") pod "23967110-3d55-4767-9fa9-230138c92e42" (UID: "23967110-3d55-4767-9fa9-230138c92e42"). InnerVolumeSpecName "kube-api-access-bb5dq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:07:30 crc kubenswrapper[4789]: I1122 08:07:30.425191 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/23967110-3d55-4767-9fa9-230138c92e42-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "23967110-3d55-4767-9fa9-230138c92e42" (UID: "23967110-3d55-4767-9fa9-230138c92e42"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:07:30 crc kubenswrapper[4789]: I1122 08:07:30.521166 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d4d9ab07-53b6-4aef-bcac-c0140d870370-serving-cert\") pod \"d4d9ab07-53b6-4aef-bcac-c0140d870370\" (UID: \"d4d9ab07-53b6-4aef-bcac-c0140d870370\") " Nov 22 08:07:30 crc kubenswrapper[4789]: I1122 08:07:30.521278 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d4d9ab07-53b6-4aef-bcac-c0140d870370-client-ca\") pod \"d4d9ab07-53b6-4aef-bcac-c0140d870370\" (UID: \"d4d9ab07-53b6-4aef-bcac-c0140d870370\") " Nov 22 08:07:30 crc kubenswrapper[4789]: I1122 08:07:30.521306 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bpjnq\" (UniqueName: \"kubernetes.io/projected/d4d9ab07-53b6-4aef-bcac-c0140d870370-kube-api-access-bpjnq\") pod \"d4d9ab07-53b6-4aef-bcac-c0140d870370\" (UID: \"d4d9ab07-53b6-4aef-bcac-c0140d870370\") " Nov 22 08:07:30 crc kubenswrapper[4789]: I1122 08:07:30.521335 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d4d9ab07-53b6-4aef-bcac-c0140d870370-config\") pod \"d4d9ab07-53b6-4aef-bcac-c0140d870370\" (UID: \"d4d9ab07-53b6-4aef-bcac-c0140d870370\") " Nov 22 08:07:30 crc kubenswrapper[4789]: I1122 08:07:30.521610 4789 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/23967110-3d55-4767-9fa9-230138c92e42-config\") on node \"crc\" DevicePath \"\"" Nov 22 08:07:30 crc kubenswrapper[4789]: I1122 08:07:30.521628 4789 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/23967110-3d55-4767-9fa9-230138c92e42-client-ca\") on node \"crc\" DevicePath \"\"" Nov 22 08:07:30 crc kubenswrapper[4789]: I1122 08:07:30.521636 4789 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/23967110-3d55-4767-9fa9-230138c92e42-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 22 08:07:30 crc kubenswrapper[4789]: I1122 08:07:30.521648 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bb5dq\" (UniqueName: \"kubernetes.io/projected/23967110-3d55-4767-9fa9-230138c92e42-kube-api-access-bb5dq\") on node \"crc\" DevicePath \"\"" Nov 22 08:07:30 crc kubenswrapper[4789]: I1122 08:07:30.521656 4789 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/23967110-3d55-4767-9fa9-230138c92e42-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 08:07:30 crc kubenswrapper[4789]: I1122 08:07:30.522354 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d4d9ab07-53b6-4aef-bcac-c0140d870370-config" (OuterVolumeSpecName: "config") pod "d4d9ab07-53b6-4aef-bcac-c0140d870370" (UID: "d4d9ab07-53b6-4aef-bcac-c0140d870370"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:07:30 crc kubenswrapper[4789]: I1122 08:07:30.522345 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d4d9ab07-53b6-4aef-bcac-c0140d870370-client-ca" (OuterVolumeSpecName: "client-ca") pod "d4d9ab07-53b6-4aef-bcac-c0140d870370" (UID: "d4d9ab07-53b6-4aef-bcac-c0140d870370"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:07:30 crc kubenswrapper[4789]: I1122 08:07:30.524797 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4d9ab07-53b6-4aef-bcac-c0140d870370-kube-api-access-bpjnq" (OuterVolumeSpecName: "kube-api-access-bpjnq") pod "d4d9ab07-53b6-4aef-bcac-c0140d870370" (UID: "d4d9ab07-53b6-4aef-bcac-c0140d870370"). InnerVolumeSpecName "kube-api-access-bpjnq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:07:30 crc kubenswrapper[4789]: I1122 08:07:30.525406 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4d9ab07-53b6-4aef-bcac-c0140d870370-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "d4d9ab07-53b6-4aef-bcac-c0140d870370" (UID: "d4d9ab07-53b6-4aef-bcac-c0140d870370"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:07:30 crc kubenswrapper[4789]: I1122 08:07:30.622706 4789 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d4d9ab07-53b6-4aef-bcac-c0140d870370-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 08:07:30 crc kubenswrapper[4789]: I1122 08:07:30.622838 4789 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d4d9ab07-53b6-4aef-bcac-c0140d870370-client-ca\") on node \"crc\" DevicePath \"\"" Nov 22 08:07:30 crc kubenswrapper[4789]: I1122 08:07:30.622854 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bpjnq\" (UniqueName: \"kubernetes.io/projected/d4d9ab07-53b6-4aef-bcac-c0140d870370-kube-api-access-bpjnq\") on node \"crc\" DevicePath \"\"" Nov 22 08:07:30 crc kubenswrapper[4789]: I1122 08:07:30.622869 4789 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d4d9ab07-53b6-4aef-bcac-c0140d870370-config\") on node \"crc\" DevicePath \"\"" Nov 22 08:07:30 crc kubenswrapper[4789]: I1122 08:07:30.828311 4789 generic.go:334] "Generic (PLEG): container finished" podID="23967110-3d55-4767-9fa9-230138c92e42" containerID="a635c33cabfc04dd7ad49a8cbc51985f284a781bc5913fb9856dae4473104bce" exitCode=0 Nov 22 08:07:30 crc kubenswrapper[4789]: I1122 08:07:30.828375 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-4s4sh" Nov 22 08:07:30 crc kubenswrapper[4789]: I1122 08:07:30.828388 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-4s4sh" event={"ID":"23967110-3d55-4767-9fa9-230138c92e42","Type":"ContainerDied","Data":"a635c33cabfc04dd7ad49a8cbc51985f284a781bc5913fb9856dae4473104bce"} Nov 22 08:07:30 crc kubenswrapper[4789]: I1122 08:07:30.828418 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-4s4sh" event={"ID":"23967110-3d55-4767-9fa9-230138c92e42","Type":"ContainerDied","Data":"26d3983f74d131481af9bae3a85a2475bc0cb6288922a712da2ae1e6904adf38"} Nov 22 08:07:30 crc kubenswrapper[4789]: I1122 08:07:30.828437 4789 scope.go:117] "RemoveContainer" containerID="a635c33cabfc04dd7ad49a8cbc51985f284a781bc5913fb9856dae4473104bce" Nov 22 08:07:30 crc kubenswrapper[4789]: I1122 08:07:30.835183 4789 generic.go:334] "Generic (PLEG): container finished" podID="d4d9ab07-53b6-4aef-bcac-c0140d870370" containerID="855070c6843b68f8678ae5705eeb15a639521a84fa045cd6e5e761646a80d863" exitCode=0 Nov 22 08:07:30 crc kubenswrapper[4789]: I1122 08:07:30.835260 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fpw88" event={"ID":"d4d9ab07-53b6-4aef-bcac-c0140d870370","Type":"ContainerDied","Data":"855070c6843b68f8678ae5705eeb15a639521a84fa045cd6e5e761646a80d863"} Nov 22 08:07:30 crc kubenswrapper[4789]: I1122 08:07:30.835297 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fpw88" event={"ID":"d4d9ab07-53b6-4aef-bcac-c0140d870370","Type":"ContainerDied","Data":"c9c0e86307f5a170b6e76d3806c5edd7f987998f93cbaa73abaafef7adf10ad8"} Nov 22 08:07:30 crc kubenswrapper[4789]: I1122 08:07:30.835388 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fpw88" Nov 22 08:07:30 crc kubenswrapper[4789]: I1122 08:07:30.856043 4789 scope.go:117] "RemoveContainer" containerID="a635c33cabfc04dd7ad49a8cbc51985f284a781bc5913fb9856dae4473104bce" Nov 22 08:07:30 crc kubenswrapper[4789]: I1122 08:07:30.856208 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-4s4sh"] Nov 22 08:07:30 crc kubenswrapper[4789]: E1122 08:07:30.856631 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a635c33cabfc04dd7ad49a8cbc51985f284a781bc5913fb9856dae4473104bce\": container with ID starting with a635c33cabfc04dd7ad49a8cbc51985f284a781bc5913fb9856dae4473104bce not found: ID does not exist" containerID="a635c33cabfc04dd7ad49a8cbc51985f284a781bc5913fb9856dae4473104bce" Nov 22 08:07:30 crc kubenswrapper[4789]: I1122 08:07:30.856742 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a635c33cabfc04dd7ad49a8cbc51985f284a781bc5913fb9856dae4473104bce"} err="failed to get container status \"a635c33cabfc04dd7ad49a8cbc51985f284a781bc5913fb9856dae4473104bce\": rpc error: code = NotFound desc = could not find container \"a635c33cabfc04dd7ad49a8cbc51985f284a781bc5913fb9856dae4473104bce\": container with ID starting with a635c33cabfc04dd7ad49a8cbc51985f284a781bc5913fb9856dae4473104bce not found: ID does not exist" Nov 22 08:07:30 crc kubenswrapper[4789]: I1122 08:07:30.856862 4789 scope.go:117] "RemoveContainer" containerID="855070c6843b68f8678ae5705eeb15a639521a84fa045cd6e5e761646a80d863" Nov 22 08:07:30 crc kubenswrapper[4789]: I1122 08:07:30.860682 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-4s4sh"] Nov 22 08:07:30 crc kubenswrapper[4789]: I1122 08:07:30.864417 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-fpw88"] Nov 22 08:07:30 crc kubenswrapper[4789]: I1122 08:07:30.868230 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-fpw88"] Nov 22 08:07:30 crc kubenswrapper[4789]: I1122 08:07:30.873130 4789 scope.go:117] "RemoveContainer" containerID="855070c6843b68f8678ae5705eeb15a639521a84fa045cd6e5e761646a80d863" Nov 22 08:07:30 crc kubenswrapper[4789]: E1122 08:07:30.873657 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"855070c6843b68f8678ae5705eeb15a639521a84fa045cd6e5e761646a80d863\": container with ID starting with 855070c6843b68f8678ae5705eeb15a639521a84fa045cd6e5e761646a80d863 not found: ID does not exist" containerID="855070c6843b68f8678ae5705eeb15a639521a84fa045cd6e5e761646a80d863" Nov 22 08:07:30 crc kubenswrapper[4789]: I1122 08:07:30.873718 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"855070c6843b68f8678ae5705eeb15a639521a84fa045cd6e5e761646a80d863"} err="failed to get container status \"855070c6843b68f8678ae5705eeb15a639521a84fa045cd6e5e761646a80d863\": rpc error: code = NotFound desc = could not find container \"855070c6843b68f8678ae5705eeb15a639521a84fa045cd6e5e761646a80d863\": container with ID starting with 855070c6843b68f8678ae5705eeb15a639521a84fa045cd6e5e761646a80d863 not found: ID does not exist" Nov 22 08:07:31 crc kubenswrapper[4789]: I1122 08:07:31.296075 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-85465fc465-t5fgl"] Nov 22 08:07:31 crc kubenswrapper[4789]: E1122 08:07:31.296344 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23967110-3d55-4767-9fa9-230138c92e42" containerName="controller-manager" Nov 22 08:07:31 crc kubenswrapper[4789]: I1122 08:07:31.296359 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="23967110-3d55-4767-9fa9-230138c92e42" containerName="controller-manager" Nov 22 08:07:31 crc kubenswrapper[4789]: E1122 08:07:31.296370 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4d9ab07-53b6-4aef-bcac-c0140d870370" containerName="route-controller-manager" Nov 22 08:07:31 crc kubenswrapper[4789]: I1122 08:07:31.296380 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4d9ab07-53b6-4aef-bcac-c0140d870370" containerName="route-controller-manager" Nov 22 08:07:31 crc kubenswrapper[4789]: E1122 08:07:31.296409 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b80556f6-5b32-486d-8459-f17bf90c4e24" containerName="registry" Nov 22 08:07:31 crc kubenswrapper[4789]: I1122 08:07:31.296426 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="b80556f6-5b32-486d-8459-f17bf90c4e24" containerName="registry" Nov 22 08:07:31 crc kubenswrapper[4789]: I1122 08:07:31.296559 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4d9ab07-53b6-4aef-bcac-c0140d870370" containerName="route-controller-manager" Nov 22 08:07:31 crc kubenswrapper[4789]: I1122 08:07:31.296841 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="23967110-3d55-4767-9fa9-230138c92e42" containerName="controller-manager" Nov 22 08:07:31 crc kubenswrapper[4789]: I1122 08:07:31.296863 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="b80556f6-5b32-486d-8459-f17bf90c4e24" containerName="registry" Nov 22 08:07:31 crc kubenswrapper[4789]: I1122 08:07:31.297399 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-85465fc465-t5fgl" Nov 22 08:07:31 crc kubenswrapper[4789]: I1122 08:07:31.299209 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-57b5d6c749-7hchc"] Nov 22 08:07:31 crc kubenswrapper[4789]: I1122 08:07:31.299733 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-57b5d6c749-7hchc" Nov 22 08:07:31 crc kubenswrapper[4789]: I1122 08:07:31.300005 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 22 08:07:31 crc kubenswrapper[4789]: I1122 08:07:31.301961 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 22 08:07:31 crc kubenswrapper[4789]: I1122 08:07:31.303621 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 22 08:07:31 crc kubenswrapper[4789]: I1122 08:07:31.303966 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 22 08:07:31 crc kubenswrapper[4789]: I1122 08:07:31.304191 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 22 08:07:31 crc kubenswrapper[4789]: I1122 08:07:31.305138 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 22 08:07:31 crc kubenswrapper[4789]: I1122 08:07:31.305174 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 22 08:07:31 crc kubenswrapper[4789]: I1122 08:07:31.305185 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 22 08:07:31 crc kubenswrapper[4789]: I1122 08:07:31.305247 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 22 08:07:31 crc kubenswrapper[4789]: I1122 08:07:31.310895 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 22 08:07:31 crc kubenswrapper[4789]: I1122 08:07:31.311698 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 22 08:07:31 crc kubenswrapper[4789]: I1122 08:07:31.312017 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 22 08:07:31 crc kubenswrapper[4789]: I1122 08:07:31.312417 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 22 08:07:31 crc kubenswrapper[4789]: I1122 08:07:31.314152 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-85465fc465-t5fgl"] Nov 22 08:07:31 crc kubenswrapper[4789]: I1122 08:07:31.320337 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-57b5d6c749-7hchc"] Nov 22 08:07:31 crc kubenswrapper[4789]: I1122 08:07:31.332086 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a2b482fc-3d5b-49e4-85d4-95e4888cad60-proxy-ca-bundles\") pod \"controller-manager-85465fc465-t5fgl\" (UID: \"a2b482fc-3d5b-49e4-85d4-95e4888cad60\") " pod="openshift-controller-manager/controller-manager-85465fc465-t5fgl" Nov 22 08:07:31 crc kubenswrapper[4789]: I1122 08:07:31.332143 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b5fb3a89-837f-44e5-af9f-11f5263612c7-serving-cert\") pod \"route-controller-manager-57b5d6c749-7hchc\" (UID: \"b5fb3a89-837f-44e5-af9f-11f5263612c7\") " pod="openshift-route-controller-manager/route-controller-manager-57b5d6c749-7hchc" Nov 22 08:07:31 crc kubenswrapper[4789]: I1122 08:07:31.332210 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a2b482fc-3d5b-49e4-85d4-95e4888cad60-config\") pod \"controller-manager-85465fc465-t5fgl\" (UID: \"a2b482fc-3d5b-49e4-85d4-95e4888cad60\") " pod="openshift-controller-manager/controller-manager-85465fc465-t5fgl" Nov 22 08:07:31 crc kubenswrapper[4789]: I1122 08:07:31.332243 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a2b482fc-3d5b-49e4-85d4-95e4888cad60-client-ca\") pod \"controller-manager-85465fc465-t5fgl\" (UID: \"a2b482fc-3d5b-49e4-85d4-95e4888cad60\") " pod="openshift-controller-manager/controller-manager-85465fc465-t5fgl" Nov 22 08:07:31 crc kubenswrapper[4789]: I1122 08:07:31.332286 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b5fb3a89-837f-44e5-af9f-11f5263612c7-client-ca\") pod \"route-controller-manager-57b5d6c749-7hchc\" (UID: \"b5fb3a89-837f-44e5-af9f-11f5263612c7\") " pod="openshift-route-controller-manager/route-controller-manager-57b5d6c749-7hchc" Nov 22 08:07:31 crc kubenswrapper[4789]: I1122 08:07:31.332322 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6hk4v\" (UniqueName: \"kubernetes.io/projected/a2b482fc-3d5b-49e4-85d4-95e4888cad60-kube-api-access-6hk4v\") pod \"controller-manager-85465fc465-t5fgl\" (UID: \"a2b482fc-3d5b-49e4-85d4-95e4888cad60\") " pod="openshift-controller-manager/controller-manager-85465fc465-t5fgl" Nov 22 08:07:31 crc kubenswrapper[4789]: I1122 08:07:31.332367 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a2b482fc-3d5b-49e4-85d4-95e4888cad60-serving-cert\") pod \"controller-manager-85465fc465-t5fgl\" (UID: \"a2b482fc-3d5b-49e4-85d4-95e4888cad60\") " pod="openshift-controller-manager/controller-manager-85465fc465-t5fgl" Nov 22 08:07:31 crc kubenswrapper[4789]: I1122 08:07:31.332397 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p6p8q\" (UniqueName: \"kubernetes.io/projected/b5fb3a89-837f-44e5-af9f-11f5263612c7-kube-api-access-p6p8q\") pod \"route-controller-manager-57b5d6c749-7hchc\" (UID: \"b5fb3a89-837f-44e5-af9f-11f5263612c7\") " pod="openshift-route-controller-manager/route-controller-manager-57b5d6c749-7hchc" Nov 22 08:07:31 crc kubenswrapper[4789]: I1122 08:07:31.332427 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5fb3a89-837f-44e5-af9f-11f5263612c7-config\") pod \"route-controller-manager-57b5d6c749-7hchc\" (UID: \"b5fb3a89-837f-44e5-af9f-11f5263612c7\") " pod="openshift-route-controller-manager/route-controller-manager-57b5d6c749-7hchc" Nov 22 08:07:31 crc kubenswrapper[4789]: I1122 08:07:31.433176 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b5fb3a89-837f-44e5-af9f-11f5263612c7-serving-cert\") pod \"route-controller-manager-57b5d6c749-7hchc\" (UID: \"b5fb3a89-837f-44e5-af9f-11f5263612c7\") " pod="openshift-route-controller-manager/route-controller-manager-57b5d6c749-7hchc" Nov 22 08:07:31 crc kubenswrapper[4789]: I1122 08:07:31.433265 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a2b482fc-3d5b-49e4-85d4-95e4888cad60-config\") pod \"controller-manager-85465fc465-t5fgl\" (UID: \"a2b482fc-3d5b-49e4-85d4-95e4888cad60\") " pod="openshift-controller-manager/controller-manager-85465fc465-t5fgl" Nov 22 08:07:31 crc kubenswrapper[4789]: I1122 08:07:31.433289 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a2b482fc-3d5b-49e4-85d4-95e4888cad60-client-ca\") pod \"controller-manager-85465fc465-t5fgl\" (UID: \"a2b482fc-3d5b-49e4-85d4-95e4888cad60\") " pod="openshift-controller-manager/controller-manager-85465fc465-t5fgl" Nov 22 08:07:31 crc kubenswrapper[4789]: I1122 08:07:31.433319 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b5fb3a89-837f-44e5-af9f-11f5263612c7-client-ca\") pod \"route-controller-manager-57b5d6c749-7hchc\" (UID: \"b5fb3a89-837f-44e5-af9f-11f5263612c7\") " pod="openshift-route-controller-manager/route-controller-manager-57b5d6c749-7hchc" Nov 22 08:07:31 crc kubenswrapper[4789]: I1122 08:07:31.433342 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6hk4v\" (UniqueName: \"kubernetes.io/projected/a2b482fc-3d5b-49e4-85d4-95e4888cad60-kube-api-access-6hk4v\") pod \"controller-manager-85465fc465-t5fgl\" (UID: \"a2b482fc-3d5b-49e4-85d4-95e4888cad60\") " pod="openshift-controller-manager/controller-manager-85465fc465-t5fgl" Nov 22 08:07:31 crc kubenswrapper[4789]: I1122 08:07:31.433371 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a2b482fc-3d5b-49e4-85d4-95e4888cad60-serving-cert\") pod \"controller-manager-85465fc465-t5fgl\" (UID: \"a2b482fc-3d5b-49e4-85d4-95e4888cad60\") " pod="openshift-controller-manager/controller-manager-85465fc465-t5fgl" Nov 22 08:07:31 crc kubenswrapper[4789]: I1122 08:07:31.433389 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p6p8q\" (UniqueName: \"kubernetes.io/projected/b5fb3a89-837f-44e5-af9f-11f5263612c7-kube-api-access-p6p8q\") pod \"route-controller-manager-57b5d6c749-7hchc\" (UID: \"b5fb3a89-837f-44e5-af9f-11f5263612c7\") " pod="openshift-route-controller-manager/route-controller-manager-57b5d6c749-7hchc" Nov 22 08:07:31 crc kubenswrapper[4789]: I1122 08:07:31.433406 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5fb3a89-837f-44e5-af9f-11f5263612c7-config\") pod \"route-controller-manager-57b5d6c749-7hchc\" (UID: \"b5fb3a89-837f-44e5-af9f-11f5263612c7\") " pod="openshift-route-controller-manager/route-controller-manager-57b5d6c749-7hchc" Nov 22 08:07:31 crc kubenswrapper[4789]: I1122 08:07:31.433434 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a2b482fc-3d5b-49e4-85d4-95e4888cad60-proxy-ca-bundles\") pod \"controller-manager-85465fc465-t5fgl\" (UID: \"a2b482fc-3d5b-49e4-85d4-95e4888cad60\") " pod="openshift-controller-manager/controller-manager-85465fc465-t5fgl" Nov 22 08:07:31 crc kubenswrapper[4789]: I1122 08:07:31.434662 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a2b482fc-3d5b-49e4-85d4-95e4888cad60-proxy-ca-bundles\") pod \"controller-manager-85465fc465-t5fgl\" (UID: \"a2b482fc-3d5b-49e4-85d4-95e4888cad60\") " pod="openshift-controller-manager/controller-manager-85465fc465-t5fgl" Nov 22 08:07:31 crc kubenswrapper[4789]: I1122 08:07:31.434833 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a2b482fc-3d5b-49e4-85d4-95e4888cad60-config\") pod \"controller-manager-85465fc465-t5fgl\" (UID: \"a2b482fc-3d5b-49e4-85d4-95e4888cad60\") " pod="openshift-controller-manager/controller-manager-85465fc465-t5fgl" Nov 22 08:07:31 crc kubenswrapper[4789]: I1122 08:07:31.435009 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b5fb3a89-837f-44e5-af9f-11f5263612c7-client-ca\") pod \"route-controller-manager-57b5d6c749-7hchc\" (UID: \"b5fb3a89-837f-44e5-af9f-11f5263612c7\") " pod="openshift-route-controller-manager/route-controller-manager-57b5d6c749-7hchc" Nov 22 08:07:31 crc kubenswrapper[4789]: I1122 08:07:31.435237 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5fb3a89-837f-44e5-af9f-11f5263612c7-config\") pod \"route-controller-manager-57b5d6c749-7hchc\" (UID: \"b5fb3a89-837f-44e5-af9f-11f5263612c7\") " pod="openshift-route-controller-manager/route-controller-manager-57b5d6c749-7hchc" Nov 22 08:07:31 crc kubenswrapper[4789]: I1122 08:07:31.435559 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a2b482fc-3d5b-49e4-85d4-95e4888cad60-client-ca\") pod \"controller-manager-85465fc465-t5fgl\" (UID: \"a2b482fc-3d5b-49e4-85d4-95e4888cad60\") " pod="openshift-controller-manager/controller-manager-85465fc465-t5fgl" Nov 22 08:07:31 crc kubenswrapper[4789]: I1122 08:07:31.438502 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b5fb3a89-837f-44e5-af9f-11f5263612c7-serving-cert\") pod \"route-controller-manager-57b5d6c749-7hchc\" (UID: \"b5fb3a89-837f-44e5-af9f-11f5263612c7\") " pod="openshift-route-controller-manager/route-controller-manager-57b5d6c749-7hchc" Nov 22 08:07:31 crc kubenswrapper[4789]: I1122 08:07:31.439055 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a2b482fc-3d5b-49e4-85d4-95e4888cad60-serving-cert\") pod \"controller-manager-85465fc465-t5fgl\" (UID: \"a2b482fc-3d5b-49e4-85d4-95e4888cad60\") " pod="openshift-controller-manager/controller-manager-85465fc465-t5fgl" Nov 22 08:07:31 crc kubenswrapper[4789]: I1122 08:07:31.450651 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p6p8q\" (UniqueName: \"kubernetes.io/projected/b5fb3a89-837f-44e5-af9f-11f5263612c7-kube-api-access-p6p8q\") pod \"route-controller-manager-57b5d6c749-7hchc\" (UID: \"b5fb3a89-837f-44e5-af9f-11f5263612c7\") " pod="openshift-route-controller-manager/route-controller-manager-57b5d6c749-7hchc" Nov 22 08:07:31 crc kubenswrapper[4789]: I1122 08:07:31.452148 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6hk4v\" (UniqueName: \"kubernetes.io/projected/a2b482fc-3d5b-49e4-85d4-95e4888cad60-kube-api-access-6hk4v\") pod \"controller-manager-85465fc465-t5fgl\" (UID: \"a2b482fc-3d5b-49e4-85d4-95e4888cad60\") " pod="openshift-controller-manager/controller-manager-85465fc465-t5fgl" Nov 22 08:07:31 crc kubenswrapper[4789]: I1122 08:07:31.616650 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-85465fc465-t5fgl" Nov 22 08:07:31 crc kubenswrapper[4789]: I1122 08:07:31.625586 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-57b5d6c749-7hchc" Nov 22 08:07:31 crc kubenswrapper[4789]: I1122 08:07:31.803042 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-85465fc465-t5fgl"] Nov 22 08:07:31 crc kubenswrapper[4789]: I1122 08:07:31.848153 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-57b5d6c749-7hchc"] Nov 22 08:07:31 crc kubenswrapper[4789]: I1122 08:07:31.851355 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-85465fc465-t5fgl" event={"ID":"a2b482fc-3d5b-49e4-85d4-95e4888cad60","Type":"ContainerStarted","Data":"4222cf6a78e72e99a5dc7dbf14d3a5b974061cdf5fda0f0e9f96f3a73caeb30e"} Nov 22 08:07:31 crc kubenswrapper[4789]: I1122 08:07:31.971981 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="23967110-3d55-4767-9fa9-230138c92e42" path="/var/lib/kubelet/pods/23967110-3d55-4767-9fa9-230138c92e42/volumes" Nov 22 08:07:31 crc kubenswrapper[4789]: I1122 08:07:31.972785 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d4d9ab07-53b6-4aef-bcac-c0140d870370" path="/var/lib/kubelet/pods/d4d9ab07-53b6-4aef-bcac-c0140d870370/volumes" Nov 22 08:07:32 crc kubenswrapper[4789]: I1122 08:07:32.862898 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-85465fc465-t5fgl" event={"ID":"a2b482fc-3d5b-49e4-85d4-95e4888cad60","Type":"ContainerStarted","Data":"ef2142a1ceed2f1ea8c281011e7697effb43a37f8ff041ee33cf1393856af3b0"} Nov 22 08:07:32 crc kubenswrapper[4789]: I1122 08:07:32.863417 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-85465fc465-t5fgl" Nov 22 08:07:32 crc kubenswrapper[4789]: I1122 08:07:32.866365 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-57b5d6c749-7hchc" event={"ID":"b5fb3a89-837f-44e5-af9f-11f5263612c7","Type":"ContainerStarted","Data":"464c4c1a0dd78daf6d843eb800af617518682621de4e70ecbcc2407c1640c2df"} Nov 22 08:07:32 crc kubenswrapper[4789]: I1122 08:07:32.866404 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-57b5d6c749-7hchc" event={"ID":"b5fb3a89-837f-44e5-af9f-11f5263612c7","Type":"ContainerStarted","Data":"60bb568d6c756d4e6bdbae8f26d85c5dea2ff62f98813a75565fe6882bdc143a"} Nov 22 08:07:32 crc kubenswrapper[4789]: I1122 08:07:32.866572 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-57b5d6c749-7hchc" Nov 22 08:07:32 crc kubenswrapper[4789]: I1122 08:07:32.871601 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-85465fc465-t5fgl" Nov 22 08:07:32 crc kubenswrapper[4789]: I1122 08:07:32.872217 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-57b5d6c749-7hchc" Nov 22 08:07:32 crc kubenswrapper[4789]: I1122 08:07:32.883413 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-85465fc465-t5fgl" podStartSLOduration=2.883399593 podStartE2EDuration="2.883399593s" podCreationTimestamp="2025-11-22 08:07:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:07:32.882271973 +0000 UTC m=+667.116672256" watchObservedRunningTime="2025-11-22 08:07:32.883399593 +0000 UTC m=+667.117799866" Nov 22 08:07:32 crc kubenswrapper[4789]: I1122 08:07:32.904599 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-57b5d6c749-7hchc" podStartSLOduration=2.9045717570000003 podStartE2EDuration="2.904571757s" podCreationTimestamp="2025-11-22 08:07:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:07:32.896961461 +0000 UTC m=+667.131361734" watchObservedRunningTime="2025-11-22 08:07:32.904571757 +0000 UTC m=+667.138972030" Nov 22 08:07:33 crc kubenswrapper[4789]: I1122 08:07:33.304736 4789 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 22 08:08:41 crc kubenswrapper[4789]: I1122 08:08:41.492370 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-jn69f"] Nov 22 08:08:41 crc kubenswrapper[4789]: I1122 08:08:41.493644 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-jn69f" Nov 22 08:08:41 crc kubenswrapper[4789]: I1122 08:08:41.496290 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Nov 22 08:08:41 crc kubenswrapper[4789]: I1122 08:08:41.496331 4789 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-759dv" Nov 22 08:08:41 crc kubenswrapper[4789]: I1122 08:08:41.496399 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-d6qv9"] Nov 22 08:08:41 crc kubenswrapper[4789]: I1122 08:08:41.497276 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-d6qv9" Nov 22 08:08:41 crc kubenswrapper[4789]: I1122 08:08:41.500838 4789 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-67nlb" Nov 22 08:08:41 crc kubenswrapper[4789]: I1122 08:08:41.501556 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-jn69f"] Nov 22 08:08:41 crc kubenswrapper[4789]: I1122 08:08:41.506500 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-txzv2"] Nov 22 08:08:41 crc kubenswrapper[4789]: I1122 08:08:41.507437 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-txzv2" Nov 22 08:08:41 crc kubenswrapper[4789]: I1122 08:08:41.509955 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-d6qv9"] Nov 22 08:08:41 crc kubenswrapper[4789]: I1122 08:08:41.510462 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Nov 22 08:08:41 crc kubenswrapper[4789]: I1122 08:08:41.511917 4789 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-xmtsn" Nov 22 08:08:41 crc kubenswrapper[4789]: I1122 08:08:41.528517 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-txzv2"] Nov 22 08:08:41 crc kubenswrapper[4789]: I1122 08:08:41.684116 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p6jtj\" (UniqueName: \"kubernetes.io/projected/6dd4b23f-fd7b-4893-850b-ae8d269b4006-kube-api-access-p6jtj\") pod \"cert-manager-5b446d88c5-d6qv9\" (UID: \"6dd4b23f-fd7b-4893-850b-ae8d269b4006\") " pod="cert-manager/cert-manager-5b446d88c5-d6qv9" Nov 22 08:08:41 crc kubenswrapper[4789]: I1122 08:08:41.684179 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wnn7f\" (UniqueName: \"kubernetes.io/projected/ca83ef6b-884d-44a3-9570-8088d240e4e8-kube-api-access-wnn7f\") pod \"cert-manager-webhook-5655c58dd6-txzv2\" (UID: \"ca83ef6b-884d-44a3-9570-8088d240e4e8\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-txzv2" Nov 22 08:08:41 crc kubenswrapper[4789]: I1122 08:08:41.684214 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ks9nw\" (UniqueName: \"kubernetes.io/projected/d8d7bb44-57f4-4daf-a00f-13e0209ecf92-kube-api-access-ks9nw\") pod \"cert-manager-cainjector-7f985d654d-jn69f\" (UID: \"d8d7bb44-57f4-4daf-a00f-13e0209ecf92\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-jn69f" Nov 22 08:08:41 crc kubenswrapper[4789]: I1122 08:08:41.784868 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p6jtj\" (UniqueName: \"kubernetes.io/projected/6dd4b23f-fd7b-4893-850b-ae8d269b4006-kube-api-access-p6jtj\") pod \"cert-manager-5b446d88c5-d6qv9\" (UID: \"6dd4b23f-fd7b-4893-850b-ae8d269b4006\") " pod="cert-manager/cert-manager-5b446d88c5-d6qv9" Nov 22 08:08:41 crc kubenswrapper[4789]: I1122 08:08:41.785121 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wnn7f\" (UniqueName: \"kubernetes.io/projected/ca83ef6b-884d-44a3-9570-8088d240e4e8-kube-api-access-wnn7f\") pod \"cert-manager-webhook-5655c58dd6-txzv2\" (UID: \"ca83ef6b-884d-44a3-9570-8088d240e4e8\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-txzv2" Nov 22 08:08:41 crc kubenswrapper[4789]: I1122 08:08:41.785237 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ks9nw\" (UniqueName: \"kubernetes.io/projected/d8d7bb44-57f4-4daf-a00f-13e0209ecf92-kube-api-access-ks9nw\") pod \"cert-manager-cainjector-7f985d654d-jn69f\" (UID: \"d8d7bb44-57f4-4daf-a00f-13e0209ecf92\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-jn69f" Nov 22 08:08:41 crc kubenswrapper[4789]: I1122 08:08:41.805153 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ks9nw\" (UniqueName: \"kubernetes.io/projected/d8d7bb44-57f4-4daf-a00f-13e0209ecf92-kube-api-access-ks9nw\") pod \"cert-manager-cainjector-7f985d654d-jn69f\" (UID: \"d8d7bb44-57f4-4daf-a00f-13e0209ecf92\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-jn69f" Nov 22 08:08:41 crc kubenswrapper[4789]: I1122 08:08:41.805518 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p6jtj\" (UniqueName: \"kubernetes.io/projected/6dd4b23f-fd7b-4893-850b-ae8d269b4006-kube-api-access-p6jtj\") pod \"cert-manager-5b446d88c5-d6qv9\" (UID: \"6dd4b23f-fd7b-4893-850b-ae8d269b4006\") " pod="cert-manager/cert-manager-5b446d88c5-d6qv9" Nov 22 08:08:41 crc kubenswrapper[4789]: I1122 08:08:41.805857 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wnn7f\" (UniqueName: \"kubernetes.io/projected/ca83ef6b-884d-44a3-9570-8088d240e4e8-kube-api-access-wnn7f\") pod \"cert-manager-webhook-5655c58dd6-txzv2\" (UID: \"ca83ef6b-884d-44a3-9570-8088d240e4e8\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-txzv2" Nov 22 08:08:41 crc kubenswrapper[4789]: I1122 08:08:41.813329 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-jn69f" Nov 22 08:08:41 crc kubenswrapper[4789]: I1122 08:08:41.829325 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-d6qv9" Nov 22 08:08:41 crc kubenswrapper[4789]: I1122 08:08:41.842234 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-txzv2" Nov 22 08:08:42 crc kubenswrapper[4789]: I1122 08:08:42.264470 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-d6qv9"] Nov 22 08:08:42 crc kubenswrapper[4789]: I1122 08:08:42.275129 4789 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 22 08:08:42 crc kubenswrapper[4789]: I1122 08:08:42.307269 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-jn69f"] Nov 22 08:08:42 crc kubenswrapper[4789]: W1122 08:08:42.311208 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd8d7bb44_57f4_4daf_a00f_13e0209ecf92.slice/crio-4967774cadbf84ea280b347e9b8a2d524743ecb70601e7d07f391253878b73e3 WatchSource:0}: Error finding container 4967774cadbf84ea280b347e9b8a2d524743ecb70601e7d07f391253878b73e3: Status 404 returned error can't find the container with id 4967774cadbf84ea280b347e9b8a2d524743ecb70601e7d07f391253878b73e3 Nov 22 08:08:42 crc kubenswrapper[4789]: I1122 08:08:42.338281 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-txzv2"] Nov 22 08:08:42 crc kubenswrapper[4789]: W1122 08:08:42.342685 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podca83ef6b_884d_44a3_9570_8088d240e4e8.slice/crio-30c0d3710c467fa0d10167e58a0da8bc2509a18aa07d2783e20c3ecae0929db4 WatchSource:0}: Error finding container 30c0d3710c467fa0d10167e58a0da8bc2509a18aa07d2783e20c3ecae0929db4: Status 404 returned error can't find the container with id 30c0d3710c467fa0d10167e58a0da8bc2509a18aa07d2783e20c3ecae0929db4 Nov 22 08:08:43 crc kubenswrapper[4789]: I1122 08:08:43.222137 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-jn69f" event={"ID":"d8d7bb44-57f4-4daf-a00f-13e0209ecf92","Type":"ContainerStarted","Data":"4967774cadbf84ea280b347e9b8a2d524743ecb70601e7d07f391253878b73e3"} Nov 22 08:08:43 crc kubenswrapper[4789]: I1122 08:08:43.223054 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-txzv2" event={"ID":"ca83ef6b-884d-44a3-9570-8088d240e4e8","Type":"ContainerStarted","Data":"30c0d3710c467fa0d10167e58a0da8bc2509a18aa07d2783e20c3ecae0929db4"} Nov 22 08:08:43 crc kubenswrapper[4789]: I1122 08:08:43.223780 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-d6qv9" event={"ID":"6dd4b23f-fd7b-4893-850b-ae8d269b4006","Type":"ContainerStarted","Data":"03befa082438430c42fbc19910117174b9b2c9cf2dfc4e6c4fef515841673ba9"} Nov 22 08:08:46 crc kubenswrapper[4789]: I1122 08:08:46.239393 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-d6qv9" event={"ID":"6dd4b23f-fd7b-4893-850b-ae8d269b4006","Type":"ContainerStarted","Data":"c667e2f464f1804fe5a1310878ce94f1c80531884b577249efa2981c5a1e9aaa"} Nov 22 08:08:46 crc kubenswrapper[4789]: I1122 08:08:46.240671 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-jn69f" event={"ID":"d8d7bb44-57f4-4daf-a00f-13e0209ecf92","Type":"ContainerStarted","Data":"4ffcd31f5f9d67f68d24b18b21d922d7ee907fabce6bf94b17a565a8fbc82663"} Nov 22 08:08:46 crc kubenswrapper[4789]: I1122 08:08:46.241777 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-txzv2" event={"ID":"ca83ef6b-884d-44a3-9570-8088d240e4e8","Type":"ContainerStarted","Data":"f4939ec848c07cac72630b3d9de58d925ead6023fae4ebd7c179f51914acb6ae"} Nov 22 08:08:46 crc kubenswrapper[4789]: I1122 08:08:46.241918 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-txzv2" Nov 22 08:08:46 crc kubenswrapper[4789]: I1122 08:08:46.258831 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-d6qv9" podStartSLOduration=2.138814874 podStartE2EDuration="5.258809088s" podCreationTimestamp="2025-11-22 08:08:41 +0000 UTC" firstStartedPulling="2025-11-22 08:08:42.274817981 +0000 UTC m=+736.509218254" lastFinishedPulling="2025-11-22 08:08:45.394812185 +0000 UTC m=+739.629212468" observedRunningTime="2025-11-22 08:08:46.25555403 +0000 UTC m=+740.489954303" watchObservedRunningTime="2025-11-22 08:08:46.258809088 +0000 UTC m=+740.493209381" Nov 22 08:08:46 crc kubenswrapper[4789]: I1122 08:08:46.295580 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-jn69f" podStartSLOduration=2.219012202 podStartE2EDuration="5.295557915s" podCreationTimestamp="2025-11-22 08:08:41 +0000 UTC" firstStartedPulling="2025-11-22 08:08:42.312829474 +0000 UTC m=+736.547229747" lastFinishedPulling="2025-11-22 08:08:45.389375177 +0000 UTC m=+739.623775460" observedRunningTime="2025-11-22 08:08:46.278820401 +0000 UTC m=+740.513220684" watchObservedRunningTime="2025-11-22 08:08:46.295557915 +0000 UTC m=+740.529958188" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.299459 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-txzv2" podStartSLOduration=7.248625792 podStartE2EDuration="10.299434067s" podCreationTimestamp="2025-11-22 08:08:41 +0000 UTC" firstStartedPulling="2025-11-22 08:08:42.345638214 +0000 UTC m=+736.580038487" lastFinishedPulling="2025-11-22 08:08:45.396446479 +0000 UTC m=+739.630846762" observedRunningTime="2025-11-22 08:08:46.296576023 +0000 UTC m=+740.530976296" watchObservedRunningTime="2025-11-22 08:08:51.299434067 +0000 UTC m=+745.533834370" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.303491 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-qmxss"] Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.303893 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" podUID="0cb4cf51-62d6-4f32-a8b4-4116555f3870" containerName="ovn-controller" containerID="cri-o://266632a371608ebd031caa58818577320156f15bcb53b8beb701f785a6b57f5a" gracePeriod=30 Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.304021 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" podUID="0cb4cf51-62d6-4f32-a8b4-4116555f3870" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://a42be6036a735dcc45137b6268cfd915cb077dea0c25fdacc35c454aae4fd3a4" gracePeriod=30 Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.304061 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" podUID="0cb4cf51-62d6-4f32-a8b4-4116555f3870" containerName="kube-rbac-proxy-node" containerID="cri-o://dc46c530762e95475561ad5c2b62c516fd6a204aa8ba1010358a8ad977d02aa1" gracePeriod=30 Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.304093 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" podUID="0cb4cf51-62d6-4f32-a8b4-4116555f3870" containerName="ovn-acl-logging" containerID="cri-o://16f5881811857276dcb1ad168c209a633185f75938f2d03755a8ae41e3b8e528" gracePeriod=30 Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.304024 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" podUID="0cb4cf51-62d6-4f32-a8b4-4116555f3870" containerName="nbdb" containerID="cri-o://3740d073f60216d0f7bcd8e638a89cf9f3fe351cd7c21ffaaa67a65a64546c33" gracePeriod=30 Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.304349 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" podUID="0cb4cf51-62d6-4f32-a8b4-4116555f3870" containerName="sbdb" containerID="cri-o://5fd3bfe1a04103047c632d783bc804e6866ca3c2f012da7653ad20d277c3a9d0" gracePeriod=30 Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.304423 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" podUID="0cb4cf51-62d6-4f32-a8b4-4116555f3870" containerName="northd" containerID="cri-o://f21f329a56796c8ad0319e24f24b6bb5f1d0f34426dd16613b9fe2819b16f69f" gracePeriod=30 Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.348005 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" podUID="0cb4cf51-62d6-4f32-a8b4-4116555f3870" containerName="ovnkube-controller" containerID="cri-o://c99a248e9a7e8df9e1be7cc7742476be75b4344274665da471fbf3022a343f28" gracePeriod=30 Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.628813 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qmxss_0cb4cf51-62d6-4f32-a8b4-4116555f3870/ovn-acl-logging/0.log" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.629369 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qmxss_0cb4cf51-62d6-4f32-a8b4-4116555f3870/ovn-controller/0.log" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.629805 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.641191 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/0cb4cf51-62d6-4f32-a8b4-4116555f3870-host-run-netns\") pod \"0cb4cf51-62d6-4f32-a8b4-4116555f3870\" (UID: \"0cb4cf51-62d6-4f32-a8b4-4116555f3870\") " Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.641257 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t7d59\" (UniqueName: \"kubernetes.io/projected/0cb4cf51-62d6-4f32-a8b4-4116555f3870-kube-api-access-t7d59\") pod \"0cb4cf51-62d6-4f32-a8b4-4116555f3870\" (UID: \"0cb4cf51-62d6-4f32-a8b4-4116555f3870\") " Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.641277 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/0cb4cf51-62d6-4f32-a8b4-4116555f3870-node-log\") pod \"0cb4cf51-62d6-4f32-a8b4-4116555f3870\" (UID: \"0cb4cf51-62d6-4f32-a8b4-4116555f3870\") " Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.641300 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/0cb4cf51-62d6-4f32-a8b4-4116555f3870-host-cni-netd\") pod \"0cb4cf51-62d6-4f32-a8b4-4116555f3870\" (UID: \"0cb4cf51-62d6-4f32-a8b4-4116555f3870\") " Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.641293 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0cb4cf51-62d6-4f32-a8b4-4116555f3870-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "0cb4cf51-62d6-4f32-a8b4-4116555f3870" (UID: "0cb4cf51-62d6-4f32-a8b4-4116555f3870"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.641324 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0cb4cf51-62d6-4f32-a8b4-4116555f3870-host-var-lib-cni-networks-ovn-kubernetes\") pod \"0cb4cf51-62d6-4f32-a8b4-4116555f3870\" (UID: \"0cb4cf51-62d6-4f32-a8b4-4116555f3870\") " Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.641360 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/0cb4cf51-62d6-4f32-a8b4-4116555f3870-run-ovn\") pod \"0cb4cf51-62d6-4f32-a8b4-4116555f3870\" (UID: \"0cb4cf51-62d6-4f32-a8b4-4116555f3870\") " Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.641362 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0cb4cf51-62d6-4f32-a8b4-4116555f3870-node-log" (OuterVolumeSpecName: "node-log") pod "0cb4cf51-62d6-4f32-a8b4-4116555f3870" (UID: "0cb4cf51-62d6-4f32-a8b4-4116555f3870"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.641380 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/0cb4cf51-62d6-4f32-a8b4-4116555f3870-log-socket\") pod \"0cb4cf51-62d6-4f32-a8b4-4116555f3870\" (UID: \"0cb4cf51-62d6-4f32-a8b4-4116555f3870\") " Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.641401 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/0cb4cf51-62d6-4f32-a8b4-4116555f3870-env-overrides\") pod \"0cb4cf51-62d6-4f32-a8b4-4116555f3870\" (UID: \"0cb4cf51-62d6-4f32-a8b4-4116555f3870\") " Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.641440 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/0cb4cf51-62d6-4f32-a8b4-4116555f3870-systemd-units\") pod \"0cb4cf51-62d6-4f32-a8b4-4116555f3870\" (UID: \"0cb4cf51-62d6-4f32-a8b4-4116555f3870\") " Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.641460 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/0cb4cf51-62d6-4f32-a8b4-4116555f3870-ovnkube-script-lib\") pod \"0cb4cf51-62d6-4f32-a8b4-4116555f3870\" (UID: \"0cb4cf51-62d6-4f32-a8b4-4116555f3870\") " Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.641474 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0cb4cf51-62d6-4f32-a8b4-4116555f3870-var-lib-openvswitch\") pod \"0cb4cf51-62d6-4f32-a8b4-4116555f3870\" (UID: \"0cb4cf51-62d6-4f32-a8b4-4116555f3870\") " Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.641493 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/0cb4cf51-62d6-4f32-a8b4-4116555f3870-run-systemd\") pod \"0cb4cf51-62d6-4f32-a8b4-4116555f3870\" (UID: \"0cb4cf51-62d6-4f32-a8b4-4116555f3870\") " Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.641520 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/0cb4cf51-62d6-4f32-a8b4-4116555f3870-host-cni-bin\") pod \"0cb4cf51-62d6-4f32-a8b4-4116555f3870\" (UID: \"0cb4cf51-62d6-4f32-a8b4-4116555f3870\") " Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.641534 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/0cb4cf51-62d6-4f32-a8b4-4116555f3870-host-slash\") pod \"0cb4cf51-62d6-4f32-a8b4-4116555f3870\" (UID: \"0cb4cf51-62d6-4f32-a8b4-4116555f3870\") " Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.641548 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0cb4cf51-62d6-4f32-a8b4-4116555f3870-etc-openvswitch\") pod \"0cb4cf51-62d6-4f32-a8b4-4116555f3870\" (UID: \"0cb4cf51-62d6-4f32-a8b4-4116555f3870\") " Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.641565 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0cb4cf51-62d6-4f32-a8b4-4116555f3870-host-run-ovn-kubernetes\") pod \"0cb4cf51-62d6-4f32-a8b4-4116555f3870\" (UID: \"0cb4cf51-62d6-4f32-a8b4-4116555f3870\") " Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.641581 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/0cb4cf51-62d6-4f32-a8b4-4116555f3870-host-kubelet\") pod \"0cb4cf51-62d6-4f32-a8b4-4116555f3870\" (UID: \"0cb4cf51-62d6-4f32-a8b4-4116555f3870\") " Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.641600 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/0cb4cf51-62d6-4f32-a8b4-4116555f3870-ovn-node-metrics-cert\") pod \"0cb4cf51-62d6-4f32-a8b4-4116555f3870\" (UID: \"0cb4cf51-62d6-4f32-a8b4-4116555f3870\") " Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.641614 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0cb4cf51-62d6-4f32-a8b4-4116555f3870-run-openvswitch\") pod \"0cb4cf51-62d6-4f32-a8b4-4116555f3870\" (UID: \"0cb4cf51-62d6-4f32-a8b4-4116555f3870\") " Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.641630 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/0cb4cf51-62d6-4f32-a8b4-4116555f3870-ovnkube-config\") pod \"0cb4cf51-62d6-4f32-a8b4-4116555f3870\" (UID: \"0cb4cf51-62d6-4f32-a8b4-4116555f3870\") " Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.641909 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0cb4cf51-62d6-4f32-a8b4-4116555f3870-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "0cb4cf51-62d6-4f32-a8b4-4116555f3870" (UID: "0cb4cf51-62d6-4f32-a8b4-4116555f3870"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.641909 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0cb4cf51-62d6-4f32-a8b4-4116555f3870-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "0cb4cf51-62d6-4f32-a8b4-4116555f3870" (UID: "0cb4cf51-62d6-4f32-a8b4-4116555f3870"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.641935 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0cb4cf51-62d6-4f32-a8b4-4116555f3870-host-slash" (OuterVolumeSpecName: "host-slash") pod "0cb4cf51-62d6-4f32-a8b4-4116555f3870" (UID: "0cb4cf51-62d6-4f32-a8b4-4116555f3870"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.641944 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0cb4cf51-62d6-4f32-a8b4-4116555f3870-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "0cb4cf51-62d6-4f32-a8b4-4116555f3870" (UID: "0cb4cf51-62d6-4f32-a8b4-4116555f3870"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.641953 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0cb4cf51-62d6-4f32-a8b4-4116555f3870-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "0cb4cf51-62d6-4f32-a8b4-4116555f3870" (UID: "0cb4cf51-62d6-4f32-a8b4-4116555f3870"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.641966 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0cb4cf51-62d6-4f32-a8b4-4116555f3870-log-socket" (OuterVolumeSpecName: "log-socket") pod "0cb4cf51-62d6-4f32-a8b4-4116555f3870" (UID: "0cb4cf51-62d6-4f32-a8b4-4116555f3870"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.641987 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0cb4cf51-62d6-4f32-a8b4-4116555f3870-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "0cb4cf51-62d6-4f32-a8b4-4116555f3870" (UID: "0cb4cf51-62d6-4f32-a8b4-4116555f3870"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.642010 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0cb4cf51-62d6-4f32-a8b4-4116555f3870-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "0cb4cf51-62d6-4f32-a8b4-4116555f3870" (UID: "0cb4cf51-62d6-4f32-a8b4-4116555f3870"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.642037 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0cb4cf51-62d6-4f32-a8b4-4116555f3870-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "0cb4cf51-62d6-4f32-a8b4-4116555f3870" (UID: "0cb4cf51-62d6-4f32-a8b4-4116555f3870"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.642063 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0cb4cf51-62d6-4f32-a8b4-4116555f3870-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "0cb4cf51-62d6-4f32-a8b4-4116555f3870" (UID: "0cb4cf51-62d6-4f32-a8b4-4116555f3870"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.642093 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0cb4cf51-62d6-4f32-a8b4-4116555f3870-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "0cb4cf51-62d6-4f32-a8b4-4116555f3870" (UID: "0cb4cf51-62d6-4f32-a8b4-4116555f3870"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.642371 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0cb4cf51-62d6-4f32-a8b4-4116555f3870-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "0cb4cf51-62d6-4f32-a8b4-4116555f3870" (UID: "0cb4cf51-62d6-4f32-a8b4-4116555f3870"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.642402 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0cb4cf51-62d6-4f32-a8b4-4116555f3870-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "0cb4cf51-62d6-4f32-a8b4-4116555f3870" (UID: "0cb4cf51-62d6-4f32-a8b4-4116555f3870"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.642409 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0cb4cf51-62d6-4f32-a8b4-4116555f3870-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "0cb4cf51-62d6-4f32-a8b4-4116555f3870" (UID: "0cb4cf51-62d6-4f32-a8b4-4116555f3870"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.642836 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0cb4cf51-62d6-4f32-a8b4-4116555f3870-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "0cb4cf51-62d6-4f32-a8b4-4116555f3870" (UID: "0cb4cf51-62d6-4f32-a8b4-4116555f3870"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.643512 4789 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/0cb4cf51-62d6-4f32-a8b4-4116555f3870-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.643554 4789 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/0cb4cf51-62d6-4f32-a8b4-4116555f3870-systemd-units\") on node \"crc\" DevicePath \"\"" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.643576 4789 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/0cb4cf51-62d6-4f32-a8b4-4116555f3870-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.643599 4789 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0cb4cf51-62d6-4f32-a8b4-4116555f3870-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.643621 4789 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/0cb4cf51-62d6-4f32-a8b4-4116555f3870-host-cni-bin\") on node \"crc\" DevicePath \"\"" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.643641 4789 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/0cb4cf51-62d6-4f32-a8b4-4116555f3870-host-slash\") on node \"crc\" DevicePath \"\"" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.643661 4789 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0cb4cf51-62d6-4f32-a8b4-4116555f3870-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.643682 4789 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0cb4cf51-62d6-4f32-a8b4-4116555f3870-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.643702 4789 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/0cb4cf51-62d6-4f32-a8b4-4116555f3870-host-kubelet\") on node \"crc\" DevicePath \"\"" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.643722 4789 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0cb4cf51-62d6-4f32-a8b4-4116555f3870-run-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.643742 4789 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/0cb4cf51-62d6-4f32-a8b4-4116555f3870-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.643810 4789 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/0cb4cf51-62d6-4f32-a8b4-4116555f3870-host-run-netns\") on node \"crc\" DevicePath \"\"" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.643831 4789 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/0cb4cf51-62d6-4f32-a8b4-4116555f3870-node-log\") on node \"crc\" DevicePath \"\"" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.643853 4789 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/0cb4cf51-62d6-4f32-a8b4-4116555f3870-host-cni-netd\") on node \"crc\" DevicePath \"\"" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.643877 4789 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0cb4cf51-62d6-4f32-a8b4-4116555f3870-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.643899 4789 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/0cb4cf51-62d6-4f32-a8b4-4116555f3870-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.643923 4789 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/0cb4cf51-62d6-4f32-a8b4-4116555f3870-log-socket\") on node \"crc\" DevicePath \"\"" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.647960 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0cb4cf51-62d6-4f32-a8b4-4116555f3870-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "0cb4cf51-62d6-4f32-a8b4-4116555f3870" (UID: "0cb4cf51-62d6-4f32-a8b4-4116555f3870"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.649299 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0cb4cf51-62d6-4f32-a8b4-4116555f3870-kube-api-access-t7d59" (OuterVolumeSpecName: "kube-api-access-t7d59") pod "0cb4cf51-62d6-4f32-a8b4-4116555f3870" (UID: "0cb4cf51-62d6-4f32-a8b4-4116555f3870"). InnerVolumeSpecName "kube-api-access-t7d59". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.666354 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0cb4cf51-62d6-4f32-a8b4-4116555f3870-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "0cb4cf51-62d6-4f32-a8b4-4116555f3870" (UID: "0cb4cf51-62d6-4f32-a8b4-4116555f3870"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.678143 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-vqhqz"] Nov 22 08:08:51 crc kubenswrapper[4789]: E1122 08:08:51.678365 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cb4cf51-62d6-4f32-a8b4-4116555f3870" containerName="ovn-controller" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.678381 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cb4cf51-62d6-4f32-a8b4-4116555f3870" containerName="ovn-controller" Nov 22 08:08:51 crc kubenswrapper[4789]: E1122 08:08:51.678398 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cb4cf51-62d6-4f32-a8b4-4116555f3870" containerName="kubecfg-setup" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.678405 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cb4cf51-62d6-4f32-a8b4-4116555f3870" containerName="kubecfg-setup" Nov 22 08:08:51 crc kubenswrapper[4789]: E1122 08:08:51.678412 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cb4cf51-62d6-4f32-a8b4-4116555f3870" containerName="nbdb" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.678419 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cb4cf51-62d6-4f32-a8b4-4116555f3870" containerName="nbdb" Nov 22 08:08:51 crc kubenswrapper[4789]: E1122 08:08:51.678428 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cb4cf51-62d6-4f32-a8b4-4116555f3870" containerName="ovn-acl-logging" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.678434 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cb4cf51-62d6-4f32-a8b4-4116555f3870" containerName="ovn-acl-logging" Nov 22 08:08:51 crc kubenswrapper[4789]: E1122 08:08:51.678443 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cb4cf51-62d6-4f32-a8b4-4116555f3870" containerName="sbdb" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.678450 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cb4cf51-62d6-4f32-a8b4-4116555f3870" containerName="sbdb" Nov 22 08:08:51 crc kubenswrapper[4789]: E1122 08:08:51.678458 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cb4cf51-62d6-4f32-a8b4-4116555f3870" containerName="northd" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.678463 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cb4cf51-62d6-4f32-a8b4-4116555f3870" containerName="northd" Nov 22 08:08:51 crc kubenswrapper[4789]: E1122 08:08:51.678473 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cb4cf51-62d6-4f32-a8b4-4116555f3870" containerName="kube-rbac-proxy-node" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.678479 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cb4cf51-62d6-4f32-a8b4-4116555f3870" containerName="kube-rbac-proxy-node" Nov 22 08:08:51 crc kubenswrapper[4789]: E1122 08:08:51.678488 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cb4cf51-62d6-4f32-a8b4-4116555f3870" containerName="kube-rbac-proxy-ovn-metrics" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.678494 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cb4cf51-62d6-4f32-a8b4-4116555f3870" containerName="kube-rbac-proxy-ovn-metrics" Nov 22 08:08:51 crc kubenswrapper[4789]: E1122 08:08:51.678506 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cb4cf51-62d6-4f32-a8b4-4116555f3870" containerName="ovnkube-controller" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.678513 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cb4cf51-62d6-4f32-a8b4-4116555f3870" containerName="ovnkube-controller" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.678601 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="0cb4cf51-62d6-4f32-a8b4-4116555f3870" containerName="ovnkube-controller" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.678613 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="0cb4cf51-62d6-4f32-a8b4-4116555f3870" containerName="sbdb" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.678620 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="0cb4cf51-62d6-4f32-a8b4-4116555f3870" containerName="ovn-controller" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.678626 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="0cb4cf51-62d6-4f32-a8b4-4116555f3870" containerName="kube-rbac-proxy-node" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.678634 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="0cb4cf51-62d6-4f32-a8b4-4116555f3870" containerName="kube-rbac-proxy-ovn-metrics" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.678642 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="0cb4cf51-62d6-4f32-a8b4-4116555f3870" containerName="nbdb" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.678650 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="0cb4cf51-62d6-4f32-a8b4-4116555f3870" containerName="northd" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.678656 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="0cb4cf51-62d6-4f32-a8b4-4116555f3870" containerName="ovn-acl-logging" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.680434 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-vqhqz" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.745051 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/7f05fa4c-01d8-4bfd-9def-ed962a22960c-host-run-netns\") pod \"ovnkube-node-vqhqz\" (UID: \"7f05fa4c-01d8-4bfd-9def-ed962a22960c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vqhqz" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.745373 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7f05fa4c-01d8-4bfd-9def-ed962a22960c-var-lib-openvswitch\") pod \"ovnkube-node-vqhqz\" (UID: \"7f05fa4c-01d8-4bfd-9def-ed962a22960c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vqhqz" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.745403 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/7f05fa4c-01d8-4bfd-9def-ed962a22960c-host-cni-bin\") pod \"ovnkube-node-vqhqz\" (UID: \"7f05fa4c-01d8-4bfd-9def-ed962a22960c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vqhqz" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.745424 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/7f05fa4c-01d8-4bfd-9def-ed962a22960c-host-cni-netd\") pod \"ovnkube-node-vqhqz\" (UID: \"7f05fa4c-01d8-4bfd-9def-ed962a22960c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vqhqz" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.745445 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rkv5r\" (UniqueName: \"kubernetes.io/projected/7f05fa4c-01d8-4bfd-9def-ed962a22960c-kube-api-access-rkv5r\") pod \"ovnkube-node-vqhqz\" (UID: \"7f05fa4c-01d8-4bfd-9def-ed962a22960c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vqhqz" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.745472 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/7f05fa4c-01d8-4bfd-9def-ed962a22960c-env-overrides\") pod \"ovnkube-node-vqhqz\" (UID: \"7f05fa4c-01d8-4bfd-9def-ed962a22960c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vqhqz" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.745500 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/7f05fa4c-01d8-4bfd-9def-ed962a22960c-host-slash\") pod \"ovnkube-node-vqhqz\" (UID: \"7f05fa4c-01d8-4bfd-9def-ed962a22960c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vqhqz" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.745523 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/7f05fa4c-01d8-4bfd-9def-ed962a22960c-log-socket\") pod \"ovnkube-node-vqhqz\" (UID: \"7f05fa4c-01d8-4bfd-9def-ed962a22960c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vqhqz" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.745543 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7f05fa4c-01d8-4bfd-9def-ed962a22960c-host-run-ovn-kubernetes\") pod \"ovnkube-node-vqhqz\" (UID: \"7f05fa4c-01d8-4bfd-9def-ed962a22960c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vqhqz" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.745562 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/7f05fa4c-01d8-4bfd-9def-ed962a22960c-ovn-node-metrics-cert\") pod \"ovnkube-node-vqhqz\" (UID: \"7f05fa4c-01d8-4bfd-9def-ed962a22960c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vqhqz" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.745586 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/7f05fa4c-01d8-4bfd-9def-ed962a22960c-systemd-units\") pod \"ovnkube-node-vqhqz\" (UID: \"7f05fa4c-01d8-4bfd-9def-ed962a22960c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vqhqz" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.745610 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7f05fa4c-01d8-4bfd-9def-ed962a22960c-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-vqhqz\" (UID: \"7f05fa4c-01d8-4bfd-9def-ed962a22960c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vqhqz" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.745652 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7f05fa4c-01d8-4bfd-9def-ed962a22960c-etc-openvswitch\") pod \"ovnkube-node-vqhqz\" (UID: \"7f05fa4c-01d8-4bfd-9def-ed962a22960c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vqhqz" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.745673 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7f05fa4c-01d8-4bfd-9def-ed962a22960c-run-openvswitch\") pod \"ovnkube-node-vqhqz\" (UID: \"7f05fa4c-01d8-4bfd-9def-ed962a22960c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vqhqz" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.745696 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/7f05fa4c-01d8-4bfd-9def-ed962a22960c-ovnkube-config\") pod \"ovnkube-node-vqhqz\" (UID: \"7f05fa4c-01d8-4bfd-9def-ed962a22960c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vqhqz" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.745716 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/7f05fa4c-01d8-4bfd-9def-ed962a22960c-run-systemd\") pod \"ovnkube-node-vqhqz\" (UID: \"7f05fa4c-01d8-4bfd-9def-ed962a22960c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vqhqz" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.745735 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/7f05fa4c-01d8-4bfd-9def-ed962a22960c-host-kubelet\") pod \"ovnkube-node-vqhqz\" (UID: \"7f05fa4c-01d8-4bfd-9def-ed962a22960c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vqhqz" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.745780 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/7f05fa4c-01d8-4bfd-9def-ed962a22960c-ovnkube-script-lib\") pod \"ovnkube-node-vqhqz\" (UID: \"7f05fa4c-01d8-4bfd-9def-ed962a22960c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vqhqz" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.745807 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/7f05fa4c-01d8-4bfd-9def-ed962a22960c-node-log\") pod \"ovnkube-node-vqhqz\" (UID: \"7f05fa4c-01d8-4bfd-9def-ed962a22960c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vqhqz" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.745837 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/7f05fa4c-01d8-4bfd-9def-ed962a22960c-run-ovn\") pod \"ovnkube-node-vqhqz\" (UID: \"7f05fa4c-01d8-4bfd-9def-ed962a22960c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vqhqz" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.745877 4789 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/0cb4cf51-62d6-4f32-a8b4-4116555f3870-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.745892 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t7d59\" (UniqueName: \"kubernetes.io/projected/0cb4cf51-62d6-4f32-a8b4-4116555f3870-kube-api-access-t7d59\") on node \"crc\" DevicePath \"\"" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.745904 4789 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/0cb4cf51-62d6-4f32-a8b4-4116555f3870-run-systemd\") on node \"crc\" DevicePath \"\"" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.846285 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-txzv2" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.846447 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7f05fa4c-01d8-4bfd-9def-ed962a22960c-etc-openvswitch\") pod \"ovnkube-node-vqhqz\" (UID: \"7f05fa4c-01d8-4bfd-9def-ed962a22960c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vqhqz" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.846531 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7f05fa4c-01d8-4bfd-9def-ed962a22960c-run-openvswitch\") pod \"ovnkube-node-vqhqz\" (UID: \"7f05fa4c-01d8-4bfd-9def-ed962a22960c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vqhqz" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.846572 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7f05fa4c-01d8-4bfd-9def-ed962a22960c-etc-openvswitch\") pod \"ovnkube-node-vqhqz\" (UID: \"7f05fa4c-01d8-4bfd-9def-ed962a22960c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vqhqz" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.846579 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/7f05fa4c-01d8-4bfd-9def-ed962a22960c-ovnkube-config\") pod \"ovnkube-node-vqhqz\" (UID: \"7f05fa4c-01d8-4bfd-9def-ed962a22960c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vqhqz" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.846640 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/7f05fa4c-01d8-4bfd-9def-ed962a22960c-run-systemd\") pod \"ovnkube-node-vqhqz\" (UID: \"7f05fa4c-01d8-4bfd-9def-ed962a22960c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vqhqz" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.846663 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/7f05fa4c-01d8-4bfd-9def-ed962a22960c-host-kubelet\") pod \"ovnkube-node-vqhqz\" (UID: \"7f05fa4c-01d8-4bfd-9def-ed962a22960c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vqhqz" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.846691 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/7f05fa4c-01d8-4bfd-9def-ed962a22960c-ovnkube-script-lib\") pod \"ovnkube-node-vqhqz\" (UID: \"7f05fa4c-01d8-4bfd-9def-ed962a22960c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vqhqz" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.846700 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7f05fa4c-01d8-4bfd-9def-ed962a22960c-run-openvswitch\") pod \"ovnkube-node-vqhqz\" (UID: \"7f05fa4c-01d8-4bfd-9def-ed962a22960c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vqhqz" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.846723 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/7f05fa4c-01d8-4bfd-9def-ed962a22960c-node-log\") pod \"ovnkube-node-vqhqz\" (UID: \"7f05fa4c-01d8-4bfd-9def-ed962a22960c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vqhqz" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.846788 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/7f05fa4c-01d8-4bfd-9def-ed962a22960c-run-ovn\") pod \"ovnkube-node-vqhqz\" (UID: \"7f05fa4c-01d8-4bfd-9def-ed962a22960c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vqhqz" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.846732 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/7f05fa4c-01d8-4bfd-9def-ed962a22960c-host-kubelet\") pod \"ovnkube-node-vqhqz\" (UID: \"7f05fa4c-01d8-4bfd-9def-ed962a22960c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vqhqz" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.846815 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/7f05fa4c-01d8-4bfd-9def-ed962a22960c-host-run-netns\") pod \"ovnkube-node-vqhqz\" (UID: \"7f05fa4c-01d8-4bfd-9def-ed962a22960c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vqhqz" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.846824 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/7f05fa4c-01d8-4bfd-9def-ed962a22960c-node-log\") pod \"ovnkube-node-vqhqz\" (UID: \"7f05fa4c-01d8-4bfd-9def-ed962a22960c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vqhqz" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.846846 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7f05fa4c-01d8-4bfd-9def-ed962a22960c-var-lib-openvswitch\") pod \"ovnkube-node-vqhqz\" (UID: \"7f05fa4c-01d8-4bfd-9def-ed962a22960c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vqhqz" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.846834 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/7f05fa4c-01d8-4bfd-9def-ed962a22960c-run-ovn\") pod \"ovnkube-node-vqhqz\" (UID: \"7f05fa4c-01d8-4bfd-9def-ed962a22960c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vqhqz" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.846871 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/7f05fa4c-01d8-4bfd-9def-ed962a22960c-host-cni-netd\") pod \"ovnkube-node-vqhqz\" (UID: \"7f05fa4c-01d8-4bfd-9def-ed962a22960c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vqhqz" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.846895 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7f05fa4c-01d8-4bfd-9def-ed962a22960c-var-lib-openvswitch\") pod \"ovnkube-node-vqhqz\" (UID: \"7f05fa4c-01d8-4bfd-9def-ed962a22960c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vqhqz" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.846709 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/7f05fa4c-01d8-4bfd-9def-ed962a22960c-run-systemd\") pod \"ovnkube-node-vqhqz\" (UID: \"7f05fa4c-01d8-4bfd-9def-ed962a22960c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vqhqz" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.846920 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/7f05fa4c-01d8-4bfd-9def-ed962a22960c-host-cni-netd\") pod \"ovnkube-node-vqhqz\" (UID: \"7f05fa4c-01d8-4bfd-9def-ed962a22960c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vqhqz" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.846922 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/7f05fa4c-01d8-4bfd-9def-ed962a22960c-host-cni-bin\") pod \"ovnkube-node-vqhqz\" (UID: \"7f05fa4c-01d8-4bfd-9def-ed962a22960c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vqhqz" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.846862 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/7f05fa4c-01d8-4bfd-9def-ed962a22960c-host-run-netns\") pod \"ovnkube-node-vqhqz\" (UID: \"7f05fa4c-01d8-4bfd-9def-ed962a22960c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vqhqz" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.846897 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/7f05fa4c-01d8-4bfd-9def-ed962a22960c-host-cni-bin\") pod \"ovnkube-node-vqhqz\" (UID: \"7f05fa4c-01d8-4bfd-9def-ed962a22960c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vqhqz" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.847269 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rkv5r\" (UniqueName: \"kubernetes.io/projected/7f05fa4c-01d8-4bfd-9def-ed962a22960c-kube-api-access-rkv5r\") pod \"ovnkube-node-vqhqz\" (UID: \"7f05fa4c-01d8-4bfd-9def-ed962a22960c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vqhqz" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.847306 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/7f05fa4c-01d8-4bfd-9def-ed962a22960c-env-overrides\") pod \"ovnkube-node-vqhqz\" (UID: \"7f05fa4c-01d8-4bfd-9def-ed962a22960c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vqhqz" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.847335 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/7f05fa4c-01d8-4bfd-9def-ed962a22960c-host-slash\") pod \"ovnkube-node-vqhqz\" (UID: \"7f05fa4c-01d8-4bfd-9def-ed962a22960c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vqhqz" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.847359 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/7f05fa4c-01d8-4bfd-9def-ed962a22960c-log-socket\") pod \"ovnkube-node-vqhqz\" (UID: \"7f05fa4c-01d8-4bfd-9def-ed962a22960c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vqhqz" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.847384 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7f05fa4c-01d8-4bfd-9def-ed962a22960c-host-run-ovn-kubernetes\") pod \"ovnkube-node-vqhqz\" (UID: \"7f05fa4c-01d8-4bfd-9def-ed962a22960c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vqhqz" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.847409 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/7f05fa4c-01d8-4bfd-9def-ed962a22960c-ovn-node-metrics-cert\") pod \"ovnkube-node-vqhqz\" (UID: \"7f05fa4c-01d8-4bfd-9def-ed962a22960c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vqhqz" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.847432 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/7f05fa4c-01d8-4bfd-9def-ed962a22960c-systemd-units\") pod \"ovnkube-node-vqhqz\" (UID: \"7f05fa4c-01d8-4bfd-9def-ed962a22960c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vqhqz" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.847436 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7f05fa4c-01d8-4bfd-9def-ed962a22960c-host-run-ovn-kubernetes\") pod \"ovnkube-node-vqhqz\" (UID: \"7f05fa4c-01d8-4bfd-9def-ed962a22960c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vqhqz" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.847444 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/7f05fa4c-01d8-4bfd-9def-ed962a22960c-log-socket\") pod \"ovnkube-node-vqhqz\" (UID: \"7f05fa4c-01d8-4bfd-9def-ed962a22960c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vqhqz" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.847455 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7f05fa4c-01d8-4bfd-9def-ed962a22960c-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-vqhqz\" (UID: \"7f05fa4c-01d8-4bfd-9def-ed962a22960c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vqhqz" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.847490 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/7f05fa4c-01d8-4bfd-9def-ed962a22960c-ovnkube-config\") pod \"ovnkube-node-vqhqz\" (UID: \"7f05fa4c-01d8-4bfd-9def-ed962a22960c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vqhqz" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.847497 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/7f05fa4c-01d8-4bfd-9def-ed962a22960c-systemd-units\") pod \"ovnkube-node-vqhqz\" (UID: \"7f05fa4c-01d8-4bfd-9def-ed962a22960c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vqhqz" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.847538 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7f05fa4c-01d8-4bfd-9def-ed962a22960c-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-vqhqz\" (UID: \"7f05fa4c-01d8-4bfd-9def-ed962a22960c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vqhqz" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.847562 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/7f05fa4c-01d8-4bfd-9def-ed962a22960c-host-slash\") pod \"ovnkube-node-vqhqz\" (UID: \"7f05fa4c-01d8-4bfd-9def-ed962a22960c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vqhqz" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.847732 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/7f05fa4c-01d8-4bfd-9def-ed962a22960c-ovnkube-script-lib\") pod \"ovnkube-node-vqhqz\" (UID: \"7f05fa4c-01d8-4bfd-9def-ed962a22960c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vqhqz" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.848465 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/7f05fa4c-01d8-4bfd-9def-ed962a22960c-env-overrides\") pod \"ovnkube-node-vqhqz\" (UID: \"7f05fa4c-01d8-4bfd-9def-ed962a22960c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vqhqz" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.851916 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/7f05fa4c-01d8-4bfd-9def-ed962a22960c-ovn-node-metrics-cert\") pod \"ovnkube-node-vqhqz\" (UID: \"7f05fa4c-01d8-4bfd-9def-ed962a22960c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vqhqz" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.878240 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rkv5r\" (UniqueName: \"kubernetes.io/projected/7f05fa4c-01d8-4bfd-9def-ed962a22960c-kube-api-access-rkv5r\") pod \"ovnkube-node-vqhqz\" (UID: \"7f05fa4c-01d8-4bfd-9def-ed962a22960c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vqhqz" Nov 22 08:08:51 crc kubenswrapper[4789]: I1122 08:08:51.995050 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-vqhqz" Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.278087 4789 generic.go:334] "Generic (PLEG): container finished" podID="7f05fa4c-01d8-4bfd-9def-ed962a22960c" containerID="5b2a3c6716f7e33bcb3ce2cc3d616cac407c8f7dbeaaf112dbda49b1451f06d3" exitCode=0 Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.278216 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vqhqz" event={"ID":"7f05fa4c-01d8-4bfd-9def-ed962a22960c","Type":"ContainerDied","Data":"5b2a3c6716f7e33bcb3ce2cc3d616cac407c8f7dbeaaf112dbda49b1451f06d3"} Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.278284 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vqhqz" event={"ID":"7f05fa4c-01d8-4bfd-9def-ed962a22960c","Type":"ContainerStarted","Data":"74deed2c075a0262dc521f917163207fdb5874f836a57679abe7b6d069161684"} Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.285814 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qmxss_0cb4cf51-62d6-4f32-a8b4-4116555f3870/ovn-acl-logging/0.log" Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.290526 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qmxss_0cb4cf51-62d6-4f32-a8b4-4116555f3870/ovn-controller/0.log" Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.291279 4789 generic.go:334] "Generic (PLEG): container finished" podID="0cb4cf51-62d6-4f32-a8b4-4116555f3870" containerID="c99a248e9a7e8df9e1be7cc7742476be75b4344274665da471fbf3022a343f28" exitCode=0 Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.291331 4789 generic.go:334] "Generic (PLEG): container finished" podID="0cb4cf51-62d6-4f32-a8b4-4116555f3870" containerID="5fd3bfe1a04103047c632d783bc804e6866ca3c2f012da7653ad20d277c3a9d0" exitCode=0 Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.291345 4789 generic.go:334] "Generic (PLEG): container finished" podID="0cb4cf51-62d6-4f32-a8b4-4116555f3870" containerID="3740d073f60216d0f7bcd8e638a89cf9f3fe351cd7c21ffaaa67a65a64546c33" exitCode=0 Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.291357 4789 generic.go:334] "Generic (PLEG): container finished" podID="0cb4cf51-62d6-4f32-a8b4-4116555f3870" containerID="f21f329a56796c8ad0319e24f24b6bb5f1d0f34426dd16613b9fe2819b16f69f" exitCode=0 Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.291380 4789 generic.go:334] "Generic (PLEG): container finished" podID="0cb4cf51-62d6-4f32-a8b4-4116555f3870" containerID="a42be6036a735dcc45137b6268cfd915cb077dea0c25fdacc35c454aae4fd3a4" exitCode=0 Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.291392 4789 generic.go:334] "Generic (PLEG): container finished" podID="0cb4cf51-62d6-4f32-a8b4-4116555f3870" containerID="dc46c530762e95475561ad5c2b62c516fd6a204aa8ba1010358a8ad977d02aa1" exitCode=0 Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.291403 4789 generic.go:334] "Generic (PLEG): container finished" podID="0cb4cf51-62d6-4f32-a8b4-4116555f3870" containerID="16f5881811857276dcb1ad168c209a633185f75938f2d03755a8ae41e3b8e528" exitCode=143 Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.291418 4789 generic.go:334] "Generic (PLEG): container finished" podID="0cb4cf51-62d6-4f32-a8b4-4116555f3870" containerID="266632a371608ebd031caa58818577320156f15bcb53b8beb701f785a6b57f5a" exitCode=143 Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.291401 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" event={"ID":"0cb4cf51-62d6-4f32-a8b4-4116555f3870","Type":"ContainerDied","Data":"c99a248e9a7e8df9e1be7cc7742476be75b4344274665da471fbf3022a343f28"} Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.291497 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" event={"ID":"0cb4cf51-62d6-4f32-a8b4-4116555f3870","Type":"ContainerDied","Data":"5fd3bfe1a04103047c632d783bc804e6866ca3c2f012da7653ad20d277c3a9d0"} Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.291441 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.291531 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" event={"ID":"0cb4cf51-62d6-4f32-a8b4-4116555f3870","Type":"ContainerDied","Data":"3740d073f60216d0f7bcd8e638a89cf9f3fe351cd7c21ffaaa67a65a64546c33"} Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.291558 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" event={"ID":"0cb4cf51-62d6-4f32-a8b4-4116555f3870","Type":"ContainerDied","Data":"f21f329a56796c8ad0319e24f24b6bb5f1d0f34426dd16613b9fe2819b16f69f"} Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.291586 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" event={"ID":"0cb4cf51-62d6-4f32-a8b4-4116555f3870","Type":"ContainerDied","Data":"a42be6036a735dcc45137b6268cfd915cb077dea0c25fdacc35c454aae4fd3a4"} Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.291618 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" event={"ID":"0cb4cf51-62d6-4f32-a8b4-4116555f3870","Type":"ContainerDied","Data":"dc46c530762e95475561ad5c2b62c516fd6a204aa8ba1010358a8ad977d02aa1"} Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.291643 4789 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"16f5881811857276dcb1ad168c209a633185f75938f2d03755a8ae41e3b8e528"} Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.291667 4789 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"266632a371608ebd031caa58818577320156f15bcb53b8beb701f785a6b57f5a"} Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.291683 4789 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5ce8875ca32df64c155691ab953221d98fdd435cf3fff8df49a2524cb6a8fca3"} Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.291702 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" event={"ID":"0cb4cf51-62d6-4f32-a8b4-4116555f3870","Type":"ContainerDied","Data":"16f5881811857276dcb1ad168c209a633185f75938f2d03755a8ae41e3b8e528"} Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.291726 4789 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c99a248e9a7e8df9e1be7cc7742476be75b4344274665da471fbf3022a343f28"} Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.291743 4789 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5fd3bfe1a04103047c632d783bc804e6866ca3c2f012da7653ad20d277c3a9d0"} Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.291797 4789 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3740d073f60216d0f7bcd8e638a89cf9f3fe351cd7c21ffaaa67a65a64546c33"} Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.291813 4789 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f21f329a56796c8ad0319e24f24b6bb5f1d0f34426dd16613b9fe2819b16f69f"} Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.291830 4789 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a42be6036a735dcc45137b6268cfd915cb077dea0c25fdacc35c454aae4fd3a4"} Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.291847 4789 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"dc46c530762e95475561ad5c2b62c516fd6a204aa8ba1010358a8ad977d02aa1"} Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.291863 4789 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"16f5881811857276dcb1ad168c209a633185f75938f2d03755a8ae41e3b8e528"} Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.291878 4789 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"266632a371608ebd031caa58818577320156f15bcb53b8beb701f785a6b57f5a"} Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.291893 4789 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5ce8875ca32df64c155691ab953221d98fdd435cf3fff8df49a2524cb6a8fca3"} Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.291914 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" event={"ID":"0cb4cf51-62d6-4f32-a8b4-4116555f3870","Type":"ContainerDied","Data":"266632a371608ebd031caa58818577320156f15bcb53b8beb701f785a6b57f5a"} Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.291938 4789 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c99a248e9a7e8df9e1be7cc7742476be75b4344274665da471fbf3022a343f28"} Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.291956 4789 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5fd3bfe1a04103047c632d783bc804e6866ca3c2f012da7653ad20d277c3a9d0"} Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.291971 4789 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3740d073f60216d0f7bcd8e638a89cf9f3fe351cd7c21ffaaa67a65a64546c33"} Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.291982 4789 scope.go:117] "RemoveContainer" containerID="c99a248e9a7e8df9e1be7cc7742476be75b4344274665da471fbf3022a343f28" Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.291987 4789 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f21f329a56796c8ad0319e24f24b6bb5f1d0f34426dd16613b9fe2819b16f69f"} Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.292003 4789 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a42be6036a735dcc45137b6268cfd915cb077dea0c25fdacc35c454aae4fd3a4"} Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.292018 4789 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"dc46c530762e95475561ad5c2b62c516fd6a204aa8ba1010358a8ad977d02aa1"} Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.292033 4789 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"16f5881811857276dcb1ad168c209a633185f75938f2d03755a8ae41e3b8e528"} Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.292053 4789 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"266632a371608ebd031caa58818577320156f15bcb53b8beb701f785a6b57f5a"} Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.292068 4789 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5ce8875ca32df64c155691ab953221d98fdd435cf3fff8df49a2524cb6a8fca3"} Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.292088 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qmxss" event={"ID":"0cb4cf51-62d6-4f32-a8b4-4116555f3870","Type":"ContainerDied","Data":"0d23c5a22905ab46e9052894cc4a7cbfb9d56d8ee528d3286a086123b9218c21"} Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.292112 4789 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c99a248e9a7e8df9e1be7cc7742476be75b4344274665da471fbf3022a343f28"} Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.292129 4789 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5fd3bfe1a04103047c632d783bc804e6866ca3c2f012da7653ad20d277c3a9d0"} Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.292143 4789 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3740d073f60216d0f7bcd8e638a89cf9f3fe351cd7c21ffaaa67a65a64546c33"} Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.292158 4789 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f21f329a56796c8ad0319e24f24b6bb5f1d0f34426dd16613b9fe2819b16f69f"} Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.292172 4789 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a42be6036a735dcc45137b6268cfd915cb077dea0c25fdacc35c454aae4fd3a4"} Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.292186 4789 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"dc46c530762e95475561ad5c2b62c516fd6a204aa8ba1010358a8ad977d02aa1"} Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.292202 4789 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"16f5881811857276dcb1ad168c209a633185f75938f2d03755a8ae41e3b8e528"} Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.292216 4789 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"266632a371608ebd031caa58818577320156f15bcb53b8beb701f785a6b57f5a"} Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.292230 4789 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5ce8875ca32df64c155691ab953221d98fdd435cf3fff8df49a2524cb6a8fca3"} Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.294884 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-rks5j_5613c209-59a0-4b8a-ac8b-5c633bc1bdc4/kube-multus/0.log" Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.294974 4789 generic.go:334] "Generic (PLEG): container finished" podID="5613c209-59a0-4b8a-ac8b-5c633bc1bdc4" containerID="569b7f705aa7fd96b15a7cd1e36497d9f4901040399fa1f49c231af8753a51fc" exitCode=2 Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.295014 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-rks5j" event={"ID":"5613c209-59a0-4b8a-ac8b-5c633bc1bdc4","Type":"ContainerDied","Data":"569b7f705aa7fd96b15a7cd1e36497d9f4901040399fa1f49c231af8753a51fc"} Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.296293 4789 scope.go:117] "RemoveContainer" containerID="569b7f705aa7fd96b15a7cd1e36497d9f4901040399fa1f49c231af8753a51fc" Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.336664 4789 scope.go:117] "RemoveContainer" containerID="5fd3bfe1a04103047c632d783bc804e6866ca3c2f012da7653ad20d277c3a9d0" Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.351564 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-qmxss"] Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.357489 4789 scope.go:117] "RemoveContainer" containerID="3740d073f60216d0f7bcd8e638a89cf9f3fe351cd7c21ffaaa67a65a64546c33" Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.361228 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-qmxss"] Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.394498 4789 scope.go:117] "RemoveContainer" containerID="f21f329a56796c8ad0319e24f24b6bb5f1d0f34426dd16613b9fe2819b16f69f" Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.413120 4789 scope.go:117] "RemoveContainer" containerID="a42be6036a735dcc45137b6268cfd915cb077dea0c25fdacc35c454aae4fd3a4" Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.425807 4789 scope.go:117] "RemoveContainer" containerID="dc46c530762e95475561ad5c2b62c516fd6a204aa8ba1010358a8ad977d02aa1" Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.467591 4789 scope.go:117] "RemoveContainer" containerID="16f5881811857276dcb1ad168c209a633185f75938f2d03755a8ae41e3b8e528" Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.490315 4789 scope.go:117] "RemoveContainer" containerID="266632a371608ebd031caa58818577320156f15bcb53b8beb701f785a6b57f5a" Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.514065 4789 scope.go:117] "RemoveContainer" containerID="5ce8875ca32df64c155691ab953221d98fdd435cf3fff8df49a2524cb6a8fca3" Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.534436 4789 scope.go:117] "RemoveContainer" containerID="c99a248e9a7e8df9e1be7cc7742476be75b4344274665da471fbf3022a343f28" Nov 22 08:08:52 crc kubenswrapper[4789]: E1122 08:08:52.534926 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c99a248e9a7e8df9e1be7cc7742476be75b4344274665da471fbf3022a343f28\": container with ID starting with c99a248e9a7e8df9e1be7cc7742476be75b4344274665da471fbf3022a343f28 not found: ID does not exist" containerID="c99a248e9a7e8df9e1be7cc7742476be75b4344274665da471fbf3022a343f28" Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.534967 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c99a248e9a7e8df9e1be7cc7742476be75b4344274665da471fbf3022a343f28"} err="failed to get container status \"c99a248e9a7e8df9e1be7cc7742476be75b4344274665da471fbf3022a343f28\": rpc error: code = NotFound desc = could not find container \"c99a248e9a7e8df9e1be7cc7742476be75b4344274665da471fbf3022a343f28\": container with ID starting with c99a248e9a7e8df9e1be7cc7742476be75b4344274665da471fbf3022a343f28 not found: ID does not exist" Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.535008 4789 scope.go:117] "RemoveContainer" containerID="5fd3bfe1a04103047c632d783bc804e6866ca3c2f012da7653ad20d277c3a9d0" Nov 22 08:08:52 crc kubenswrapper[4789]: E1122 08:08:52.535315 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5fd3bfe1a04103047c632d783bc804e6866ca3c2f012da7653ad20d277c3a9d0\": container with ID starting with 5fd3bfe1a04103047c632d783bc804e6866ca3c2f012da7653ad20d277c3a9d0 not found: ID does not exist" containerID="5fd3bfe1a04103047c632d783bc804e6866ca3c2f012da7653ad20d277c3a9d0" Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.535349 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5fd3bfe1a04103047c632d783bc804e6866ca3c2f012da7653ad20d277c3a9d0"} err="failed to get container status \"5fd3bfe1a04103047c632d783bc804e6866ca3c2f012da7653ad20d277c3a9d0\": rpc error: code = NotFound desc = could not find container \"5fd3bfe1a04103047c632d783bc804e6866ca3c2f012da7653ad20d277c3a9d0\": container with ID starting with 5fd3bfe1a04103047c632d783bc804e6866ca3c2f012da7653ad20d277c3a9d0 not found: ID does not exist" Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.535382 4789 scope.go:117] "RemoveContainer" containerID="3740d073f60216d0f7bcd8e638a89cf9f3fe351cd7c21ffaaa67a65a64546c33" Nov 22 08:08:52 crc kubenswrapper[4789]: E1122 08:08:52.535646 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3740d073f60216d0f7bcd8e638a89cf9f3fe351cd7c21ffaaa67a65a64546c33\": container with ID starting with 3740d073f60216d0f7bcd8e638a89cf9f3fe351cd7c21ffaaa67a65a64546c33 not found: ID does not exist" containerID="3740d073f60216d0f7bcd8e638a89cf9f3fe351cd7c21ffaaa67a65a64546c33" Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.535672 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3740d073f60216d0f7bcd8e638a89cf9f3fe351cd7c21ffaaa67a65a64546c33"} err="failed to get container status \"3740d073f60216d0f7bcd8e638a89cf9f3fe351cd7c21ffaaa67a65a64546c33\": rpc error: code = NotFound desc = could not find container \"3740d073f60216d0f7bcd8e638a89cf9f3fe351cd7c21ffaaa67a65a64546c33\": container with ID starting with 3740d073f60216d0f7bcd8e638a89cf9f3fe351cd7c21ffaaa67a65a64546c33 not found: ID does not exist" Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.535688 4789 scope.go:117] "RemoveContainer" containerID="f21f329a56796c8ad0319e24f24b6bb5f1d0f34426dd16613b9fe2819b16f69f" Nov 22 08:08:52 crc kubenswrapper[4789]: E1122 08:08:52.535958 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f21f329a56796c8ad0319e24f24b6bb5f1d0f34426dd16613b9fe2819b16f69f\": container with ID starting with f21f329a56796c8ad0319e24f24b6bb5f1d0f34426dd16613b9fe2819b16f69f not found: ID does not exist" containerID="f21f329a56796c8ad0319e24f24b6bb5f1d0f34426dd16613b9fe2819b16f69f" Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.535985 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f21f329a56796c8ad0319e24f24b6bb5f1d0f34426dd16613b9fe2819b16f69f"} err="failed to get container status \"f21f329a56796c8ad0319e24f24b6bb5f1d0f34426dd16613b9fe2819b16f69f\": rpc error: code = NotFound desc = could not find container \"f21f329a56796c8ad0319e24f24b6bb5f1d0f34426dd16613b9fe2819b16f69f\": container with ID starting with f21f329a56796c8ad0319e24f24b6bb5f1d0f34426dd16613b9fe2819b16f69f not found: ID does not exist" Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.536004 4789 scope.go:117] "RemoveContainer" containerID="a42be6036a735dcc45137b6268cfd915cb077dea0c25fdacc35c454aae4fd3a4" Nov 22 08:08:52 crc kubenswrapper[4789]: E1122 08:08:52.536225 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a42be6036a735dcc45137b6268cfd915cb077dea0c25fdacc35c454aae4fd3a4\": container with ID starting with a42be6036a735dcc45137b6268cfd915cb077dea0c25fdacc35c454aae4fd3a4 not found: ID does not exist" containerID="a42be6036a735dcc45137b6268cfd915cb077dea0c25fdacc35c454aae4fd3a4" Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.536246 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a42be6036a735dcc45137b6268cfd915cb077dea0c25fdacc35c454aae4fd3a4"} err="failed to get container status \"a42be6036a735dcc45137b6268cfd915cb077dea0c25fdacc35c454aae4fd3a4\": rpc error: code = NotFound desc = could not find container \"a42be6036a735dcc45137b6268cfd915cb077dea0c25fdacc35c454aae4fd3a4\": container with ID starting with a42be6036a735dcc45137b6268cfd915cb077dea0c25fdacc35c454aae4fd3a4 not found: ID does not exist" Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.536258 4789 scope.go:117] "RemoveContainer" containerID="dc46c530762e95475561ad5c2b62c516fd6a204aa8ba1010358a8ad977d02aa1" Nov 22 08:08:52 crc kubenswrapper[4789]: E1122 08:08:52.536464 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dc46c530762e95475561ad5c2b62c516fd6a204aa8ba1010358a8ad977d02aa1\": container with ID starting with dc46c530762e95475561ad5c2b62c516fd6a204aa8ba1010358a8ad977d02aa1 not found: ID does not exist" containerID="dc46c530762e95475561ad5c2b62c516fd6a204aa8ba1010358a8ad977d02aa1" Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.536488 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc46c530762e95475561ad5c2b62c516fd6a204aa8ba1010358a8ad977d02aa1"} err="failed to get container status \"dc46c530762e95475561ad5c2b62c516fd6a204aa8ba1010358a8ad977d02aa1\": rpc error: code = NotFound desc = could not find container \"dc46c530762e95475561ad5c2b62c516fd6a204aa8ba1010358a8ad977d02aa1\": container with ID starting with dc46c530762e95475561ad5c2b62c516fd6a204aa8ba1010358a8ad977d02aa1 not found: ID does not exist" Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.536503 4789 scope.go:117] "RemoveContainer" containerID="16f5881811857276dcb1ad168c209a633185f75938f2d03755a8ae41e3b8e528" Nov 22 08:08:52 crc kubenswrapper[4789]: E1122 08:08:52.536716 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"16f5881811857276dcb1ad168c209a633185f75938f2d03755a8ae41e3b8e528\": container with ID starting with 16f5881811857276dcb1ad168c209a633185f75938f2d03755a8ae41e3b8e528 not found: ID does not exist" containerID="16f5881811857276dcb1ad168c209a633185f75938f2d03755a8ae41e3b8e528" Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.536737 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"16f5881811857276dcb1ad168c209a633185f75938f2d03755a8ae41e3b8e528"} err="failed to get container status \"16f5881811857276dcb1ad168c209a633185f75938f2d03755a8ae41e3b8e528\": rpc error: code = NotFound desc = could not find container \"16f5881811857276dcb1ad168c209a633185f75938f2d03755a8ae41e3b8e528\": container with ID starting with 16f5881811857276dcb1ad168c209a633185f75938f2d03755a8ae41e3b8e528 not found: ID does not exist" Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.536770 4789 scope.go:117] "RemoveContainer" containerID="266632a371608ebd031caa58818577320156f15bcb53b8beb701f785a6b57f5a" Nov 22 08:08:52 crc kubenswrapper[4789]: E1122 08:08:52.537077 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"266632a371608ebd031caa58818577320156f15bcb53b8beb701f785a6b57f5a\": container with ID starting with 266632a371608ebd031caa58818577320156f15bcb53b8beb701f785a6b57f5a not found: ID does not exist" containerID="266632a371608ebd031caa58818577320156f15bcb53b8beb701f785a6b57f5a" Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.537100 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"266632a371608ebd031caa58818577320156f15bcb53b8beb701f785a6b57f5a"} err="failed to get container status \"266632a371608ebd031caa58818577320156f15bcb53b8beb701f785a6b57f5a\": rpc error: code = NotFound desc = could not find container \"266632a371608ebd031caa58818577320156f15bcb53b8beb701f785a6b57f5a\": container with ID starting with 266632a371608ebd031caa58818577320156f15bcb53b8beb701f785a6b57f5a not found: ID does not exist" Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.537113 4789 scope.go:117] "RemoveContainer" containerID="5ce8875ca32df64c155691ab953221d98fdd435cf3fff8df49a2524cb6a8fca3" Nov 22 08:08:52 crc kubenswrapper[4789]: E1122 08:08:52.537512 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5ce8875ca32df64c155691ab953221d98fdd435cf3fff8df49a2524cb6a8fca3\": container with ID starting with 5ce8875ca32df64c155691ab953221d98fdd435cf3fff8df49a2524cb6a8fca3 not found: ID does not exist" containerID="5ce8875ca32df64c155691ab953221d98fdd435cf3fff8df49a2524cb6a8fca3" Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.537551 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ce8875ca32df64c155691ab953221d98fdd435cf3fff8df49a2524cb6a8fca3"} err="failed to get container status \"5ce8875ca32df64c155691ab953221d98fdd435cf3fff8df49a2524cb6a8fca3\": rpc error: code = NotFound desc = could not find container \"5ce8875ca32df64c155691ab953221d98fdd435cf3fff8df49a2524cb6a8fca3\": container with ID starting with 5ce8875ca32df64c155691ab953221d98fdd435cf3fff8df49a2524cb6a8fca3 not found: ID does not exist" Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.537575 4789 scope.go:117] "RemoveContainer" containerID="c99a248e9a7e8df9e1be7cc7742476be75b4344274665da471fbf3022a343f28" Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.537814 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c99a248e9a7e8df9e1be7cc7742476be75b4344274665da471fbf3022a343f28"} err="failed to get container status \"c99a248e9a7e8df9e1be7cc7742476be75b4344274665da471fbf3022a343f28\": rpc error: code = NotFound desc = could not find container \"c99a248e9a7e8df9e1be7cc7742476be75b4344274665da471fbf3022a343f28\": container with ID starting with c99a248e9a7e8df9e1be7cc7742476be75b4344274665da471fbf3022a343f28 not found: ID does not exist" Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.537834 4789 scope.go:117] "RemoveContainer" containerID="5fd3bfe1a04103047c632d783bc804e6866ca3c2f012da7653ad20d277c3a9d0" Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.538039 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5fd3bfe1a04103047c632d783bc804e6866ca3c2f012da7653ad20d277c3a9d0"} err="failed to get container status \"5fd3bfe1a04103047c632d783bc804e6866ca3c2f012da7653ad20d277c3a9d0\": rpc error: code = NotFound desc = could not find container \"5fd3bfe1a04103047c632d783bc804e6866ca3c2f012da7653ad20d277c3a9d0\": container with ID starting with 5fd3bfe1a04103047c632d783bc804e6866ca3c2f012da7653ad20d277c3a9d0 not found: ID does not exist" Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.538058 4789 scope.go:117] "RemoveContainer" containerID="3740d073f60216d0f7bcd8e638a89cf9f3fe351cd7c21ffaaa67a65a64546c33" Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.538259 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3740d073f60216d0f7bcd8e638a89cf9f3fe351cd7c21ffaaa67a65a64546c33"} err="failed to get container status \"3740d073f60216d0f7bcd8e638a89cf9f3fe351cd7c21ffaaa67a65a64546c33\": rpc error: code = NotFound desc = could not find container \"3740d073f60216d0f7bcd8e638a89cf9f3fe351cd7c21ffaaa67a65a64546c33\": container with ID starting with 3740d073f60216d0f7bcd8e638a89cf9f3fe351cd7c21ffaaa67a65a64546c33 not found: ID does not exist" Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.538286 4789 scope.go:117] "RemoveContainer" containerID="f21f329a56796c8ad0319e24f24b6bb5f1d0f34426dd16613b9fe2819b16f69f" Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.538606 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f21f329a56796c8ad0319e24f24b6bb5f1d0f34426dd16613b9fe2819b16f69f"} err="failed to get container status \"f21f329a56796c8ad0319e24f24b6bb5f1d0f34426dd16613b9fe2819b16f69f\": rpc error: code = NotFound desc = could not find container \"f21f329a56796c8ad0319e24f24b6bb5f1d0f34426dd16613b9fe2819b16f69f\": container with ID starting with f21f329a56796c8ad0319e24f24b6bb5f1d0f34426dd16613b9fe2819b16f69f not found: ID does not exist" Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.538629 4789 scope.go:117] "RemoveContainer" containerID="a42be6036a735dcc45137b6268cfd915cb077dea0c25fdacc35c454aae4fd3a4" Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.538848 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a42be6036a735dcc45137b6268cfd915cb077dea0c25fdacc35c454aae4fd3a4"} err="failed to get container status \"a42be6036a735dcc45137b6268cfd915cb077dea0c25fdacc35c454aae4fd3a4\": rpc error: code = NotFound desc = could not find container \"a42be6036a735dcc45137b6268cfd915cb077dea0c25fdacc35c454aae4fd3a4\": container with ID starting with a42be6036a735dcc45137b6268cfd915cb077dea0c25fdacc35c454aae4fd3a4 not found: ID does not exist" Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.538864 4789 scope.go:117] "RemoveContainer" containerID="dc46c530762e95475561ad5c2b62c516fd6a204aa8ba1010358a8ad977d02aa1" Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.539161 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc46c530762e95475561ad5c2b62c516fd6a204aa8ba1010358a8ad977d02aa1"} err="failed to get container status \"dc46c530762e95475561ad5c2b62c516fd6a204aa8ba1010358a8ad977d02aa1\": rpc error: code = NotFound desc = could not find container \"dc46c530762e95475561ad5c2b62c516fd6a204aa8ba1010358a8ad977d02aa1\": container with ID starting with dc46c530762e95475561ad5c2b62c516fd6a204aa8ba1010358a8ad977d02aa1 not found: ID does not exist" Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.539184 4789 scope.go:117] "RemoveContainer" containerID="16f5881811857276dcb1ad168c209a633185f75938f2d03755a8ae41e3b8e528" Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.539398 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"16f5881811857276dcb1ad168c209a633185f75938f2d03755a8ae41e3b8e528"} err="failed to get container status \"16f5881811857276dcb1ad168c209a633185f75938f2d03755a8ae41e3b8e528\": rpc error: code = NotFound desc = could not find container \"16f5881811857276dcb1ad168c209a633185f75938f2d03755a8ae41e3b8e528\": container with ID starting with 16f5881811857276dcb1ad168c209a633185f75938f2d03755a8ae41e3b8e528 not found: ID does not exist" Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.539431 4789 scope.go:117] "RemoveContainer" containerID="266632a371608ebd031caa58818577320156f15bcb53b8beb701f785a6b57f5a" Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.539842 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"266632a371608ebd031caa58818577320156f15bcb53b8beb701f785a6b57f5a"} err="failed to get container status \"266632a371608ebd031caa58818577320156f15bcb53b8beb701f785a6b57f5a\": rpc error: code = NotFound desc = could not find container \"266632a371608ebd031caa58818577320156f15bcb53b8beb701f785a6b57f5a\": container with ID starting with 266632a371608ebd031caa58818577320156f15bcb53b8beb701f785a6b57f5a not found: ID does not exist" Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.539893 4789 scope.go:117] "RemoveContainer" containerID="5ce8875ca32df64c155691ab953221d98fdd435cf3fff8df49a2524cb6a8fca3" Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.540159 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ce8875ca32df64c155691ab953221d98fdd435cf3fff8df49a2524cb6a8fca3"} err="failed to get container status \"5ce8875ca32df64c155691ab953221d98fdd435cf3fff8df49a2524cb6a8fca3\": rpc error: code = NotFound desc = could not find container \"5ce8875ca32df64c155691ab953221d98fdd435cf3fff8df49a2524cb6a8fca3\": container with ID starting with 5ce8875ca32df64c155691ab953221d98fdd435cf3fff8df49a2524cb6a8fca3 not found: ID does not exist" Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.540180 4789 scope.go:117] "RemoveContainer" containerID="c99a248e9a7e8df9e1be7cc7742476be75b4344274665da471fbf3022a343f28" Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.540354 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c99a248e9a7e8df9e1be7cc7742476be75b4344274665da471fbf3022a343f28"} err="failed to get container status \"c99a248e9a7e8df9e1be7cc7742476be75b4344274665da471fbf3022a343f28\": rpc error: code = NotFound desc = could not find container \"c99a248e9a7e8df9e1be7cc7742476be75b4344274665da471fbf3022a343f28\": container with ID starting with c99a248e9a7e8df9e1be7cc7742476be75b4344274665da471fbf3022a343f28 not found: ID does not exist" Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.540374 4789 scope.go:117] "RemoveContainer" containerID="5fd3bfe1a04103047c632d783bc804e6866ca3c2f012da7653ad20d277c3a9d0" Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.540579 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5fd3bfe1a04103047c632d783bc804e6866ca3c2f012da7653ad20d277c3a9d0"} err="failed to get container status \"5fd3bfe1a04103047c632d783bc804e6866ca3c2f012da7653ad20d277c3a9d0\": rpc error: code = NotFound desc = could not find container \"5fd3bfe1a04103047c632d783bc804e6866ca3c2f012da7653ad20d277c3a9d0\": container with ID starting with 5fd3bfe1a04103047c632d783bc804e6866ca3c2f012da7653ad20d277c3a9d0 not found: ID does not exist" Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.540607 4789 scope.go:117] "RemoveContainer" containerID="3740d073f60216d0f7bcd8e638a89cf9f3fe351cd7c21ffaaa67a65a64546c33" Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.540860 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3740d073f60216d0f7bcd8e638a89cf9f3fe351cd7c21ffaaa67a65a64546c33"} err="failed to get container status \"3740d073f60216d0f7bcd8e638a89cf9f3fe351cd7c21ffaaa67a65a64546c33\": rpc error: code = NotFound desc = could not find container \"3740d073f60216d0f7bcd8e638a89cf9f3fe351cd7c21ffaaa67a65a64546c33\": container with ID starting with 3740d073f60216d0f7bcd8e638a89cf9f3fe351cd7c21ffaaa67a65a64546c33 not found: ID does not exist" Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.540884 4789 scope.go:117] "RemoveContainer" containerID="f21f329a56796c8ad0319e24f24b6bb5f1d0f34426dd16613b9fe2819b16f69f" Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.541113 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f21f329a56796c8ad0319e24f24b6bb5f1d0f34426dd16613b9fe2819b16f69f"} err="failed to get container status \"f21f329a56796c8ad0319e24f24b6bb5f1d0f34426dd16613b9fe2819b16f69f\": rpc error: code = NotFound desc = could not find container \"f21f329a56796c8ad0319e24f24b6bb5f1d0f34426dd16613b9fe2819b16f69f\": container with ID starting with f21f329a56796c8ad0319e24f24b6bb5f1d0f34426dd16613b9fe2819b16f69f not found: ID does not exist" Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.541134 4789 scope.go:117] "RemoveContainer" containerID="a42be6036a735dcc45137b6268cfd915cb077dea0c25fdacc35c454aae4fd3a4" Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.541361 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a42be6036a735dcc45137b6268cfd915cb077dea0c25fdacc35c454aae4fd3a4"} err="failed to get container status \"a42be6036a735dcc45137b6268cfd915cb077dea0c25fdacc35c454aae4fd3a4\": rpc error: code = NotFound desc = could not find container \"a42be6036a735dcc45137b6268cfd915cb077dea0c25fdacc35c454aae4fd3a4\": container with ID starting with a42be6036a735dcc45137b6268cfd915cb077dea0c25fdacc35c454aae4fd3a4 not found: ID does not exist" Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.541382 4789 scope.go:117] "RemoveContainer" containerID="dc46c530762e95475561ad5c2b62c516fd6a204aa8ba1010358a8ad977d02aa1" Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.541579 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc46c530762e95475561ad5c2b62c516fd6a204aa8ba1010358a8ad977d02aa1"} err="failed to get container status \"dc46c530762e95475561ad5c2b62c516fd6a204aa8ba1010358a8ad977d02aa1\": rpc error: code = NotFound desc = could not find container \"dc46c530762e95475561ad5c2b62c516fd6a204aa8ba1010358a8ad977d02aa1\": container with ID starting with dc46c530762e95475561ad5c2b62c516fd6a204aa8ba1010358a8ad977d02aa1 not found: ID does not exist" Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.541599 4789 scope.go:117] "RemoveContainer" containerID="16f5881811857276dcb1ad168c209a633185f75938f2d03755a8ae41e3b8e528" Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.541803 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"16f5881811857276dcb1ad168c209a633185f75938f2d03755a8ae41e3b8e528"} err="failed to get container status \"16f5881811857276dcb1ad168c209a633185f75938f2d03755a8ae41e3b8e528\": rpc error: code = NotFound desc = could not find container \"16f5881811857276dcb1ad168c209a633185f75938f2d03755a8ae41e3b8e528\": container with ID starting with 16f5881811857276dcb1ad168c209a633185f75938f2d03755a8ae41e3b8e528 not found: ID does not exist" Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.541831 4789 scope.go:117] "RemoveContainer" containerID="266632a371608ebd031caa58818577320156f15bcb53b8beb701f785a6b57f5a" Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.542132 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"266632a371608ebd031caa58818577320156f15bcb53b8beb701f785a6b57f5a"} err="failed to get container status \"266632a371608ebd031caa58818577320156f15bcb53b8beb701f785a6b57f5a\": rpc error: code = NotFound desc = could not find container \"266632a371608ebd031caa58818577320156f15bcb53b8beb701f785a6b57f5a\": container with ID starting with 266632a371608ebd031caa58818577320156f15bcb53b8beb701f785a6b57f5a not found: ID does not exist" Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.542155 4789 scope.go:117] "RemoveContainer" containerID="5ce8875ca32df64c155691ab953221d98fdd435cf3fff8df49a2524cb6a8fca3" Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.542544 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ce8875ca32df64c155691ab953221d98fdd435cf3fff8df49a2524cb6a8fca3"} err="failed to get container status \"5ce8875ca32df64c155691ab953221d98fdd435cf3fff8df49a2524cb6a8fca3\": rpc error: code = NotFound desc = could not find container \"5ce8875ca32df64c155691ab953221d98fdd435cf3fff8df49a2524cb6a8fca3\": container with ID starting with 5ce8875ca32df64c155691ab953221d98fdd435cf3fff8df49a2524cb6a8fca3 not found: ID does not exist" Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.542569 4789 scope.go:117] "RemoveContainer" containerID="c99a248e9a7e8df9e1be7cc7742476be75b4344274665da471fbf3022a343f28" Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.542800 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c99a248e9a7e8df9e1be7cc7742476be75b4344274665da471fbf3022a343f28"} err="failed to get container status \"c99a248e9a7e8df9e1be7cc7742476be75b4344274665da471fbf3022a343f28\": rpc error: code = NotFound desc = could not find container \"c99a248e9a7e8df9e1be7cc7742476be75b4344274665da471fbf3022a343f28\": container with ID starting with c99a248e9a7e8df9e1be7cc7742476be75b4344274665da471fbf3022a343f28 not found: ID does not exist" Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.542822 4789 scope.go:117] "RemoveContainer" containerID="5fd3bfe1a04103047c632d783bc804e6866ca3c2f012da7653ad20d277c3a9d0" Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.543098 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5fd3bfe1a04103047c632d783bc804e6866ca3c2f012da7653ad20d277c3a9d0"} err="failed to get container status \"5fd3bfe1a04103047c632d783bc804e6866ca3c2f012da7653ad20d277c3a9d0\": rpc error: code = NotFound desc = could not find container \"5fd3bfe1a04103047c632d783bc804e6866ca3c2f012da7653ad20d277c3a9d0\": container with ID starting with 5fd3bfe1a04103047c632d783bc804e6866ca3c2f012da7653ad20d277c3a9d0 not found: ID does not exist" Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.543133 4789 scope.go:117] "RemoveContainer" containerID="3740d073f60216d0f7bcd8e638a89cf9f3fe351cd7c21ffaaa67a65a64546c33" Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.543330 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3740d073f60216d0f7bcd8e638a89cf9f3fe351cd7c21ffaaa67a65a64546c33"} err="failed to get container status \"3740d073f60216d0f7bcd8e638a89cf9f3fe351cd7c21ffaaa67a65a64546c33\": rpc error: code = NotFound desc = could not find container \"3740d073f60216d0f7bcd8e638a89cf9f3fe351cd7c21ffaaa67a65a64546c33\": container with ID starting with 3740d073f60216d0f7bcd8e638a89cf9f3fe351cd7c21ffaaa67a65a64546c33 not found: ID does not exist" Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.543355 4789 scope.go:117] "RemoveContainer" containerID="f21f329a56796c8ad0319e24f24b6bb5f1d0f34426dd16613b9fe2819b16f69f" Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.543539 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f21f329a56796c8ad0319e24f24b6bb5f1d0f34426dd16613b9fe2819b16f69f"} err="failed to get container status \"f21f329a56796c8ad0319e24f24b6bb5f1d0f34426dd16613b9fe2819b16f69f\": rpc error: code = NotFound desc = could not find container \"f21f329a56796c8ad0319e24f24b6bb5f1d0f34426dd16613b9fe2819b16f69f\": container with ID starting with f21f329a56796c8ad0319e24f24b6bb5f1d0f34426dd16613b9fe2819b16f69f not found: ID does not exist" Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.543560 4789 scope.go:117] "RemoveContainer" containerID="a42be6036a735dcc45137b6268cfd915cb077dea0c25fdacc35c454aae4fd3a4" Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.543770 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a42be6036a735dcc45137b6268cfd915cb077dea0c25fdacc35c454aae4fd3a4"} err="failed to get container status \"a42be6036a735dcc45137b6268cfd915cb077dea0c25fdacc35c454aae4fd3a4\": rpc error: code = NotFound desc = could not find container \"a42be6036a735dcc45137b6268cfd915cb077dea0c25fdacc35c454aae4fd3a4\": container with ID starting with a42be6036a735dcc45137b6268cfd915cb077dea0c25fdacc35c454aae4fd3a4 not found: ID does not exist" Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.543791 4789 scope.go:117] "RemoveContainer" containerID="dc46c530762e95475561ad5c2b62c516fd6a204aa8ba1010358a8ad977d02aa1" Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.543997 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc46c530762e95475561ad5c2b62c516fd6a204aa8ba1010358a8ad977d02aa1"} err="failed to get container status \"dc46c530762e95475561ad5c2b62c516fd6a204aa8ba1010358a8ad977d02aa1\": rpc error: code = NotFound desc = could not find container \"dc46c530762e95475561ad5c2b62c516fd6a204aa8ba1010358a8ad977d02aa1\": container with ID starting with dc46c530762e95475561ad5c2b62c516fd6a204aa8ba1010358a8ad977d02aa1 not found: ID does not exist" Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.544019 4789 scope.go:117] "RemoveContainer" containerID="16f5881811857276dcb1ad168c209a633185f75938f2d03755a8ae41e3b8e528" Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.544213 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"16f5881811857276dcb1ad168c209a633185f75938f2d03755a8ae41e3b8e528"} err="failed to get container status \"16f5881811857276dcb1ad168c209a633185f75938f2d03755a8ae41e3b8e528\": rpc error: code = NotFound desc = could not find container \"16f5881811857276dcb1ad168c209a633185f75938f2d03755a8ae41e3b8e528\": container with ID starting with 16f5881811857276dcb1ad168c209a633185f75938f2d03755a8ae41e3b8e528 not found: ID does not exist" Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.544233 4789 scope.go:117] "RemoveContainer" containerID="266632a371608ebd031caa58818577320156f15bcb53b8beb701f785a6b57f5a" Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.544410 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"266632a371608ebd031caa58818577320156f15bcb53b8beb701f785a6b57f5a"} err="failed to get container status \"266632a371608ebd031caa58818577320156f15bcb53b8beb701f785a6b57f5a\": rpc error: code = NotFound desc = could not find container \"266632a371608ebd031caa58818577320156f15bcb53b8beb701f785a6b57f5a\": container with ID starting with 266632a371608ebd031caa58818577320156f15bcb53b8beb701f785a6b57f5a not found: ID does not exist" Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.544431 4789 scope.go:117] "RemoveContainer" containerID="5ce8875ca32df64c155691ab953221d98fdd435cf3fff8df49a2524cb6a8fca3" Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.544634 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ce8875ca32df64c155691ab953221d98fdd435cf3fff8df49a2524cb6a8fca3"} err="failed to get container status \"5ce8875ca32df64c155691ab953221d98fdd435cf3fff8df49a2524cb6a8fca3\": rpc error: code = NotFound desc = could not find container \"5ce8875ca32df64c155691ab953221d98fdd435cf3fff8df49a2524cb6a8fca3\": container with ID starting with 5ce8875ca32df64c155691ab953221d98fdd435cf3fff8df49a2524cb6a8fca3 not found: ID does not exist" Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.544655 4789 scope.go:117] "RemoveContainer" containerID="c99a248e9a7e8df9e1be7cc7742476be75b4344274665da471fbf3022a343f28" Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.544851 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c99a248e9a7e8df9e1be7cc7742476be75b4344274665da471fbf3022a343f28"} err="failed to get container status \"c99a248e9a7e8df9e1be7cc7742476be75b4344274665da471fbf3022a343f28\": rpc error: code = NotFound desc = could not find container \"c99a248e9a7e8df9e1be7cc7742476be75b4344274665da471fbf3022a343f28\": container with ID starting with c99a248e9a7e8df9e1be7cc7742476be75b4344274665da471fbf3022a343f28 not found: ID does not exist" Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.544873 4789 scope.go:117] "RemoveContainer" containerID="5fd3bfe1a04103047c632d783bc804e6866ca3c2f012da7653ad20d277c3a9d0" Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.545098 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5fd3bfe1a04103047c632d783bc804e6866ca3c2f012da7653ad20d277c3a9d0"} err="failed to get container status \"5fd3bfe1a04103047c632d783bc804e6866ca3c2f012da7653ad20d277c3a9d0\": rpc error: code = NotFound desc = could not find container \"5fd3bfe1a04103047c632d783bc804e6866ca3c2f012da7653ad20d277c3a9d0\": container with ID starting with 5fd3bfe1a04103047c632d783bc804e6866ca3c2f012da7653ad20d277c3a9d0 not found: ID does not exist" Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.545124 4789 scope.go:117] "RemoveContainer" containerID="3740d073f60216d0f7bcd8e638a89cf9f3fe351cd7c21ffaaa67a65a64546c33" Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.545562 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3740d073f60216d0f7bcd8e638a89cf9f3fe351cd7c21ffaaa67a65a64546c33"} err="failed to get container status \"3740d073f60216d0f7bcd8e638a89cf9f3fe351cd7c21ffaaa67a65a64546c33\": rpc error: code = NotFound desc = could not find container \"3740d073f60216d0f7bcd8e638a89cf9f3fe351cd7c21ffaaa67a65a64546c33\": container with ID starting with 3740d073f60216d0f7bcd8e638a89cf9f3fe351cd7c21ffaaa67a65a64546c33 not found: ID does not exist" Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.545587 4789 scope.go:117] "RemoveContainer" containerID="f21f329a56796c8ad0319e24f24b6bb5f1d0f34426dd16613b9fe2819b16f69f" Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.545821 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f21f329a56796c8ad0319e24f24b6bb5f1d0f34426dd16613b9fe2819b16f69f"} err="failed to get container status \"f21f329a56796c8ad0319e24f24b6bb5f1d0f34426dd16613b9fe2819b16f69f\": rpc error: code = NotFound desc = could not find container \"f21f329a56796c8ad0319e24f24b6bb5f1d0f34426dd16613b9fe2819b16f69f\": container with ID starting with f21f329a56796c8ad0319e24f24b6bb5f1d0f34426dd16613b9fe2819b16f69f not found: ID does not exist" Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.545844 4789 scope.go:117] "RemoveContainer" containerID="a42be6036a735dcc45137b6268cfd915cb077dea0c25fdacc35c454aae4fd3a4" Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.546072 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a42be6036a735dcc45137b6268cfd915cb077dea0c25fdacc35c454aae4fd3a4"} err="failed to get container status \"a42be6036a735dcc45137b6268cfd915cb077dea0c25fdacc35c454aae4fd3a4\": rpc error: code = NotFound desc = could not find container \"a42be6036a735dcc45137b6268cfd915cb077dea0c25fdacc35c454aae4fd3a4\": container with ID starting with a42be6036a735dcc45137b6268cfd915cb077dea0c25fdacc35c454aae4fd3a4 not found: ID does not exist" Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.546105 4789 scope.go:117] "RemoveContainer" containerID="dc46c530762e95475561ad5c2b62c516fd6a204aa8ba1010358a8ad977d02aa1" Nov 22 08:08:52 crc kubenswrapper[4789]: I1122 08:08:52.546321 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc46c530762e95475561ad5c2b62c516fd6a204aa8ba1010358a8ad977d02aa1"} err="failed to get container status \"dc46c530762e95475561ad5c2b62c516fd6a204aa8ba1010358a8ad977d02aa1\": rpc error: code = NotFound desc = could not find container \"dc46c530762e95475561ad5c2b62c516fd6a204aa8ba1010358a8ad977d02aa1\": container with ID starting with dc46c530762e95475561ad5c2b62c516fd6a204aa8ba1010358a8ad977d02aa1 not found: ID does not exist" Nov 22 08:08:53 crc kubenswrapper[4789]: I1122 08:08:53.304291 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-rks5j_5613c209-59a0-4b8a-ac8b-5c633bc1bdc4/kube-multus/0.log" Nov 22 08:08:53 crc kubenswrapper[4789]: I1122 08:08:53.304636 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-rks5j" event={"ID":"5613c209-59a0-4b8a-ac8b-5c633bc1bdc4","Type":"ContainerStarted","Data":"ee43b6258e408e86f7100a460969c1bc8beefcd1a3714dcac278a7c7488d9ac8"} Nov 22 08:08:53 crc kubenswrapper[4789]: I1122 08:08:53.312455 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vqhqz" event={"ID":"7f05fa4c-01d8-4bfd-9def-ed962a22960c","Type":"ContainerStarted","Data":"9583cfac25eb8510ad2fb0588a45b94c0d48fee29e5a3ac1983c1d8e092943e9"} Nov 22 08:08:53 crc kubenswrapper[4789]: I1122 08:08:53.312501 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vqhqz" event={"ID":"7f05fa4c-01d8-4bfd-9def-ed962a22960c","Type":"ContainerStarted","Data":"7b48971633f562e30aad9cf285a28c523dd490d6bb5052d98dc145eca54ad8d3"} Nov 22 08:08:53 crc kubenswrapper[4789]: I1122 08:08:53.312513 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vqhqz" event={"ID":"7f05fa4c-01d8-4bfd-9def-ed962a22960c","Type":"ContainerStarted","Data":"710f6d27eef6fda7d1ffec8096277f06933a47a2d32b7f4a28df662c493fb747"} Nov 22 08:08:53 crc kubenswrapper[4789]: I1122 08:08:53.312527 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vqhqz" event={"ID":"7f05fa4c-01d8-4bfd-9def-ed962a22960c","Type":"ContainerStarted","Data":"cad27da8590b3b5192542b391449adbbe618f7c0b35bbd76f86a03a375c7cf11"} Nov 22 08:08:53 crc kubenswrapper[4789]: I1122 08:08:53.312538 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vqhqz" event={"ID":"7f05fa4c-01d8-4bfd-9def-ed962a22960c","Type":"ContainerStarted","Data":"a852140cb6591bac49234c0d26e76a99f8b5aae5c6deb02dea46fe251c7281dc"} Nov 22 08:08:53 crc kubenswrapper[4789]: I1122 08:08:53.312549 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vqhqz" event={"ID":"7f05fa4c-01d8-4bfd-9def-ed962a22960c","Type":"ContainerStarted","Data":"00b728faa5f0d0d565d5ad70f50018ab6294f0c866d5e9e55eeb933ecef187fd"} Nov 22 08:08:53 crc kubenswrapper[4789]: I1122 08:08:53.972285 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0cb4cf51-62d6-4f32-a8b4-4116555f3870" path="/var/lib/kubelet/pods/0cb4cf51-62d6-4f32-a8b4-4116555f3870/volumes" Nov 22 08:08:56 crc kubenswrapper[4789]: I1122 08:08:56.333597 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vqhqz" event={"ID":"7f05fa4c-01d8-4bfd-9def-ed962a22960c","Type":"ContainerStarted","Data":"3dc588ec19ab482cfeafc417be3ae062236c976ecfc2791334dcb65bae05a988"} Nov 22 08:08:58 crc kubenswrapper[4789]: I1122 08:08:58.349721 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vqhqz" event={"ID":"7f05fa4c-01d8-4bfd-9def-ed962a22960c","Type":"ContainerStarted","Data":"969f0966650f0c4baaf8f4db047a59c5b830753cc8ff1a6ccd1c7931e677fc80"} Nov 22 08:08:58 crc kubenswrapper[4789]: I1122 08:08:58.350845 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-vqhqz" Nov 22 08:08:58 crc kubenswrapper[4789]: I1122 08:08:58.351150 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-vqhqz" Nov 22 08:08:58 crc kubenswrapper[4789]: I1122 08:08:58.351166 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-vqhqz" Nov 22 08:08:58 crc kubenswrapper[4789]: I1122 08:08:58.389508 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-vqhqz" Nov 22 08:08:58 crc kubenswrapper[4789]: I1122 08:08:58.392438 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-vqhqz" podStartSLOduration=7.392415486 podStartE2EDuration="7.392415486s" podCreationTimestamp="2025-11-22 08:08:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:08:58.387483732 +0000 UTC m=+752.621884035" watchObservedRunningTime="2025-11-22 08:08:58.392415486 +0000 UTC m=+752.626815799" Nov 22 08:08:58 crc kubenswrapper[4789]: I1122 08:08:58.396471 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-vqhqz" Nov 22 08:09:05 crc kubenswrapper[4789]: I1122 08:09:05.372609 4789 patch_prober.go:28] interesting pod/machine-config-daemon-pxvrm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 08:09:05 crc kubenswrapper[4789]: I1122 08:09:05.373231 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 08:09:22 crc kubenswrapper[4789]: I1122 08:09:22.025031 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-vqhqz" Nov 22 08:09:35 crc kubenswrapper[4789]: I1122 08:09:35.372411 4789 patch_prober.go:28] interesting pod/machine-config-daemon-pxvrm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 08:09:35 crc kubenswrapper[4789]: I1122 08:09:35.373095 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 08:09:35 crc kubenswrapper[4789]: I1122 08:09:35.831092 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ewkwwn"] Nov 22 08:09:35 crc kubenswrapper[4789]: I1122 08:09:35.832057 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ewkwwn" Nov 22 08:09:35 crc kubenswrapper[4789]: I1122 08:09:35.833352 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 22 08:09:35 crc kubenswrapper[4789]: I1122 08:09:35.838899 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ewkwwn"] Nov 22 08:09:35 crc kubenswrapper[4789]: I1122 08:09:35.926412 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4vdmn\" (UniqueName: \"kubernetes.io/projected/a4bfa2cc-45a0-4c18-9a13-19fd8cc06420-kube-api-access-4vdmn\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ewkwwn\" (UID: \"a4bfa2cc-45a0-4c18-9a13-19fd8cc06420\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ewkwwn" Nov 22 08:09:35 crc kubenswrapper[4789]: I1122 08:09:35.926508 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a4bfa2cc-45a0-4c18-9a13-19fd8cc06420-bundle\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ewkwwn\" (UID: \"a4bfa2cc-45a0-4c18-9a13-19fd8cc06420\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ewkwwn" Nov 22 08:09:35 crc kubenswrapper[4789]: I1122 08:09:35.926551 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a4bfa2cc-45a0-4c18-9a13-19fd8cc06420-util\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ewkwwn\" (UID: \"a4bfa2cc-45a0-4c18-9a13-19fd8cc06420\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ewkwwn" Nov 22 08:09:36 crc kubenswrapper[4789]: I1122 08:09:36.027517 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a4bfa2cc-45a0-4c18-9a13-19fd8cc06420-bundle\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ewkwwn\" (UID: \"a4bfa2cc-45a0-4c18-9a13-19fd8cc06420\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ewkwwn" Nov 22 08:09:36 crc kubenswrapper[4789]: I1122 08:09:36.027577 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a4bfa2cc-45a0-4c18-9a13-19fd8cc06420-util\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ewkwwn\" (UID: \"a4bfa2cc-45a0-4c18-9a13-19fd8cc06420\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ewkwwn" Nov 22 08:09:36 crc kubenswrapper[4789]: I1122 08:09:36.027637 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4vdmn\" (UniqueName: \"kubernetes.io/projected/a4bfa2cc-45a0-4c18-9a13-19fd8cc06420-kube-api-access-4vdmn\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ewkwwn\" (UID: \"a4bfa2cc-45a0-4c18-9a13-19fd8cc06420\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ewkwwn" Nov 22 08:09:36 crc kubenswrapper[4789]: I1122 08:09:36.028090 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a4bfa2cc-45a0-4c18-9a13-19fd8cc06420-bundle\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ewkwwn\" (UID: \"a4bfa2cc-45a0-4c18-9a13-19fd8cc06420\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ewkwwn" Nov 22 08:09:36 crc kubenswrapper[4789]: I1122 08:09:36.028198 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a4bfa2cc-45a0-4c18-9a13-19fd8cc06420-util\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ewkwwn\" (UID: \"a4bfa2cc-45a0-4c18-9a13-19fd8cc06420\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ewkwwn" Nov 22 08:09:36 crc kubenswrapper[4789]: I1122 08:09:36.057387 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4vdmn\" (UniqueName: \"kubernetes.io/projected/a4bfa2cc-45a0-4c18-9a13-19fd8cc06420-kube-api-access-4vdmn\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ewkwwn\" (UID: \"a4bfa2cc-45a0-4c18-9a13-19fd8cc06420\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ewkwwn" Nov 22 08:09:36 crc kubenswrapper[4789]: I1122 08:09:36.179330 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ewkwwn" Nov 22 08:09:36 crc kubenswrapper[4789]: I1122 08:09:36.622206 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ewkwwn"] Nov 22 08:09:36 crc kubenswrapper[4789]: W1122 08:09:36.632055 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda4bfa2cc_45a0_4c18_9a13_19fd8cc06420.slice/crio-dccf07f11bbe6b4037b6d72942ebd539b9536bda2f0c748c3caf06ffe71da6cc WatchSource:0}: Error finding container dccf07f11bbe6b4037b6d72942ebd539b9536bda2f0c748c3caf06ffe71da6cc: Status 404 returned error can't find the container with id dccf07f11bbe6b4037b6d72942ebd539b9536bda2f0c748c3caf06ffe71da6cc Nov 22 08:09:37 crc kubenswrapper[4789]: I1122 08:09:37.565629 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ewkwwn" event={"ID":"a4bfa2cc-45a0-4c18-9a13-19fd8cc06420","Type":"ContainerStarted","Data":"094d4cddf6f803374b217c4cb37138605b6e340b4c6c6b7a742046d03cfff171"} Nov 22 08:09:37 crc kubenswrapper[4789]: I1122 08:09:37.565715 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ewkwwn" event={"ID":"a4bfa2cc-45a0-4c18-9a13-19fd8cc06420","Type":"ContainerStarted","Data":"dccf07f11bbe6b4037b6d72942ebd539b9536bda2f0c748c3caf06ffe71da6cc"} Nov 22 08:09:38 crc kubenswrapper[4789]: I1122 08:09:38.572608 4789 generic.go:334] "Generic (PLEG): container finished" podID="a4bfa2cc-45a0-4c18-9a13-19fd8cc06420" containerID="094d4cddf6f803374b217c4cb37138605b6e340b4c6c6b7a742046d03cfff171" exitCode=0 Nov 22 08:09:38 crc kubenswrapper[4789]: I1122 08:09:38.572675 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ewkwwn" event={"ID":"a4bfa2cc-45a0-4c18-9a13-19fd8cc06420","Type":"ContainerDied","Data":"094d4cddf6f803374b217c4cb37138605b6e340b4c6c6b7a742046d03cfff171"} Nov 22 08:09:38 crc kubenswrapper[4789]: I1122 08:09:38.969464 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-2sp2r"] Nov 22 08:09:38 crc kubenswrapper[4789]: I1122 08:09:38.971122 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2sp2r" Nov 22 08:09:38 crc kubenswrapper[4789]: I1122 08:09:38.977017 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2sp2r"] Nov 22 08:09:39 crc kubenswrapper[4789]: I1122 08:09:39.068257 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5746cb8d-5058-45e7-9f5d-f1e90a666973-utilities\") pod \"redhat-operators-2sp2r\" (UID: \"5746cb8d-5058-45e7-9f5d-f1e90a666973\") " pod="openshift-marketplace/redhat-operators-2sp2r" Nov 22 08:09:39 crc kubenswrapper[4789]: I1122 08:09:39.068385 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-szz4q\" (UniqueName: \"kubernetes.io/projected/5746cb8d-5058-45e7-9f5d-f1e90a666973-kube-api-access-szz4q\") pod \"redhat-operators-2sp2r\" (UID: \"5746cb8d-5058-45e7-9f5d-f1e90a666973\") " pod="openshift-marketplace/redhat-operators-2sp2r" Nov 22 08:09:39 crc kubenswrapper[4789]: I1122 08:09:39.068412 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5746cb8d-5058-45e7-9f5d-f1e90a666973-catalog-content\") pod \"redhat-operators-2sp2r\" (UID: \"5746cb8d-5058-45e7-9f5d-f1e90a666973\") " pod="openshift-marketplace/redhat-operators-2sp2r" Nov 22 08:09:39 crc kubenswrapper[4789]: I1122 08:09:39.169094 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5746cb8d-5058-45e7-9f5d-f1e90a666973-utilities\") pod \"redhat-operators-2sp2r\" (UID: \"5746cb8d-5058-45e7-9f5d-f1e90a666973\") " pod="openshift-marketplace/redhat-operators-2sp2r" Nov 22 08:09:39 crc kubenswrapper[4789]: I1122 08:09:39.169146 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5746cb8d-5058-45e7-9f5d-f1e90a666973-catalog-content\") pod \"redhat-operators-2sp2r\" (UID: \"5746cb8d-5058-45e7-9f5d-f1e90a666973\") " pod="openshift-marketplace/redhat-operators-2sp2r" Nov 22 08:09:39 crc kubenswrapper[4789]: I1122 08:09:39.169167 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-szz4q\" (UniqueName: \"kubernetes.io/projected/5746cb8d-5058-45e7-9f5d-f1e90a666973-kube-api-access-szz4q\") pod \"redhat-operators-2sp2r\" (UID: \"5746cb8d-5058-45e7-9f5d-f1e90a666973\") " pod="openshift-marketplace/redhat-operators-2sp2r" Nov 22 08:09:39 crc kubenswrapper[4789]: I1122 08:09:39.169553 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5746cb8d-5058-45e7-9f5d-f1e90a666973-utilities\") pod \"redhat-operators-2sp2r\" (UID: \"5746cb8d-5058-45e7-9f5d-f1e90a666973\") " pod="openshift-marketplace/redhat-operators-2sp2r" Nov 22 08:09:39 crc kubenswrapper[4789]: I1122 08:09:39.169597 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5746cb8d-5058-45e7-9f5d-f1e90a666973-catalog-content\") pod \"redhat-operators-2sp2r\" (UID: \"5746cb8d-5058-45e7-9f5d-f1e90a666973\") " pod="openshift-marketplace/redhat-operators-2sp2r" Nov 22 08:09:39 crc kubenswrapper[4789]: I1122 08:09:39.191420 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-szz4q\" (UniqueName: \"kubernetes.io/projected/5746cb8d-5058-45e7-9f5d-f1e90a666973-kube-api-access-szz4q\") pod \"redhat-operators-2sp2r\" (UID: \"5746cb8d-5058-45e7-9f5d-f1e90a666973\") " pod="openshift-marketplace/redhat-operators-2sp2r" Nov 22 08:09:39 crc kubenswrapper[4789]: I1122 08:09:39.290560 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2sp2r" Nov 22 08:09:39 crc kubenswrapper[4789]: I1122 08:09:39.370080 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-gvdqx"] Nov 22 08:09:39 crc kubenswrapper[4789]: I1122 08:09:39.373478 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gvdqx" Nov 22 08:09:39 crc kubenswrapper[4789]: I1122 08:09:39.382349 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gvdqx"] Nov 22 08:09:39 crc kubenswrapper[4789]: I1122 08:09:39.474257 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rdg68\" (UniqueName: \"kubernetes.io/projected/02ccf358-42e4-48bc-9748-de9e2c46e2b5-kube-api-access-rdg68\") pod \"community-operators-gvdqx\" (UID: \"02ccf358-42e4-48bc-9748-de9e2c46e2b5\") " pod="openshift-marketplace/community-operators-gvdqx" Nov 22 08:09:39 crc kubenswrapper[4789]: I1122 08:09:39.474315 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02ccf358-42e4-48bc-9748-de9e2c46e2b5-utilities\") pod \"community-operators-gvdqx\" (UID: \"02ccf358-42e4-48bc-9748-de9e2c46e2b5\") " pod="openshift-marketplace/community-operators-gvdqx" Nov 22 08:09:39 crc kubenswrapper[4789]: I1122 08:09:39.474534 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02ccf358-42e4-48bc-9748-de9e2c46e2b5-catalog-content\") pod \"community-operators-gvdqx\" (UID: \"02ccf358-42e4-48bc-9748-de9e2c46e2b5\") " pod="openshift-marketplace/community-operators-gvdqx" Nov 22 08:09:39 crc kubenswrapper[4789]: I1122 08:09:39.575309 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02ccf358-42e4-48bc-9748-de9e2c46e2b5-catalog-content\") pod \"community-operators-gvdqx\" (UID: \"02ccf358-42e4-48bc-9748-de9e2c46e2b5\") " pod="openshift-marketplace/community-operators-gvdqx" Nov 22 08:09:39 crc kubenswrapper[4789]: I1122 08:09:39.575351 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdg68\" (UniqueName: \"kubernetes.io/projected/02ccf358-42e4-48bc-9748-de9e2c46e2b5-kube-api-access-rdg68\") pod \"community-operators-gvdqx\" (UID: \"02ccf358-42e4-48bc-9748-de9e2c46e2b5\") " pod="openshift-marketplace/community-operators-gvdqx" Nov 22 08:09:39 crc kubenswrapper[4789]: I1122 08:09:39.575382 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02ccf358-42e4-48bc-9748-de9e2c46e2b5-utilities\") pod \"community-operators-gvdqx\" (UID: \"02ccf358-42e4-48bc-9748-de9e2c46e2b5\") " pod="openshift-marketplace/community-operators-gvdqx" Nov 22 08:09:39 crc kubenswrapper[4789]: I1122 08:09:39.575860 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02ccf358-42e4-48bc-9748-de9e2c46e2b5-utilities\") pod \"community-operators-gvdqx\" (UID: \"02ccf358-42e4-48bc-9748-de9e2c46e2b5\") " pod="openshift-marketplace/community-operators-gvdqx" Nov 22 08:09:39 crc kubenswrapper[4789]: I1122 08:09:39.576075 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02ccf358-42e4-48bc-9748-de9e2c46e2b5-catalog-content\") pod \"community-operators-gvdqx\" (UID: \"02ccf358-42e4-48bc-9748-de9e2c46e2b5\") " pod="openshift-marketplace/community-operators-gvdqx" Nov 22 08:09:39 crc kubenswrapper[4789]: I1122 08:09:39.596038 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdg68\" (UniqueName: \"kubernetes.io/projected/02ccf358-42e4-48bc-9748-de9e2c46e2b5-kube-api-access-rdg68\") pod \"community-operators-gvdqx\" (UID: \"02ccf358-42e4-48bc-9748-de9e2c46e2b5\") " pod="openshift-marketplace/community-operators-gvdqx" Nov 22 08:09:39 crc kubenswrapper[4789]: I1122 08:09:39.699841 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gvdqx" Nov 22 08:09:39 crc kubenswrapper[4789]: I1122 08:09:39.786609 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2sp2r"] Nov 22 08:09:39 crc kubenswrapper[4789]: W1122 08:09:39.798224 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5746cb8d_5058_45e7_9f5d_f1e90a666973.slice/crio-e77cabeec9f2e1eaaf50b2c017401fdd2670c80bb3219eb79cf737dc97255fa9 WatchSource:0}: Error finding container e77cabeec9f2e1eaaf50b2c017401fdd2670c80bb3219eb79cf737dc97255fa9: Status 404 returned error can't find the container with id e77cabeec9f2e1eaaf50b2c017401fdd2670c80bb3219eb79cf737dc97255fa9 Nov 22 08:09:39 crc kubenswrapper[4789]: I1122 08:09:39.938411 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gvdqx"] Nov 22 08:09:40 crc kubenswrapper[4789]: W1122 08:09:40.457276 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod02ccf358_42e4_48bc_9748_de9e2c46e2b5.slice/crio-c84b01e4a2bbd9279664c59069b740c757e543b91d21dd9250e1668cb5575b6c WatchSource:0}: Error finding container c84b01e4a2bbd9279664c59069b740c757e543b91d21dd9250e1668cb5575b6c: Status 404 returned error can't find the container with id c84b01e4a2bbd9279664c59069b740c757e543b91d21dd9250e1668cb5575b6c Nov 22 08:09:40 crc kubenswrapper[4789]: I1122 08:09:40.585231 4789 generic.go:334] "Generic (PLEG): container finished" podID="5746cb8d-5058-45e7-9f5d-f1e90a666973" containerID="0153cb94cea2a99ed67964df338bc6fa0590653475132e4e8739a97dcf1b835d" exitCode=0 Nov 22 08:09:40 crc kubenswrapper[4789]: I1122 08:09:40.585335 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2sp2r" event={"ID":"5746cb8d-5058-45e7-9f5d-f1e90a666973","Type":"ContainerDied","Data":"0153cb94cea2a99ed67964df338bc6fa0590653475132e4e8739a97dcf1b835d"} Nov 22 08:09:40 crc kubenswrapper[4789]: I1122 08:09:40.585398 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2sp2r" event={"ID":"5746cb8d-5058-45e7-9f5d-f1e90a666973","Type":"ContainerStarted","Data":"e77cabeec9f2e1eaaf50b2c017401fdd2670c80bb3219eb79cf737dc97255fa9"} Nov 22 08:09:40 crc kubenswrapper[4789]: I1122 08:09:40.586363 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gvdqx" event={"ID":"02ccf358-42e4-48bc-9748-de9e2c46e2b5","Type":"ContainerStarted","Data":"c84b01e4a2bbd9279664c59069b740c757e543b91d21dd9250e1668cb5575b6c"} Nov 22 08:09:41 crc kubenswrapper[4789]: I1122 08:09:41.592225 4789 generic.go:334] "Generic (PLEG): container finished" podID="02ccf358-42e4-48bc-9748-de9e2c46e2b5" containerID="14b0b12f7ba5a10d0e389a88a9f09a496bfc0b2ded7fd2b8a0e5e3167ab448ec" exitCode=0 Nov 22 08:09:41 crc kubenswrapper[4789]: I1122 08:09:41.592282 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gvdqx" event={"ID":"02ccf358-42e4-48bc-9748-de9e2c46e2b5","Type":"ContainerDied","Data":"14b0b12f7ba5a10d0e389a88a9f09a496bfc0b2ded7fd2b8a0e5e3167ab448ec"} Nov 22 08:09:42 crc kubenswrapper[4789]: I1122 08:09:42.569833 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-b2vc5"] Nov 22 08:09:42 crc kubenswrapper[4789]: I1122 08:09:42.571833 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b2vc5" Nov 22 08:09:42 crc kubenswrapper[4789]: I1122 08:09:42.578478 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-b2vc5"] Nov 22 08:09:42 crc kubenswrapper[4789]: I1122 08:09:42.602466 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2sp2r" event={"ID":"5746cb8d-5058-45e7-9f5d-f1e90a666973","Type":"ContainerStarted","Data":"550ec18fb1967ed632ad463b992a06481679dfe9e29e9a9bb7a1c4a651630667"} Nov 22 08:09:42 crc kubenswrapper[4789]: I1122 08:09:42.605131 4789 generic.go:334] "Generic (PLEG): container finished" podID="a4bfa2cc-45a0-4c18-9a13-19fd8cc06420" containerID="31add9134b2d37f7e7829519ca3ae69864c19852b82297bfab6883af0080baf2" exitCode=0 Nov 22 08:09:42 crc kubenswrapper[4789]: I1122 08:09:42.605168 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ewkwwn" event={"ID":"a4bfa2cc-45a0-4c18-9a13-19fd8cc06420","Type":"ContainerDied","Data":"31add9134b2d37f7e7829519ca3ae69864c19852b82297bfab6883af0080baf2"} Nov 22 08:09:42 crc kubenswrapper[4789]: I1122 08:09:42.715268 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3009aea9-8c5a-42cc-b9d6-dca6ce289616-utilities\") pod \"redhat-marketplace-b2vc5\" (UID: \"3009aea9-8c5a-42cc-b9d6-dca6ce289616\") " pod="openshift-marketplace/redhat-marketplace-b2vc5" Nov 22 08:09:42 crc kubenswrapper[4789]: I1122 08:09:42.715348 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dv64c\" (UniqueName: \"kubernetes.io/projected/3009aea9-8c5a-42cc-b9d6-dca6ce289616-kube-api-access-dv64c\") pod \"redhat-marketplace-b2vc5\" (UID: \"3009aea9-8c5a-42cc-b9d6-dca6ce289616\") " pod="openshift-marketplace/redhat-marketplace-b2vc5" Nov 22 08:09:42 crc kubenswrapper[4789]: I1122 08:09:42.715386 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3009aea9-8c5a-42cc-b9d6-dca6ce289616-catalog-content\") pod \"redhat-marketplace-b2vc5\" (UID: \"3009aea9-8c5a-42cc-b9d6-dca6ce289616\") " pod="openshift-marketplace/redhat-marketplace-b2vc5" Nov 22 08:09:42 crc kubenswrapper[4789]: I1122 08:09:42.816686 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dv64c\" (UniqueName: \"kubernetes.io/projected/3009aea9-8c5a-42cc-b9d6-dca6ce289616-kube-api-access-dv64c\") pod \"redhat-marketplace-b2vc5\" (UID: \"3009aea9-8c5a-42cc-b9d6-dca6ce289616\") " pod="openshift-marketplace/redhat-marketplace-b2vc5" Nov 22 08:09:42 crc kubenswrapper[4789]: I1122 08:09:42.816782 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3009aea9-8c5a-42cc-b9d6-dca6ce289616-catalog-content\") pod \"redhat-marketplace-b2vc5\" (UID: \"3009aea9-8c5a-42cc-b9d6-dca6ce289616\") " pod="openshift-marketplace/redhat-marketplace-b2vc5" Nov 22 08:09:42 crc kubenswrapper[4789]: I1122 08:09:42.816856 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3009aea9-8c5a-42cc-b9d6-dca6ce289616-utilities\") pod \"redhat-marketplace-b2vc5\" (UID: \"3009aea9-8c5a-42cc-b9d6-dca6ce289616\") " pod="openshift-marketplace/redhat-marketplace-b2vc5" Nov 22 08:09:42 crc kubenswrapper[4789]: I1122 08:09:42.817330 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3009aea9-8c5a-42cc-b9d6-dca6ce289616-utilities\") pod \"redhat-marketplace-b2vc5\" (UID: \"3009aea9-8c5a-42cc-b9d6-dca6ce289616\") " pod="openshift-marketplace/redhat-marketplace-b2vc5" Nov 22 08:09:42 crc kubenswrapper[4789]: I1122 08:09:42.817450 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3009aea9-8c5a-42cc-b9d6-dca6ce289616-catalog-content\") pod \"redhat-marketplace-b2vc5\" (UID: \"3009aea9-8c5a-42cc-b9d6-dca6ce289616\") " pod="openshift-marketplace/redhat-marketplace-b2vc5" Nov 22 08:09:42 crc kubenswrapper[4789]: I1122 08:09:42.846409 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dv64c\" (UniqueName: \"kubernetes.io/projected/3009aea9-8c5a-42cc-b9d6-dca6ce289616-kube-api-access-dv64c\") pod \"redhat-marketplace-b2vc5\" (UID: \"3009aea9-8c5a-42cc-b9d6-dca6ce289616\") " pod="openshift-marketplace/redhat-marketplace-b2vc5" Nov 22 08:09:42 crc kubenswrapper[4789]: I1122 08:09:42.899288 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b2vc5" Nov 22 08:09:43 crc kubenswrapper[4789]: I1122 08:09:43.308248 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-b2vc5"] Nov 22 08:09:43 crc kubenswrapper[4789]: I1122 08:09:43.613469 4789 generic.go:334] "Generic (PLEG): container finished" podID="3009aea9-8c5a-42cc-b9d6-dca6ce289616" containerID="12dd9f00b45f9af5bc3a2013a1bc72aedca04f0abdca53e2aad3f61ba6317adf" exitCode=0 Nov 22 08:09:43 crc kubenswrapper[4789]: I1122 08:09:43.613528 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b2vc5" event={"ID":"3009aea9-8c5a-42cc-b9d6-dca6ce289616","Type":"ContainerDied","Data":"12dd9f00b45f9af5bc3a2013a1bc72aedca04f0abdca53e2aad3f61ba6317adf"} Nov 22 08:09:43 crc kubenswrapper[4789]: I1122 08:09:43.613578 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b2vc5" event={"ID":"3009aea9-8c5a-42cc-b9d6-dca6ce289616","Type":"ContainerStarted","Data":"f37838bb7c96aae8e6ae01824e0400f2f611df8ec3bb695d3276531022eadc10"} Nov 22 08:09:43 crc kubenswrapper[4789]: I1122 08:09:43.620265 4789 generic.go:334] "Generic (PLEG): container finished" podID="a4bfa2cc-45a0-4c18-9a13-19fd8cc06420" containerID="c5ac7a5279a8cbd2a49f45bf3ee937f2fd41788160b08840ec9d14f2bf432d03" exitCode=0 Nov 22 08:09:43 crc kubenswrapper[4789]: I1122 08:09:43.620338 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ewkwwn" event={"ID":"a4bfa2cc-45a0-4c18-9a13-19fd8cc06420","Type":"ContainerDied","Data":"c5ac7a5279a8cbd2a49f45bf3ee937f2fd41788160b08840ec9d14f2bf432d03"} Nov 22 08:09:43 crc kubenswrapper[4789]: I1122 08:09:43.623084 4789 generic.go:334] "Generic (PLEG): container finished" podID="5746cb8d-5058-45e7-9f5d-f1e90a666973" containerID="550ec18fb1967ed632ad463b992a06481679dfe9e29e9a9bb7a1c4a651630667" exitCode=0 Nov 22 08:09:43 crc kubenswrapper[4789]: I1122 08:09:43.623163 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2sp2r" event={"ID":"5746cb8d-5058-45e7-9f5d-f1e90a666973","Type":"ContainerDied","Data":"550ec18fb1967ed632ad463b992a06481679dfe9e29e9a9bb7a1c4a651630667"} Nov 22 08:09:43 crc kubenswrapper[4789]: I1122 08:09:43.625561 4789 generic.go:334] "Generic (PLEG): container finished" podID="02ccf358-42e4-48bc-9748-de9e2c46e2b5" containerID="f1edced96ee24f085cf1b90b51e8f1a237f7df87fac0ad16e2a711fb5ba1e12b" exitCode=0 Nov 22 08:09:43 crc kubenswrapper[4789]: I1122 08:09:43.625595 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gvdqx" event={"ID":"02ccf358-42e4-48bc-9748-de9e2c46e2b5","Type":"ContainerDied","Data":"f1edced96ee24f085cf1b90b51e8f1a237f7df87fac0ad16e2a711fb5ba1e12b"} Nov 22 08:09:44 crc kubenswrapper[4789]: I1122 08:09:44.634001 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2sp2r" event={"ID":"5746cb8d-5058-45e7-9f5d-f1e90a666973","Type":"ContainerStarted","Data":"0366eac1b416707d6c0bad3298fa20229f0ea18c7a7108c770f31c84752976a8"} Nov 22 08:09:44 crc kubenswrapper[4789]: I1122 08:09:44.636992 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gvdqx" event={"ID":"02ccf358-42e4-48bc-9748-de9e2c46e2b5","Type":"ContainerStarted","Data":"9acec1deb69b06863e9efed3a816d35d4d8e45becf4da0a9a2cbd3ccf162a98f"} Nov 22 08:09:44 crc kubenswrapper[4789]: I1122 08:09:44.654079 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-2sp2r" podStartSLOduration=3.210314148 podStartE2EDuration="6.654065509s" podCreationTimestamp="2025-11-22 08:09:38 +0000 UTC" firstStartedPulling="2025-11-22 08:09:40.5870708 +0000 UTC m=+794.821471073" lastFinishedPulling="2025-11-22 08:09:44.030822161 +0000 UTC m=+798.265222434" observedRunningTime="2025-11-22 08:09:44.651264143 +0000 UTC m=+798.885664416" watchObservedRunningTime="2025-11-22 08:09:44.654065509 +0000 UTC m=+798.888465782" Nov 22 08:09:44 crc kubenswrapper[4789]: I1122 08:09:44.673044 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-gvdqx" podStartSLOduration=3.189768306 podStartE2EDuration="5.673025334s" podCreationTimestamp="2025-11-22 08:09:39 +0000 UTC" firstStartedPulling="2025-11-22 08:09:41.611043226 +0000 UTC m=+795.845443499" lastFinishedPulling="2025-11-22 08:09:44.094300264 +0000 UTC m=+798.328700527" observedRunningTime="2025-11-22 08:09:44.671058931 +0000 UTC m=+798.905459224" watchObservedRunningTime="2025-11-22 08:09:44.673025334 +0000 UTC m=+798.907425607" Nov 22 08:09:44 crc kubenswrapper[4789]: I1122 08:09:44.836933 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ewkwwn" Nov 22 08:09:44 crc kubenswrapper[4789]: I1122 08:09:44.950874 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a4bfa2cc-45a0-4c18-9a13-19fd8cc06420-util\") pod \"a4bfa2cc-45a0-4c18-9a13-19fd8cc06420\" (UID: \"a4bfa2cc-45a0-4c18-9a13-19fd8cc06420\") " Nov 22 08:09:44 crc kubenswrapper[4789]: I1122 08:09:44.950943 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a4bfa2cc-45a0-4c18-9a13-19fd8cc06420-bundle\") pod \"a4bfa2cc-45a0-4c18-9a13-19fd8cc06420\" (UID: \"a4bfa2cc-45a0-4c18-9a13-19fd8cc06420\") " Nov 22 08:09:44 crc kubenswrapper[4789]: I1122 08:09:44.951049 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4vdmn\" (UniqueName: \"kubernetes.io/projected/a4bfa2cc-45a0-4c18-9a13-19fd8cc06420-kube-api-access-4vdmn\") pod \"a4bfa2cc-45a0-4c18-9a13-19fd8cc06420\" (UID: \"a4bfa2cc-45a0-4c18-9a13-19fd8cc06420\") " Nov 22 08:09:44 crc kubenswrapper[4789]: I1122 08:09:44.952423 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a4bfa2cc-45a0-4c18-9a13-19fd8cc06420-bundle" (OuterVolumeSpecName: "bundle") pod "a4bfa2cc-45a0-4c18-9a13-19fd8cc06420" (UID: "a4bfa2cc-45a0-4c18-9a13-19fd8cc06420"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:09:44 crc kubenswrapper[4789]: I1122 08:09:44.956191 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a4bfa2cc-45a0-4c18-9a13-19fd8cc06420-kube-api-access-4vdmn" (OuterVolumeSpecName: "kube-api-access-4vdmn") pod "a4bfa2cc-45a0-4c18-9a13-19fd8cc06420" (UID: "a4bfa2cc-45a0-4c18-9a13-19fd8cc06420"). InnerVolumeSpecName "kube-api-access-4vdmn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:09:44 crc kubenswrapper[4789]: I1122 08:09:44.962556 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a4bfa2cc-45a0-4c18-9a13-19fd8cc06420-util" (OuterVolumeSpecName: "util") pod "a4bfa2cc-45a0-4c18-9a13-19fd8cc06420" (UID: "a4bfa2cc-45a0-4c18-9a13-19fd8cc06420"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:09:45 crc kubenswrapper[4789]: I1122 08:09:45.052950 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4vdmn\" (UniqueName: \"kubernetes.io/projected/a4bfa2cc-45a0-4c18-9a13-19fd8cc06420-kube-api-access-4vdmn\") on node \"crc\" DevicePath \"\"" Nov 22 08:09:45 crc kubenswrapper[4789]: I1122 08:09:45.053004 4789 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a4bfa2cc-45a0-4c18-9a13-19fd8cc06420-util\") on node \"crc\" DevicePath \"\"" Nov 22 08:09:45 crc kubenswrapper[4789]: I1122 08:09:45.053017 4789 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a4bfa2cc-45a0-4c18-9a13-19fd8cc06420-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:09:45 crc kubenswrapper[4789]: I1122 08:09:45.642942 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ewkwwn" event={"ID":"a4bfa2cc-45a0-4c18-9a13-19fd8cc06420","Type":"ContainerDied","Data":"dccf07f11bbe6b4037b6d72942ebd539b9536bda2f0c748c3caf06ffe71da6cc"} Nov 22 08:09:45 crc kubenswrapper[4789]: I1122 08:09:45.642981 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ewkwwn" Nov 22 08:09:45 crc kubenswrapper[4789]: I1122 08:09:45.642986 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dccf07f11bbe6b4037b6d72942ebd539b9536bda2f0c748c3caf06ffe71da6cc" Nov 22 08:09:45 crc kubenswrapper[4789]: I1122 08:09:45.644486 4789 generic.go:334] "Generic (PLEG): container finished" podID="3009aea9-8c5a-42cc-b9d6-dca6ce289616" containerID="8a60b113d1b061d3884e76a8f511e9b3560cba34c88992b11967e321d5646322" exitCode=0 Nov 22 08:09:45 crc kubenswrapper[4789]: I1122 08:09:45.644515 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b2vc5" event={"ID":"3009aea9-8c5a-42cc-b9d6-dca6ce289616","Type":"ContainerDied","Data":"8a60b113d1b061d3884e76a8f511e9b3560cba34c88992b11967e321d5646322"} Nov 22 08:09:46 crc kubenswrapper[4789]: I1122 08:09:46.359742 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-557fdffb88-cfdtx"] Nov 22 08:09:46 crc kubenswrapper[4789]: E1122 08:09:46.360344 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4bfa2cc-45a0-4c18-9a13-19fd8cc06420" containerName="extract" Nov 22 08:09:46 crc kubenswrapper[4789]: I1122 08:09:46.360358 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4bfa2cc-45a0-4c18-9a13-19fd8cc06420" containerName="extract" Nov 22 08:09:46 crc kubenswrapper[4789]: E1122 08:09:46.360378 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4bfa2cc-45a0-4c18-9a13-19fd8cc06420" containerName="pull" Nov 22 08:09:46 crc kubenswrapper[4789]: I1122 08:09:46.360385 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4bfa2cc-45a0-4c18-9a13-19fd8cc06420" containerName="pull" Nov 22 08:09:46 crc kubenswrapper[4789]: E1122 08:09:46.360405 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4bfa2cc-45a0-4c18-9a13-19fd8cc06420" containerName="util" Nov 22 08:09:46 crc kubenswrapper[4789]: I1122 08:09:46.360411 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4bfa2cc-45a0-4c18-9a13-19fd8cc06420" containerName="util" Nov 22 08:09:46 crc kubenswrapper[4789]: I1122 08:09:46.360521 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4bfa2cc-45a0-4c18-9a13-19fd8cc06420" containerName="extract" Nov 22 08:09:46 crc kubenswrapper[4789]: I1122 08:09:46.360919 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-557fdffb88-cfdtx" Nov 22 08:09:46 crc kubenswrapper[4789]: I1122 08:09:46.363098 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Nov 22 08:09:46 crc kubenswrapper[4789]: I1122 08:09:46.363198 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-ftssp" Nov 22 08:09:46 crc kubenswrapper[4789]: I1122 08:09:46.364641 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Nov 22 08:09:46 crc kubenswrapper[4789]: I1122 08:09:46.368198 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s52fd\" (UniqueName: \"kubernetes.io/projected/42d193be-d22b-4731-baed-4773d41c095d-kube-api-access-s52fd\") pod \"nmstate-operator-557fdffb88-cfdtx\" (UID: \"42d193be-d22b-4731-baed-4773d41c095d\") " pod="openshift-nmstate/nmstate-operator-557fdffb88-cfdtx" Nov 22 08:09:46 crc kubenswrapper[4789]: I1122 08:09:46.377003 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-557fdffb88-cfdtx"] Nov 22 08:09:46 crc kubenswrapper[4789]: I1122 08:09:46.468902 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s52fd\" (UniqueName: \"kubernetes.io/projected/42d193be-d22b-4731-baed-4773d41c095d-kube-api-access-s52fd\") pod \"nmstate-operator-557fdffb88-cfdtx\" (UID: \"42d193be-d22b-4731-baed-4773d41c095d\") " pod="openshift-nmstate/nmstate-operator-557fdffb88-cfdtx" Nov 22 08:09:46 crc kubenswrapper[4789]: I1122 08:09:46.493737 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s52fd\" (UniqueName: \"kubernetes.io/projected/42d193be-d22b-4731-baed-4773d41c095d-kube-api-access-s52fd\") pod \"nmstate-operator-557fdffb88-cfdtx\" (UID: \"42d193be-d22b-4731-baed-4773d41c095d\") " pod="openshift-nmstate/nmstate-operator-557fdffb88-cfdtx" Nov 22 08:09:46 crc kubenswrapper[4789]: I1122 08:09:46.653256 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b2vc5" event={"ID":"3009aea9-8c5a-42cc-b9d6-dca6ce289616","Type":"ContainerStarted","Data":"118ff9b20cd28a73ebbfeaea76cf9386d79de7513149c06e7011cfaeb91de3ba"} Nov 22 08:09:46 crc kubenswrapper[4789]: I1122 08:09:46.672479 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-b2vc5" podStartSLOduration=2.262466538 podStartE2EDuration="4.672451558s" podCreationTimestamp="2025-11-22 08:09:42 +0000 UTC" firstStartedPulling="2025-11-22 08:09:43.615355793 +0000 UTC m=+797.849756066" lastFinishedPulling="2025-11-22 08:09:46.025340813 +0000 UTC m=+800.259741086" observedRunningTime="2025-11-22 08:09:46.668603474 +0000 UTC m=+800.903003757" watchObservedRunningTime="2025-11-22 08:09:46.672451558 +0000 UTC m=+800.906851851" Nov 22 08:09:46 crc kubenswrapper[4789]: I1122 08:09:46.673098 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-557fdffb88-cfdtx" Nov 22 08:09:47 crc kubenswrapper[4789]: I1122 08:09:47.111879 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-557fdffb88-cfdtx"] Nov 22 08:09:47 crc kubenswrapper[4789]: I1122 08:09:47.660367 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-557fdffb88-cfdtx" event={"ID":"42d193be-d22b-4731-baed-4773d41c095d","Type":"ContainerStarted","Data":"f100f279ad3efb713767c6524e3f458928fa50d05847aac35eb894654ec190aa"} Nov 22 08:09:49 crc kubenswrapper[4789]: I1122 08:09:49.291025 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-2sp2r" Nov 22 08:09:49 crc kubenswrapper[4789]: I1122 08:09:49.291366 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-2sp2r" Nov 22 08:09:49 crc kubenswrapper[4789]: I1122 08:09:49.675236 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-557fdffb88-cfdtx" event={"ID":"42d193be-d22b-4731-baed-4773d41c095d","Type":"ContainerStarted","Data":"66ba7fe33e206a325d927d58f4729a50be55be88fb0334b9094b508a2595408a"} Nov 22 08:09:49 crc kubenswrapper[4789]: I1122 08:09:49.694636 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-557fdffb88-cfdtx" podStartSLOduration=1.322584736 podStartE2EDuration="3.694615795s" podCreationTimestamp="2025-11-22 08:09:46 +0000 UTC" firstStartedPulling="2025-11-22 08:09:47.122391832 +0000 UTC m=+801.356792105" lastFinishedPulling="2025-11-22 08:09:49.494422891 +0000 UTC m=+803.728823164" observedRunningTime="2025-11-22 08:09:49.691944614 +0000 UTC m=+803.926344887" watchObservedRunningTime="2025-11-22 08:09:49.694615795 +0000 UTC m=+803.929016068" Nov 22 08:09:49 crc kubenswrapper[4789]: I1122 08:09:49.700370 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-gvdqx" Nov 22 08:09:49 crc kubenswrapper[4789]: I1122 08:09:49.700678 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-gvdqx" Nov 22 08:09:49 crc kubenswrapper[4789]: I1122 08:09:49.746992 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-gvdqx" Nov 22 08:09:50 crc kubenswrapper[4789]: I1122 08:09:50.332267 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-2sp2r" podUID="5746cb8d-5058-45e7-9f5d-f1e90a666973" containerName="registry-server" probeResult="failure" output=< Nov 22 08:09:50 crc kubenswrapper[4789]: timeout: failed to connect service ":50051" within 1s Nov 22 08:09:50 crc kubenswrapper[4789]: > Nov 22 08:09:50 crc kubenswrapper[4789]: I1122 08:09:50.737207 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-gvdqx" Nov 22 08:09:52 crc kubenswrapper[4789]: I1122 08:09:52.899856 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-b2vc5" Nov 22 08:09:52 crc kubenswrapper[4789]: I1122 08:09:52.900258 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-b2vc5" Nov 22 08:09:52 crc kubenswrapper[4789]: I1122 08:09:52.934904 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-b2vc5" Nov 22 08:09:53 crc kubenswrapper[4789]: I1122 08:09:53.749012 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-b2vc5" Nov 22 08:09:53 crc kubenswrapper[4789]: I1122 08:09:53.760164 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gvdqx"] Nov 22 08:09:53 crc kubenswrapper[4789]: I1122 08:09:53.760419 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-gvdqx" podUID="02ccf358-42e4-48bc-9748-de9e2c46e2b5" containerName="registry-server" containerID="cri-o://9acec1deb69b06863e9efed3a816d35d4d8e45becf4da0a9a2cbd3ccf162a98f" gracePeriod=2 Nov 22 08:09:54 crc kubenswrapper[4789]: I1122 08:09:54.107154 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gvdqx" Nov 22 08:09:54 crc kubenswrapper[4789]: I1122 08:09:54.266524 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02ccf358-42e4-48bc-9748-de9e2c46e2b5-catalog-content\") pod \"02ccf358-42e4-48bc-9748-de9e2c46e2b5\" (UID: \"02ccf358-42e4-48bc-9748-de9e2c46e2b5\") " Nov 22 08:09:54 crc kubenswrapper[4789]: I1122 08:09:54.266727 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rdg68\" (UniqueName: \"kubernetes.io/projected/02ccf358-42e4-48bc-9748-de9e2c46e2b5-kube-api-access-rdg68\") pod \"02ccf358-42e4-48bc-9748-de9e2c46e2b5\" (UID: \"02ccf358-42e4-48bc-9748-de9e2c46e2b5\") " Nov 22 08:09:54 crc kubenswrapper[4789]: I1122 08:09:54.266952 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02ccf358-42e4-48bc-9748-de9e2c46e2b5-utilities\") pod \"02ccf358-42e4-48bc-9748-de9e2c46e2b5\" (UID: \"02ccf358-42e4-48bc-9748-de9e2c46e2b5\") " Nov 22 08:09:54 crc kubenswrapper[4789]: I1122 08:09:54.268386 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/02ccf358-42e4-48bc-9748-de9e2c46e2b5-utilities" (OuterVolumeSpecName: "utilities") pod "02ccf358-42e4-48bc-9748-de9e2c46e2b5" (UID: "02ccf358-42e4-48bc-9748-de9e2c46e2b5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:09:54 crc kubenswrapper[4789]: I1122 08:09:54.271849 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/02ccf358-42e4-48bc-9748-de9e2c46e2b5-kube-api-access-rdg68" (OuterVolumeSpecName: "kube-api-access-rdg68") pod "02ccf358-42e4-48bc-9748-de9e2c46e2b5" (UID: "02ccf358-42e4-48bc-9748-de9e2c46e2b5"). InnerVolumeSpecName "kube-api-access-rdg68". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:09:54 crc kubenswrapper[4789]: I1122 08:09:54.328313 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/02ccf358-42e4-48bc-9748-de9e2c46e2b5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "02ccf358-42e4-48bc-9748-de9e2c46e2b5" (UID: "02ccf358-42e4-48bc-9748-de9e2c46e2b5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:09:54 crc kubenswrapper[4789]: I1122 08:09:54.369285 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02ccf358-42e4-48bc-9748-de9e2c46e2b5-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 08:09:54 crc kubenswrapper[4789]: I1122 08:09:54.369336 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02ccf358-42e4-48bc-9748-de9e2c46e2b5-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 08:09:54 crc kubenswrapper[4789]: I1122 08:09:54.369359 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rdg68\" (UniqueName: \"kubernetes.io/projected/02ccf358-42e4-48bc-9748-de9e2c46e2b5-kube-api-access-rdg68\") on node \"crc\" DevicePath \"\"" Nov 22 08:09:54 crc kubenswrapper[4789]: I1122 08:09:54.708515 4789 generic.go:334] "Generic (PLEG): container finished" podID="02ccf358-42e4-48bc-9748-de9e2c46e2b5" containerID="9acec1deb69b06863e9efed3a816d35d4d8e45becf4da0a9a2cbd3ccf162a98f" exitCode=0 Nov 22 08:09:54 crc kubenswrapper[4789]: I1122 08:09:54.708584 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gvdqx" Nov 22 08:09:54 crc kubenswrapper[4789]: I1122 08:09:54.708824 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gvdqx" event={"ID":"02ccf358-42e4-48bc-9748-de9e2c46e2b5","Type":"ContainerDied","Data":"9acec1deb69b06863e9efed3a816d35d4d8e45becf4da0a9a2cbd3ccf162a98f"} Nov 22 08:09:54 crc kubenswrapper[4789]: I1122 08:09:54.708891 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gvdqx" event={"ID":"02ccf358-42e4-48bc-9748-de9e2c46e2b5","Type":"ContainerDied","Data":"c84b01e4a2bbd9279664c59069b740c757e543b91d21dd9250e1668cb5575b6c"} Nov 22 08:09:54 crc kubenswrapper[4789]: I1122 08:09:54.708921 4789 scope.go:117] "RemoveContainer" containerID="9acec1deb69b06863e9efed3a816d35d4d8e45becf4da0a9a2cbd3ccf162a98f" Nov 22 08:09:54 crc kubenswrapper[4789]: I1122 08:09:54.737583 4789 scope.go:117] "RemoveContainer" containerID="f1edced96ee24f085cf1b90b51e8f1a237f7df87fac0ad16e2a711fb5ba1e12b" Nov 22 08:09:54 crc kubenswrapper[4789]: I1122 08:09:54.740563 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gvdqx"] Nov 22 08:09:54 crc kubenswrapper[4789]: I1122 08:09:54.753368 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-gvdqx"] Nov 22 08:09:54 crc kubenswrapper[4789]: I1122 08:09:54.769898 4789 scope.go:117] "RemoveContainer" containerID="14b0b12f7ba5a10d0e389a88a9f09a496bfc0b2ded7fd2b8a0e5e3167ab448ec" Nov 22 08:09:54 crc kubenswrapper[4789]: I1122 08:09:54.792076 4789 scope.go:117] "RemoveContainer" containerID="9acec1deb69b06863e9efed3a816d35d4d8e45becf4da0a9a2cbd3ccf162a98f" Nov 22 08:09:54 crc kubenswrapper[4789]: E1122 08:09:54.792476 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9acec1deb69b06863e9efed3a816d35d4d8e45becf4da0a9a2cbd3ccf162a98f\": container with ID starting with 9acec1deb69b06863e9efed3a816d35d4d8e45becf4da0a9a2cbd3ccf162a98f not found: ID does not exist" containerID="9acec1deb69b06863e9efed3a816d35d4d8e45becf4da0a9a2cbd3ccf162a98f" Nov 22 08:09:54 crc kubenswrapper[4789]: I1122 08:09:54.792567 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9acec1deb69b06863e9efed3a816d35d4d8e45becf4da0a9a2cbd3ccf162a98f"} err="failed to get container status \"9acec1deb69b06863e9efed3a816d35d4d8e45becf4da0a9a2cbd3ccf162a98f\": rpc error: code = NotFound desc = could not find container \"9acec1deb69b06863e9efed3a816d35d4d8e45becf4da0a9a2cbd3ccf162a98f\": container with ID starting with 9acec1deb69b06863e9efed3a816d35d4d8e45becf4da0a9a2cbd3ccf162a98f not found: ID does not exist" Nov 22 08:09:54 crc kubenswrapper[4789]: I1122 08:09:54.792651 4789 scope.go:117] "RemoveContainer" containerID="f1edced96ee24f085cf1b90b51e8f1a237f7df87fac0ad16e2a711fb5ba1e12b" Nov 22 08:09:54 crc kubenswrapper[4789]: E1122 08:09:54.793033 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f1edced96ee24f085cf1b90b51e8f1a237f7df87fac0ad16e2a711fb5ba1e12b\": container with ID starting with f1edced96ee24f085cf1b90b51e8f1a237f7df87fac0ad16e2a711fb5ba1e12b not found: ID does not exist" containerID="f1edced96ee24f085cf1b90b51e8f1a237f7df87fac0ad16e2a711fb5ba1e12b" Nov 22 08:09:54 crc kubenswrapper[4789]: I1122 08:09:54.793101 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f1edced96ee24f085cf1b90b51e8f1a237f7df87fac0ad16e2a711fb5ba1e12b"} err="failed to get container status \"f1edced96ee24f085cf1b90b51e8f1a237f7df87fac0ad16e2a711fb5ba1e12b\": rpc error: code = NotFound desc = could not find container \"f1edced96ee24f085cf1b90b51e8f1a237f7df87fac0ad16e2a711fb5ba1e12b\": container with ID starting with f1edced96ee24f085cf1b90b51e8f1a237f7df87fac0ad16e2a711fb5ba1e12b not found: ID does not exist" Nov 22 08:09:54 crc kubenswrapper[4789]: I1122 08:09:54.793144 4789 scope.go:117] "RemoveContainer" containerID="14b0b12f7ba5a10d0e389a88a9f09a496bfc0b2ded7fd2b8a0e5e3167ab448ec" Nov 22 08:09:54 crc kubenswrapper[4789]: E1122 08:09:54.793554 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"14b0b12f7ba5a10d0e389a88a9f09a496bfc0b2ded7fd2b8a0e5e3167ab448ec\": container with ID starting with 14b0b12f7ba5a10d0e389a88a9f09a496bfc0b2ded7fd2b8a0e5e3167ab448ec not found: ID does not exist" containerID="14b0b12f7ba5a10d0e389a88a9f09a496bfc0b2ded7fd2b8a0e5e3167ab448ec" Nov 22 08:09:54 crc kubenswrapper[4789]: I1122 08:09:54.793594 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"14b0b12f7ba5a10d0e389a88a9f09a496bfc0b2ded7fd2b8a0e5e3167ab448ec"} err="failed to get container status \"14b0b12f7ba5a10d0e389a88a9f09a496bfc0b2ded7fd2b8a0e5e3167ab448ec\": rpc error: code = NotFound desc = could not find container \"14b0b12f7ba5a10d0e389a88a9f09a496bfc0b2ded7fd2b8a0e5e3167ab448ec\": container with ID starting with 14b0b12f7ba5a10d0e389a88a9f09a496bfc0b2ded7fd2b8a0e5e3167ab448ec not found: ID does not exist" Nov 22 08:09:55 crc kubenswrapper[4789]: I1122 08:09:55.972885 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="02ccf358-42e4-48bc-9748-de9e2c46e2b5" path="/var/lib/kubelet/pods/02ccf358-42e4-48bc-9748-de9e2c46e2b5/volumes" Nov 22 08:09:56 crc kubenswrapper[4789]: I1122 08:09:56.279060 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-5dcf9c57c5-t6j77"] Nov 22 08:09:56 crc kubenswrapper[4789]: E1122 08:09:56.279312 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02ccf358-42e4-48bc-9748-de9e2c46e2b5" containerName="extract-utilities" Nov 22 08:09:56 crc kubenswrapper[4789]: I1122 08:09:56.279355 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="02ccf358-42e4-48bc-9748-de9e2c46e2b5" containerName="extract-utilities" Nov 22 08:09:56 crc kubenswrapper[4789]: E1122 08:09:56.279381 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02ccf358-42e4-48bc-9748-de9e2c46e2b5" containerName="registry-server" Nov 22 08:09:56 crc kubenswrapper[4789]: I1122 08:09:56.279389 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="02ccf358-42e4-48bc-9748-de9e2c46e2b5" containerName="registry-server" Nov 22 08:09:56 crc kubenswrapper[4789]: E1122 08:09:56.279407 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02ccf358-42e4-48bc-9748-de9e2c46e2b5" containerName="extract-content" Nov 22 08:09:56 crc kubenswrapper[4789]: I1122 08:09:56.279416 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="02ccf358-42e4-48bc-9748-de9e2c46e2b5" containerName="extract-content" Nov 22 08:09:56 crc kubenswrapper[4789]: I1122 08:09:56.279535 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="02ccf358-42e4-48bc-9748-de9e2c46e2b5" containerName="registry-server" Nov 22 08:09:56 crc kubenswrapper[4789]: I1122 08:09:56.280224 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-t6j77" Nov 22 08:09:56 crc kubenswrapper[4789]: I1122 08:09:56.298112 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-5dcf9c57c5-t6j77"] Nov 22 08:09:56 crc kubenswrapper[4789]: I1122 08:09:56.300776 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-lvmvm" Nov 22 08:09:56 crc kubenswrapper[4789]: I1122 08:09:56.304427 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-hccc7"] Nov 22 08:09:56 crc kubenswrapper[4789]: I1122 08:09:56.305287 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-hccc7" Nov 22 08:09:56 crc kubenswrapper[4789]: I1122 08:09:56.310499 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-6b89b748d8-59p65"] Nov 22 08:09:56 crc kubenswrapper[4789]: I1122 08:09:56.311485 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-59p65" Nov 22 08:09:56 crc kubenswrapper[4789]: I1122 08:09:56.316124 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Nov 22 08:09:56 crc kubenswrapper[4789]: I1122 08:09:56.323862 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6b89b748d8-59p65"] Nov 22 08:09:56 crc kubenswrapper[4789]: I1122 08:09:56.398814 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/d94d3cb7-9188-42a2-b911-4f7665349830-nmstate-lock\") pod \"nmstate-handler-hccc7\" (UID: \"d94d3cb7-9188-42a2-b911-4f7665349830\") " pod="openshift-nmstate/nmstate-handler-hccc7" Nov 22 08:09:56 crc kubenswrapper[4789]: I1122 08:09:56.398864 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mlhpw\" (UniqueName: \"kubernetes.io/projected/4dc70724-5816-4152-8d2f-763faeddd54f-kube-api-access-mlhpw\") pod \"nmstate-webhook-6b89b748d8-59p65\" (UID: \"4dc70724-5816-4152-8d2f-763faeddd54f\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-59p65" Nov 22 08:09:56 crc kubenswrapper[4789]: I1122 08:09:56.398906 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t4567\" (UniqueName: \"kubernetes.io/projected/d94d3cb7-9188-42a2-b911-4f7665349830-kube-api-access-t4567\") pod \"nmstate-handler-hccc7\" (UID: \"d94d3cb7-9188-42a2-b911-4f7665349830\") " pod="openshift-nmstate/nmstate-handler-hccc7" Nov 22 08:09:56 crc kubenswrapper[4789]: I1122 08:09:56.398925 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/d94d3cb7-9188-42a2-b911-4f7665349830-dbus-socket\") pod \"nmstate-handler-hccc7\" (UID: \"d94d3cb7-9188-42a2-b911-4f7665349830\") " pod="openshift-nmstate/nmstate-handler-hccc7" Nov 22 08:09:56 crc kubenswrapper[4789]: I1122 08:09:56.398943 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6vmvw\" (UniqueName: \"kubernetes.io/projected/15f327f3-5cc9-4ce4-b291-efdffea58145-kube-api-access-6vmvw\") pod \"nmstate-metrics-5dcf9c57c5-t6j77\" (UID: \"15f327f3-5cc9-4ce4-b291-efdffea58145\") " pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-t6j77" Nov 22 08:09:56 crc kubenswrapper[4789]: I1122 08:09:56.398957 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/d94d3cb7-9188-42a2-b911-4f7665349830-ovs-socket\") pod \"nmstate-handler-hccc7\" (UID: \"d94d3cb7-9188-42a2-b911-4f7665349830\") " pod="openshift-nmstate/nmstate-handler-hccc7" Nov 22 08:09:56 crc kubenswrapper[4789]: I1122 08:09:56.398989 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/4dc70724-5816-4152-8d2f-763faeddd54f-tls-key-pair\") pod \"nmstate-webhook-6b89b748d8-59p65\" (UID: \"4dc70724-5816-4152-8d2f-763faeddd54f\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-59p65" Nov 22 08:09:56 crc kubenswrapper[4789]: I1122 08:09:56.410772 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5874bd7bc5-pfnkv"] Nov 22 08:09:56 crc kubenswrapper[4789]: I1122 08:09:56.411395 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-pfnkv" Nov 22 08:09:56 crc kubenswrapper[4789]: I1122 08:09:56.419772 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-b6bkl" Nov 22 08:09:56 crc kubenswrapper[4789]: I1122 08:09:56.419994 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Nov 22 08:09:56 crc kubenswrapper[4789]: I1122 08:09:56.424184 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Nov 22 08:09:56 crc kubenswrapper[4789]: I1122 08:09:56.434469 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5874bd7bc5-pfnkv"] Nov 22 08:09:56 crc kubenswrapper[4789]: I1122 08:09:56.500586 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/d94d3cb7-9188-42a2-b911-4f7665349830-nmstate-lock\") pod \"nmstate-handler-hccc7\" (UID: \"d94d3cb7-9188-42a2-b911-4f7665349830\") " pod="openshift-nmstate/nmstate-handler-hccc7" Nov 22 08:09:56 crc kubenswrapper[4789]: I1122 08:09:56.500691 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mlhpw\" (UniqueName: \"kubernetes.io/projected/4dc70724-5816-4152-8d2f-763faeddd54f-kube-api-access-mlhpw\") pod \"nmstate-webhook-6b89b748d8-59p65\" (UID: \"4dc70724-5816-4152-8d2f-763faeddd54f\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-59p65" Nov 22 08:09:56 crc kubenswrapper[4789]: I1122 08:09:56.500729 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/d94d3cb7-9188-42a2-b911-4f7665349830-nmstate-lock\") pod \"nmstate-handler-hccc7\" (UID: \"d94d3cb7-9188-42a2-b911-4f7665349830\") " pod="openshift-nmstate/nmstate-handler-hccc7" Nov 22 08:09:56 crc kubenswrapper[4789]: I1122 08:09:56.501357 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t4567\" (UniqueName: \"kubernetes.io/projected/d94d3cb7-9188-42a2-b911-4f7665349830-kube-api-access-t4567\") pod \"nmstate-handler-hccc7\" (UID: \"d94d3cb7-9188-42a2-b911-4f7665349830\") " pod="openshift-nmstate/nmstate-handler-hccc7" Nov 22 08:09:56 crc kubenswrapper[4789]: I1122 08:09:56.501676 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/d94d3cb7-9188-42a2-b911-4f7665349830-dbus-socket\") pod \"nmstate-handler-hccc7\" (UID: \"d94d3cb7-9188-42a2-b911-4f7665349830\") " pod="openshift-nmstate/nmstate-handler-hccc7" Nov 22 08:09:56 crc kubenswrapper[4789]: I1122 08:09:56.501725 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6vmvw\" (UniqueName: \"kubernetes.io/projected/15f327f3-5cc9-4ce4-b291-efdffea58145-kube-api-access-6vmvw\") pod \"nmstate-metrics-5dcf9c57c5-t6j77\" (UID: \"15f327f3-5cc9-4ce4-b291-efdffea58145\") " pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-t6j77" Nov 22 08:09:56 crc kubenswrapper[4789]: I1122 08:09:56.501979 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/d94d3cb7-9188-42a2-b911-4f7665349830-dbus-socket\") pod \"nmstate-handler-hccc7\" (UID: \"d94d3cb7-9188-42a2-b911-4f7665349830\") " pod="openshift-nmstate/nmstate-handler-hccc7" Nov 22 08:09:56 crc kubenswrapper[4789]: I1122 08:09:56.501995 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/d94d3cb7-9188-42a2-b911-4f7665349830-ovs-socket\") pod \"nmstate-handler-hccc7\" (UID: \"d94d3cb7-9188-42a2-b911-4f7665349830\") " pod="openshift-nmstate/nmstate-handler-hccc7" Nov 22 08:09:56 crc kubenswrapper[4789]: I1122 08:09:56.502050 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/d94d3cb7-9188-42a2-b911-4f7665349830-ovs-socket\") pod \"nmstate-handler-hccc7\" (UID: \"d94d3cb7-9188-42a2-b911-4f7665349830\") " pod="openshift-nmstate/nmstate-handler-hccc7" Nov 22 08:09:56 crc kubenswrapper[4789]: I1122 08:09:56.502072 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/c88315b2-ec94-4272-91f1-3127418ae55c-nginx-conf\") pod \"nmstate-console-plugin-5874bd7bc5-pfnkv\" (UID: \"c88315b2-ec94-4272-91f1-3127418ae55c\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-pfnkv" Nov 22 08:09:56 crc kubenswrapper[4789]: I1122 08:09:56.502111 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/c88315b2-ec94-4272-91f1-3127418ae55c-plugin-serving-cert\") pod \"nmstate-console-plugin-5874bd7bc5-pfnkv\" (UID: \"c88315b2-ec94-4272-91f1-3127418ae55c\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-pfnkv" Nov 22 08:09:56 crc kubenswrapper[4789]: I1122 08:09:56.502170 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/4dc70724-5816-4152-8d2f-763faeddd54f-tls-key-pair\") pod \"nmstate-webhook-6b89b748d8-59p65\" (UID: \"4dc70724-5816-4152-8d2f-763faeddd54f\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-59p65" Nov 22 08:09:56 crc kubenswrapper[4789]: I1122 08:09:56.502231 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4rlhf\" (UniqueName: \"kubernetes.io/projected/c88315b2-ec94-4272-91f1-3127418ae55c-kube-api-access-4rlhf\") pod \"nmstate-console-plugin-5874bd7bc5-pfnkv\" (UID: \"c88315b2-ec94-4272-91f1-3127418ae55c\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-pfnkv" Nov 22 08:09:56 crc kubenswrapper[4789]: I1122 08:09:56.512085 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/4dc70724-5816-4152-8d2f-763faeddd54f-tls-key-pair\") pod \"nmstate-webhook-6b89b748d8-59p65\" (UID: \"4dc70724-5816-4152-8d2f-763faeddd54f\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-59p65" Nov 22 08:09:56 crc kubenswrapper[4789]: I1122 08:09:56.522911 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mlhpw\" (UniqueName: \"kubernetes.io/projected/4dc70724-5816-4152-8d2f-763faeddd54f-kube-api-access-mlhpw\") pod \"nmstate-webhook-6b89b748d8-59p65\" (UID: \"4dc70724-5816-4152-8d2f-763faeddd54f\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-59p65" Nov 22 08:09:56 crc kubenswrapper[4789]: I1122 08:09:56.528900 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6vmvw\" (UniqueName: \"kubernetes.io/projected/15f327f3-5cc9-4ce4-b291-efdffea58145-kube-api-access-6vmvw\") pod \"nmstate-metrics-5dcf9c57c5-t6j77\" (UID: \"15f327f3-5cc9-4ce4-b291-efdffea58145\") " pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-t6j77" Nov 22 08:09:56 crc kubenswrapper[4789]: I1122 08:09:56.531366 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t4567\" (UniqueName: \"kubernetes.io/projected/d94d3cb7-9188-42a2-b911-4f7665349830-kube-api-access-t4567\") pod \"nmstate-handler-hccc7\" (UID: \"d94d3cb7-9188-42a2-b911-4f7665349830\") " pod="openshift-nmstate/nmstate-handler-hccc7" Nov 22 08:09:56 crc kubenswrapper[4789]: I1122 08:09:56.557670 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-b2vc5"] Nov 22 08:09:56 crc kubenswrapper[4789]: I1122 08:09:56.557991 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-b2vc5" podUID="3009aea9-8c5a-42cc-b9d6-dca6ce289616" containerName="registry-server" containerID="cri-o://118ff9b20cd28a73ebbfeaea76cf9386d79de7513149c06e7011cfaeb91de3ba" gracePeriod=2 Nov 22 08:09:56 crc kubenswrapper[4789]: I1122 08:09:56.598803 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-t6j77" Nov 22 08:09:56 crc kubenswrapper[4789]: I1122 08:09:56.604892 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4rlhf\" (UniqueName: \"kubernetes.io/projected/c88315b2-ec94-4272-91f1-3127418ae55c-kube-api-access-4rlhf\") pod \"nmstate-console-plugin-5874bd7bc5-pfnkv\" (UID: \"c88315b2-ec94-4272-91f1-3127418ae55c\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-pfnkv" Nov 22 08:09:56 crc kubenswrapper[4789]: I1122 08:09:56.605023 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/c88315b2-ec94-4272-91f1-3127418ae55c-nginx-conf\") pod \"nmstate-console-plugin-5874bd7bc5-pfnkv\" (UID: \"c88315b2-ec94-4272-91f1-3127418ae55c\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-pfnkv" Nov 22 08:09:56 crc kubenswrapper[4789]: I1122 08:09:56.605064 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/c88315b2-ec94-4272-91f1-3127418ae55c-plugin-serving-cert\") pod \"nmstate-console-plugin-5874bd7bc5-pfnkv\" (UID: \"c88315b2-ec94-4272-91f1-3127418ae55c\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-pfnkv" Nov 22 08:09:56 crc kubenswrapper[4789]: E1122 08:09:56.605236 4789 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Nov 22 08:09:56 crc kubenswrapper[4789]: E1122 08:09:56.605334 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c88315b2-ec94-4272-91f1-3127418ae55c-plugin-serving-cert podName:c88315b2-ec94-4272-91f1-3127418ae55c nodeName:}" failed. No retries permitted until 2025-11-22 08:09:57.105298446 +0000 UTC m=+811.339698719 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/c88315b2-ec94-4272-91f1-3127418ae55c-plugin-serving-cert") pod "nmstate-console-plugin-5874bd7bc5-pfnkv" (UID: "c88315b2-ec94-4272-91f1-3127418ae55c") : secret "plugin-serving-cert" not found Nov 22 08:09:56 crc kubenswrapper[4789]: I1122 08:09:56.606961 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/c88315b2-ec94-4272-91f1-3127418ae55c-nginx-conf\") pod \"nmstate-console-plugin-5874bd7bc5-pfnkv\" (UID: \"c88315b2-ec94-4272-91f1-3127418ae55c\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-pfnkv" Nov 22 08:09:56 crc kubenswrapper[4789]: I1122 08:09:56.623856 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4rlhf\" (UniqueName: \"kubernetes.io/projected/c88315b2-ec94-4272-91f1-3127418ae55c-kube-api-access-4rlhf\") pod \"nmstate-console-plugin-5874bd7bc5-pfnkv\" (UID: \"c88315b2-ec94-4272-91f1-3127418ae55c\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-pfnkv" Nov 22 08:09:56 crc kubenswrapper[4789]: I1122 08:09:56.641812 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-hccc7" Nov 22 08:09:56 crc kubenswrapper[4789]: I1122 08:09:56.645515 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-59p65" Nov 22 08:09:56 crc kubenswrapper[4789]: I1122 08:09:56.650233 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-5c449cb8b8-9r629"] Nov 22 08:09:56 crc kubenswrapper[4789]: I1122 08:09:56.651169 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5c449cb8b8-9r629" Nov 22 08:09:56 crc kubenswrapper[4789]: I1122 08:09:56.679463 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-5c449cb8b8-9r629"] Nov 22 08:09:56 crc kubenswrapper[4789]: W1122 08:09:56.762116 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd94d3cb7_9188_42a2_b911_4f7665349830.slice/crio-bf046adf56a5d3cd1c368227a1fb9376da816effdbc5f25061726bb6b332c3bc WatchSource:0}: Error finding container bf046adf56a5d3cd1c368227a1fb9376da816effdbc5f25061726bb6b332c3bc: Status 404 returned error can't find the container with id bf046adf56a5d3cd1c368227a1fb9376da816effdbc5f25061726bb6b332c3bc Nov 22 08:09:56 crc kubenswrapper[4789]: I1122 08:09:56.796101 4789 generic.go:334] "Generic (PLEG): container finished" podID="3009aea9-8c5a-42cc-b9d6-dca6ce289616" containerID="118ff9b20cd28a73ebbfeaea76cf9386d79de7513149c06e7011cfaeb91de3ba" exitCode=0 Nov 22 08:09:56 crc kubenswrapper[4789]: I1122 08:09:56.796148 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b2vc5" event={"ID":"3009aea9-8c5a-42cc-b9d6-dca6ce289616","Type":"ContainerDied","Data":"118ff9b20cd28a73ebbfeaea76cf9386d79de7513149c06e7011cfaeb91de3ba"} Nov 22 08:09:56 crc kubenswrapper[4789]: I1122 08:09:56.807623 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/62e3cf19-cd3a-438b-978b-0f5f300506dc-trusted-ca-bundle\") pod \"console-5c449cb8b8-9r629\" (UID: \"62e3cf19-cd3a-438b-978b-0f5f300506dc\") " pod="openshift-console/console-5c449cb8b8-9r629" Nov 22 08:09:56 crc kubenswrapper[4789]: I1122 08:09:56.807669 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/62e3cf19-cd3a-438b-978b-0f5f300506dc-console-oauth-config\") pod \"console-5c449cb8b8-9r629\" (UID: \"62e3cf19-cd3a-438b-978b-0f5f300506dc\") " pod="openshift-console/console-5c449cb8b8-9r629" Nov 22 08:09:56 crc kubenswrapper[4789]: I1122 08:09:56.807692 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/62e3cf19-cd3a-438b-978b-0f5f300506dc-console-config\") pod \"console-5c449cb8b8-9r629\" (UID: \"62e3cf19-cd3a-438b-978b-0f5f300506dc\") " pod="openshift-console/console-5c449cb8b8-9r629" Nov 22 08:09:56 crc kubenswrapper[4789]: I1122 08:09:56.807921 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/62e3cf19-cd3a-438b-978b-0f5f300506dc-oauth-serving-cert\") pod \"console-5c449cb8b8-9r629\" (UID: \"62e3cf19-cd3a-438b-978b-0f5f300506dc\") " pod="openshift-console/console-5c449cb8b8-9r629" Nov 22 08:09:56 crc kubenswrapper[4789]: I1122 08:09:56.807968 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/62e3cf19-cd3a-438b-978b-0f5f300506dc-console-serving-cert\") pod \"console-5c449cb8b8-9r629\" (UID: \"62e3cf19-cd3a-438b-978b-0f5f300506dc\") " pod="openshift-console/console-5c449cb8b8-9r629" Nov 22 08:09:56 crc kubenswrapper[4789]: I1122 08:09:56.807998 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/62e3cf19-cd3a-438b-978b-0f5f300506dc-service-ca\") pod \"console-5c449cb8b8-9r629\" (UID: \"62e3cf19-cd3a-438b-978b-0f5f300506dc\") " pod="openshift-console/console-5c449cb8b8-9r629" Nov 22 08:09:56 crc kubenswrapper[4789]: I1122 08:09:56.808074 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w4z4n\" (UniqueName: \"kubernetes.io/projected/62e3cf19-cd3a-438b-978b-0f5f300506dc-kube-api-access-w4z4n\") pod \"console-5c449cb8b8-9r629\" (UID: \"62e3cf19-cd3a-438b-978b-0f5f300506dc\") " pod="openshift-console/console-5c449cb8b8-9r629" Nov 22 08:09:56 crc kubenswrapper[4789]: I1122 08:09:56.910484 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/62e3cf19-cd3a-438b-978b-0f5f300506dc-trusted-ca-bundle\") pod \"console-5c449cb8b8-9r629\" (UID: \"62e3cf19-cd3a-438b-978b-0f5f300506dc\") " pod="openshift-console/console-5c449cb8b8-9r629" Nov 22 08:09:56 crc kubenswrapper[4789]: I1122 08:09:56.910526 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/62e3cf19-cd3a-438b-978b-0f5f300506dc-console-oauth-config\") pod \"console-5c449cb8b8-9r629\" (UID: \"62e3cf19-cd3a-438b-978b-0f5f300506dc\") " pod="openshift-console/console-5c449cb8b8-9r629" Nov 22 08:09:56 crc kubenswrapper[4789]: I1122 08:09:56.910543 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/62e3cf19-cd3a-438b-978b-0f5f300506dc-console-config\") pod \"console-5c449cb8b8-9r629\" (UID: \"62e3cf19-cd3a-438b-978b-0f5f300506dc\") " pod="openshift-console/console-5c449cb8b8-9r629" Nov 22 08:09:56 crc kubenswrapper[4789]: I1122 08:09:56.910614 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/62e3cf19-cd3a-438b-978b-0f5f300506dc-oauth-serving-cert\") pod \"console-5c449cb8b8-9r629\" (UID: \"62e3cf19-cd3a-438b-978b-0f5f300506dc\") " pod="openshift-console/console-5c449cb8b8-9r629" Nov 22 08:09:56 crc kubenswrapper[4789]: I1122 08:09:56.910638 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/62e3cf19-cd3a-438b-978b-0f5f300506dc-console-serving-cert\") pod \"console-5c449cb8b8-9r629\" (UID: \"62e3cf19-cd3a-438b-978b-0f5f300506dc\") " pod="openshift-console/console-5c449cb8b8-9r629" Nov 22 08:09:56 crc kubenswrapper[4789]: I1122 08:09:56.910659 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/62e3cf19-cd3a-438b-978b-0f5f300506dc-service-ca\") pod \"console-5c449cb8b8-9r629\" (UID: \"62e3cf19-cd3a-438b-978b-0f5f300506dc\") " pod="openshift-console/console-5c449cb8b8-9r629" Nov 22 08:09:56 crc kubenswrapper[4789]: I1122 08:09:56.910700 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w4z4n\" (UniqueName: \"kubernetes.io/projected/62e3cf19-cd3a-438b-978b-0f5f300506dc-kube-api-access-w4z4n\") pod \"console-5c449cb8b8-9r629\" (UID: \"62e3cf19-cd3a-438b-978b-0f5f300506dc\") " pod="openshift-console/console-5c449cb8b8-9r629" Nov 22 08:09:56 crc kubenswrapper[4789]: I1122 08:09:56.912536 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/62e3cf19-cd3a-438b-978b-0f5f300506dc-trusted-ca-bundle\") pod \"console-5c449cb8b8-9r629\" (UID: \"62e3cf19-cd3a-438b-978b-0f5f300506dc\") " pod="openshift-console/console-5c449cb8b8-9r629" Nov 22 08:09:56 crc kubenswrapper[4789]: I1122 08:09:56.913221 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/62e3cf19-cd3a-438b-978b-0f5f300506dc-console-config\") pod \"console-5c449cb8b8-9r629\" (UID: \"62e3cf19-cd3a-438b-978b-0f5f300506dc\") " pod="openshift-console/console-5c449cb8b8-9r629" Nov 22 08:09:56 crc kubenswrapper[4789]: I1122 08:09:56.913899 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/62e3cf19-cd3a-438b-978b-0f5f300506dc-service-ca\") pod \"console-5c449cb8b8-9r629\" (UID: \"62e3cf19-cd3a-438b-978b-0f5f300506dc\") " pod="openshift-console/console-5c449cb8b8-9r629" Nov 22 08:09:56 crc kubenswrapper[4789]: I1122 08:09:56.914367 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/62e3cf19-cd3a-438b-978b-0f5f300506dc-oauth-serving-cert\") pod \"console-5c449cb8b8-9r629\" (UID: \"62e3cf19-cd3a-438b-978b-0f5f300506dc\") " pod="openshift-console/console-5c449cb8b8-9r629" Nov 22 08:09:56 crc kubenswrapper[4789]: I1122 08:09:56.916311 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/62e3cf19-cd3a-438b-978b-0f5f300506dc-console-oauth-config\") pod \"console-5c449cb8b8-9r629\" (UID: \"62e3cf19-cd3a-438b-978b-0f5f300506dc\") " pod="openshift-console/console-5c449cb8b8-9r629" Nov 22 08:09:56 crc kubenswrapper[4789]: I1122 08:09:56.919331 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6b89b748d8-59p65"] Nov 22 08:09:56 crc kubenswrapper[4789]: I1122 08:09:56.920408 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/62e3cf19-cd3a-438b-978b-0f5f300506dc-console-serving-cert\") pod \"console-5c449cb8b8-9r629\" (UID: \"62e3cf19-cd3a-438b-978b-0f5f300506dc\") " pod="openshift-console/console-5c449cb8b8-9r629" Nov 22 08:09:56 crc kubenswrapper[4789]: W1122 08:09:56.925971 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4dc70724_5816_4152_8d2f_763faeddd54f.slice/crio-e7cb7c4807f19d5008e0c376bda0e32017607b652c67603d36b8231f94abef82 WatchSource:0}: Error finding container e7cb7c4807f19d5008e0c376bda0e32017607b652c67603d36b8231f94abef82: Status 404 returned error can't find the container with id e7cb7c4807f19d5008e0c376bda0e32017607b652c67603d36b8231f94abef82 Nov 22 08:09:56 crc kubenswrapper[4789]: I1122 08:09:56.931710 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w4z4n\" (UniqueName: \"kubernetes.io/projected/62e3cf19-cd3a-438b-978b-0f5f300506dc-kube-api-access-w4z4n\") pod \"console-5c449cb8b8-9r629\" (UID: \"62e3cf19-cd3a-438b-978b-0f5f300506dc\") " pod="openshift-console/console-5c449cb8b8-9r629" Nov 22 08:09:56 crc kubenswrapper[4789]: I1122 08:09:56.968068 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b2vc5" Nov 22 08:09:57 crc kubenswrapper[4789]: I1122 08:09:57.067978 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-5dcf9c57c5-t6j77"] Nov 22 08:09:57 crc kubenswrapper[4789]: W1122 08:09:57.072505 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod15f327f3_5cc9_4ce4_b291_efdffea58145.slice/crio-76cb1ac32664e15f9935bce63e9fcc4fdd216dc38c7657c9f81ecac674fd1e54 WatchSource:0}: Error finding container 76cb1ac32664e15f9935bce63e9fcc4fdd216dc38c7657c9f81ecac674fd1e54: Status 404 returned error can't find the container with id 76cb1ac32664e15f9935bce63e9fcc4fdd216dc38c7657c9f81ecac674fd1e54 Nov 22 08:09:57 crc kubenswrapper[4789]: I1122 08:09:57.077290 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5c449cb8b8-9r629" Nov 22 08:09:57 crc kubenswrapper[4789]: I1122 08:09:57.112242 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3009aea9-8c5a-42cc-b9d6-dca6ce289616-catalog-content\") pod \"3009aea9-8c5a-42cc-b9d6-dca6ce289616\" (UID: \"3009aea9-8c5a-42cc-b9d6-dca6ce289616\") " Nov 22 08:09:57 crc kubenswrapper[4789]: I1122 08:09:57.112323 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dv64c\" (UniqueName: \"kubernetes.io/projected/3009aea9-8c5a-42cc-b9d6-dca6ce289616-kube-api-access-dv64c\") pod \"3009aea9-8c5a-42cc-b9d6-dca6ce289616\" (UID: \"3009aea9-8c5a-42cc-b9d6-dca6ce289616\") " Nov 22 08:09:57 crc kubenswrapper[4789]: I1122 08:09:57.112511 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3009aea9-8c5a-42cc-b9d6-dca6ce289616-utilities\") pod \"3009aea9-8c5a-42cc-b9d6-dca6ce289616\" (UID: \"3009aea9-8c5a-42cc-b9d6-dca6ce289616\") " Nov 22 08:09:57 crc kubenswrapper[4789]: I1122 08:09:57.113211 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/c88315b2-ec94-4272-91f1-3127418ae55c-plugin-serving-cert\") pod \"nmstate-console-plugin-5874bd7bc5-pfnkv\" (UID: \"c88315b2-ec94-4272-91f1-3127418ae55c\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-pfnkv" Nov 22 08:09:57 crc kubenswrapper[4789]: I1122 08:09:57.114252 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3009aea9-8c5a-42cc-b9d6-dca6ce289616-utilities" (OuterVolumeSpecName: "utilities") pod "3009aea9-8c5a-42cc-b9d6-dca6ce289616" (UID: "3009aea9-8c5a-42cc-b9d6-dca6ce289616"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:09:57 crc kubenswrapper[4789]: I1122 08:09:57.116891 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3009aea9-8c5a-42cc-b9d6-dca6ce289616-kube-api-access-dv64c" (OuterVolumeSpecName: "kube-api-access-dv64c") pod "3009aea9-8c5a-42cc-b9d6-dca6ce289616" (UID: "3009aea9-8c5a-42cc-b9d6-dca6ce289616"). InnerVolumeSpecName "kube-api-access-dv64c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:09:57 crc kubenswrapper[4789]: I1122 08:09:57.117253 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/c88315b2-ec94-4272-91f1-3127418ae55c-plugin-serving-cert\") pod \"nmstate-console-plugin-5874bd7bc5-pfnkv\" (UID: \"c88315b2-ec94-4272-91f1-3127418ae55c\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-pfnkv" Nov 22 08:09:57 crc kubenswrapper[4789]: I1122 08:09:57.131417 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3009aea9-8c5a-42cc-b9d6-dca6ce289616-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3009aea9-8c5a-42cc-b9d6-dca6ce289616" (UID: "3009aea9-8c5a-42cc-b9d6-dca6ce289616"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:09:57 crc kubenswrapper[4789]: I1122 08:09:57.214176 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3009aea9-8c5a-42cc-b9d6-dca6ce289616-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 08:09:57 crc kubenswrapper[4789]: I1122 08:09:57.214504 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dv64c\" (UniqueName: \"kubernetes.io/projected/3009aea9-8c5a-42cc-b9d6-dca6ce289616-kube-api-access-dv64c\") on node \"crc\" DevicePath \"\"" Nov 22 08:09:57 crc kubenswrapper[4789]: I1122 08:09:57.214517 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3009aea9-8c5a-42cc-b9d6-dca6ce289616-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 08:09:57 crc kubenswrapper[4789]: I1122 08:09:57.236158 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-5c449cb8b8-9r629"] Nov 22 08:09:57 crc kubenswrapper[4789]: W1122 08:09:57.241614 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod62e3cf19_cd3a_438b_978b_0f5f300506dc.slice/crio-5734b406ef9b4c4dbb32f97d639fb49d5e7852186f6c7df808fcb8b1fc2398c6 WatchSource:0}: Error finding container 5734b406ef9b4c4dbb32f97d639fb49d5e7852186f6c7df808fcb8b1fc2398c6: Status 404 returned error can't find the container with id 5734b406ef9b4c4dbb32f97d639fb49d5e7852186f6c7df808fcb8b1fc2398c6 Nov 22 08:09:57 crc kubenswrapper[4789]: I1122 08:09:57.326594 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-pfnkv" Nov 22 08:09:57 crc kubenswrapper[4789]: I1122 08:09:57.512595 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5874bd7bc5-pfnkv"] Nov 22 08:09:57 crc kubenswrapper[4789]: W1122 08:09:57.523459 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc88315b2_ec94_4272_91f1_3127418ae55c.slice/crio-6ff7d70627ee9767a5a121f0e878224ff1ebf6f6dbe47d1b22e364deb2261817 WatchSource:0}: Error finding container 6ff7d70627ee9767a5a121f0e878224ff1ebf6f6dbe47d1b22e364deb2261817: Status 404 returned error can't find the container with id 6ff7d70627ee9767a5a121f0e878224ff1ebf6f6dbe47d1b22e364deb2261817 Nov 22 08:09:57 crc kubenswrapper[4789]: I1122 08:09:57.805087 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-59p65" event={"ID":"4dc70724-5816-4152-8d2f-763faeddd54f","Type":"ContainerStarted","Data":"e7cb7c4807f19d5008e0c376bda0e32017607b652c67603d36b8231f94abef82"} Nov 22 08:09:57 crc kubenswrapper[4789]: I1122 08:09:57.806313 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-hccc7" event={"ID":"d94d3cb7-9188-42a2-b911-4f7665349830","Type":"ContainerStarted","Data":"bf046adf56a5d3cd1c368227a1fb9376da816effdbc5f25061726bb6b332c3bc"} Nov 22 08:09:57 crc kubenswrapper[4789]: I1122 08:09:57.808565 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b2vc5" event={"ID":"3009aea9-8c5a-42cc-b9d6-dca6ce289616","Type":"ContainerDied","Data":"f37838bb7c96aae8e6ae01824e0400f2f611df8ec3bb695d3276531022eadc10"} Nov 22 08:09:57 crc kubenswrapper[4789]: I1122 08:09:57.808612 4789 scope.go:117] "RemoveContainer" containerID="118ff9b20cd28a73ebbfeaea76cf9386d79de7513149c06e7011cfaeb91de3ba" Nov 22 08:09:57 crc kubenswrapper[4789]: I1122 08:09:57.808830 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b2vc5" Nov 22 08:09:57 crc kubenswrapper[4789]: I1122 08:09:57.810055 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5c449cb8b8-9r629" event={"ID":"62e3cf19-cd3a-438b-978b-0f5f300506dc","Type":"ContainerStarted","Data":"14ed2afd74b837dc7ad740caf0c0ca52466c76914e55dee55668824b2adf85ff"} Nov 22 08:09:57 crc kubenswrapper[4789]: I1122 08:09:57.810096 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5c449cb8b8-9r629" event={"ID":"62e3cf19-cd3a-438b-978b-0f5f300506dc","Type":"ContainerStarted","Data":"5734b406ef9b4c4dbb32f97d639fb49d5e7852186f6c7df808fcb8b1fc2398c6"} Nov 22 08:09:57 crc kubenswrapper[4789]: I1122 08:09:57.814478 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-t6j77" event={"ID":"15f327f3-5cc9-4ce4-b291-efdffea58145","Type":"ContainerStarted","Data":"76cb1ac32664e15f9935bce63e9fcc4fdd216dc38c7657c9f81ecac674fd1e54"} Nov 22 08:09:57 crc kubenswrapper[4789]: I1122 08:09:57.815790 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-pfnkv" event={"ID":"c88315b2-ec94-4272-91f1-3127418ae55c","Type":"ContainerStarted","Data":"6ff7d70627ee9767a5a121f0e878224ff1ebf6f6dbe47d1b22e364deb2261817"} Nov 22 08:09:57 crc kubenswrapper[4789]: I1122 08:09:57.836493 4789 scope.go:117] "RemoveContainer" containerID="8a60b113d1b061d3884e76a8f511e9b3560cba34c88992b11967e321d5646322" Nov 22 08:09:57 crc kubenswrapper[4789]: I1122 08:09:57.847090 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-5c449cb8b8-9r629" podStartSLOduration=1.8470675939999999 podStartE2EDuration="1.847067594s" podCreationTimestamp="2025-11-22 08:09:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:09:57.831241425 +0000 UTC m=+812.065641708" watchObservedRunningTime="2025-11-22 08:09:57.847067594 +0000 UTC m=+812.081467857" Nov 22 08:09:57 crc kubenswrapper[4789]: I1122 08:09:57.864895 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-b2vc5"] Nov 22 08:09:57 crc kubenswrapper[4789]: I1122 08:09:57.868317 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-b2vc5"] Nov 22 08:09:57 crc kubenswrapper[4789]: I1122 08:09:57.877188 4789 scope.go:117] "RemoveContainer" containerID="12dd9f00b45f9af5bc3a2013a1bc72aedca04f0abdca53e2aad3f61ba6317adf" Nov 22 08:09:57 crc kubenswrapper[4789]: I1122 08:09:57.976058 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3009aea9-8c5a-42cc-b9d6-dca6ce289616" path="/var/lib/kubelet/pods/3009aea9-8c5a-42cc-b9d6-dca6ce289616/volumes" Nov 22 08:09:59 crc kubenswrapper[4789]: I1122 08:09:59.332837 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-2sp2r" Nov 22 08:09:59 crc kubenswrapper[4789]: I1122 08:09:59.375808 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-2sp2r" Nov 22 08:10:00 crc kubenswrapper[4789]: I1122 08:10:00.156551 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-2sp2r"] Nov 22 08:10:00 crc kubenswrapper[4789]: I1122 08:10:00.837855 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-59p65" event={"ID":"4dc70724-5816-4152-8d2f-763faeddd54f","Type":"ContainerStarted","Data":"d1b9d291ff595617390e274bfed545d219b0734f73db9e3b5393fa485ec51c63"} Nov 22 08:10:00 crc kubenswrapper[4789]: I1122 08:10:00.838258 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-59p65" Nov 22 08:10:00 crc kubenswrapper[4789]: I1122 08:10:00.840850 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-hccc7" event={"ID":"d94d3cb7-9188-42a2-b911-4f7665349830","Type":"ContainerStarted","Data":"12a8330c3a9f4ccb5140a86e44bc817483cc0c781e8e35fa24045b0bdd950629"} Nov 22 08:10:00 crc kubenswrapper[4789]: I1122 08:10:00.841023 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-hccc7" Nov 22 08:10:00 crc kubenswrapper[4789]: I1122 08:10:00.843677 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-t6j77" event={"ID":"15f327f3-5cc9-4ce4-b291-efdffea58145","Type":"ContainerStarted","Data":"bea5c393c78c1a40f9e8a026797fd552bfa5658d44c45bc79fa643ef61cba4d9"} Nov 22 08:10:00 crc kubenswrapper[4789]: I1122 08:10:00.865133 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-pfnkv" event={"ID":"c88315b2-ec94-4272-91f1-3127418ae55c","Type":"ContainerStarted","Data":"6d02df6f6c8f81ebefcd422dcd665b304e6660e8a3950cb16293b332f3e24b05"} Nov 22 08:10:00 crc kubenswrapper[4789]: I1122 08:10:00.865685 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-59p65" podStartSLOduration=2.093744411 podStartE2EDuration="4.865620494s" podCreationTimestamp="2025-11-22 08:09:56 +0000 UTC" firstStartedPulling="2025-11-22 08:09:56.928696425 +0000 UTC m=+811.163096698" lastFinishedPulling="2025-11-22 08:09:59.700572508 +0000 UTC m=+813.934972781" observedRunningTime="2025-11-22 08:10:00.859315192 +0000 UTC m=+815.093715465" watchObservedRunningTime="2025-11-22 08:10:00.865620494 +0000 UTC m=+815.100020767" Nov 22 08:10:00 crc kubenswrapper[4789]: I1122 08:10:00.866298 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-2sp2r" podUID="5746cb8d-5058-45e7-9f5d-f1e90a666973" containerName="registry-server" containerID="cri-o://0366eac1b416707d6c0bad3298fa20229f0ea18c7a7108c770f31c84752976a8" gracePeriod=2 Nov 22 08:10:00 crc kubenswrapper[4789]: I1122 08:10:00.884965 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-hccc7" podStartSLOduration=1.963714971 podStartE2EDuration="4.884936448s" podCreationTimestamp="2025-11-22 08:09:56 +0000 UTC" firstStartedPulling="2025-11-22 08:09:56.77599867 +0000 UTC m=+811.010398943" lastFinishedPulling="2025-11-22 08:09:59.697220147 +0000 UTC m=+813.931620420" observedRunningTime="2025-11-22 08:10:00.87652564 +0000 UTC m=+815.110925913" watchObservedRunningTime="2025-11-22 08:10:00.884936448 +0000 UTC m=+815.119336731" Nov 22 08:10:00 crc kubenswrapper[4789]: I1122 08:10:00.895885 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-pfnkv" podStartSLOduration=2.725736337 podStartE2EDuration="4.895856205s" podCreationTimestamp="2025-11-22 08:09:56 +0000 UTC" firstStartedPulling="2025-11-22 08:09:57.52530785 +0000 UTC m=+811.759708123" lastFinishedPulling="2025-11-22 08:09:59.695427718 +0000 UTC m=+813.929827991" observedRunningTime="2025-11-22 08:10:00.888603087 +0000 UTC m=+815.123003360" watchObservedRunningTime="2025-11-22 08:10:00.895856205 +0000 UTC m=+815.130256468" Nov 22 08:10:01 crc kubenswrapper[4789]: I1122 08:10:01.234580 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2sp2r" Nov 22 08:10:01 crc kubenswrapper[4789]: I1122 08:10:01.367816 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5746cb8d-5058-45e7-9f5d-f1e90a666973-utilities\") pod \"5746cb8d-5058-45e7-9f5d-f1e90a666973\" (UID: \"5746cb8d-5058-45e7-9f5d-f1e90a666973\") " Nov 22 08:10:01 crc kubenswrapper[4789]: I1122 08:10:01.367885 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-szz4q\" (UniqueName: \"kubernetes.io/projected/5746cb8d-5058-45e7-9f5d-f1e90a666973-kube-api-access-szz4q\") pod \"5746cb8d-5058-45e7-9f5d-f1e90a666973\" (UID: \"5746cb8d-5058-45e7-9f5d-f1e90a666973\") " Nov 22 08:10:01 crc kubenswrapper[4789]: I1122 08:10:01.367916 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5746cb8d-5058-45e7-9f5d-f1e90a666973-catalog-content\") pod \"5746cb8d-5058-45e7-9f5d-f1e90a666973\" (UID: \"5746cb8d-5058-45e7-9f5d-f1e90a666973\") " Nov 22 08:10:01 crc kubenswrapper[4789]: I1122 08:10:01.369081 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5746cb8d-5058-45e7-9f5d-f1e90a666973-utilities" (OuterVolumeSpecName: "utilities") pod "5746cb8d-5058-45e7-9f5d-f1e90a666973" (UID: "5746cb8d-5058-45e7-9f5d-f1e90a666973"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:10:01 crc kubenswrapper[4789]: I1122 08:10:01.375589 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5746cb8d-5058-45e7-9f5d-f1e90a666973-kube-api-access-szz4q" (OuterVolumeSpecName: "kube-api-access-szz4q") pod "5746cb8d-5058-45e7-9f5d-f1e90a666973" (UID: "5746cb8d-5058-45e7-9f5d-f1e90a666973"). InnerVolumeSpecName "kube-api-access-szz4q". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:10:01 crc kubenswrapper[4789]: I1122 08:10:01.451687 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5746cb8d-5058-45e7-9f5d-f1e90a666973-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5746cb8d-5058-45e7-9f5d-f1e90a666973" (UID: "5746cb8d-5058-45e7-9f5d-f1e90a666973"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:10:01 crc kubenswrapper[4789]: I1122 08:10:01.469113 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5746cb8d-5058-45e7-9f5d-f1e90a666973-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 08:10:01 crc kubenswrapper[4789]: I1122 08:10:01.469134 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-szz4q\" (UniqueName: \"kubernetes.io/projected/5746cb8d-5058-45e7-9f5d-f1e90a666973-kube-api-access-szz4q\") on node \"crc\" DevicePath \"\"" Nov 22 08:10:01 crc kubenswrapper[4789]: I1122 08:10:01.469172 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5746cb8d-5058-45e7-9f5d-f1e90a666973-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 08:10:01 crc kubenswrapper[4789]: I1122 08:10:01.874537 4789 generic.go:334] "Generic (PLEG): container finished" podID="5746cb8d-5058-45e7-9f5d-f1e90a666973" containerID="0366eac1b416707d6c0bad3298fa20229f0ea18c7a7108c770f31c84752976a8" exitCode=0 Nov 22 08:10:01 crc kubenswrapper[4789]: I1122 08:10:01.874648 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2sp2r" Nov 22 08:10:01 crc kubenswrapper[4789]: I1122 08:10:01.874650 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2sp2r" event={"ID":"5746cb8d-5058-45e7-9f5d-f1e90a666973","Type":"ContainerDied","Data":"0366eac1b416707d6c0bad3298fa20229f0ea18c7a7108c770f31c84752976a8"} Nov 22 08:10:01 crc kubenswrapper[4789]: I1122 08:10:01.874732 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2sp2r" event={"ID":"5746cb8d-5058-45e7-9f5d-f1e90a666973","Type":"ContainerDied","Data":"e77cabeec9f2e1eaaf50b2c017401fdd2670c80bb3219eb79cf737dc97255fa9"} Nov 22 08:10:01 crc kubenswrapper[4789]: I1122 08:10:01.874828 4789 scope.go:117] "RemoveContainer" containerID="0366eac1b416707d6c0bad3298fa20229f0ea18c7a7108c770f31c84752976a8" Nov 22 08:10:01 crc kubenswrapper[4789]: I1122 08:10:01.886827 4789 scope.go:117] "RemoveContainer" containerID="550ec18fb1967ed632ad463b992a06481679dfe9e29e9a9bb7a1c4a651630667" Nov 22 08:10:01 crc kubenswrapper[4789]: I1122 08:10:01.906650 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-2sp2r"] Nov 22 08:10:01 crc kubenswrapper[4789]: I1122 08:10:01.909306 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-2sp2r"] Nov 22 08:10:01 crc kubenswrapper[4789]: I1122 08:10:01.932931 4789 scope.go:117] "RemoveContainer" containerID="0153cb94cea2a99ed67964df338bc6fa0590653475132e4e8739a97dcf1b835d" Nov 22 08:10:01 crc kubenswrapper[4789]: I1122 08:10:01.964508 4789 scope.go:117] "RemoveContainer" containerID="0366eac1b416707d6c0bad3298fa20229f0ea18c7a7108c770f31c84752976a8" Nov 22 08:10:01 crc kubenswrapper[4789]: E1122 08:10:01.965088 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0366eac1b416707d6c0bad3298fa20229f0ea18c7a7108c770f31c84752976a8\": container with ID starting with 0366eac1b416707d6c0bad3298fa20229f0ea18c7a7108c770f31c84752976a8 not found: ID does not exist" containerID="0366eac1b416707d6c0bad3298fa20229f0ea18c7a7108c770f31c84752976a8" Nov 22 08:10:01 crc kubenswrapper[4789]: I1122 08:10:01.965123 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0366eac1b416707d6c0bad3298fa20229f0ea18c7a7108c770f31c84752976a8"} err="failed to get container status \"0366eac1b416707d6c0bad3298fa20229f0ea18c7a7108c770f31c84752976a8\": rpc error: code = NotFound desc = could not find container \"0366eac1b416707d6c0bad3298fa20229f0ea18c7a7108c770f31c84752976a8\": container with ID starting with 0366eac1b416707d6c0bad3298fa20229f0ea18c7a7108c770f31c84752976a8 not found: ID does not exist" Nov 22 08:10:01 crc kubenswrapper[4789]: I1122 08:10:01.965147 4789 scope.go:117] "RemoveContainer" containerID="550ec18fb1967ed632ad463b992a06481679dfe9e29e9a9bb7a1c4a651630667" Nov 22 08:10:01 crc kubenswrapper[4789]: E1122 08:10:01.965620 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"550ec18fb1967ed632ad463b992a06481679dfe9e29e9a9bb7a1c4a651630667\": container with ID starting with 550ec18fb1967ed632ad463b992a06481679dfe9e29e9a9bb7a1c4a651630667 not found: ID does not exist" containerID="550ec18fb1967ed632ad463b992a06481679dfe9e29e9a9bb7a1c4a651630667" Nov 22 08:10:01 crc kubenswrapper[4789]: I1122 08:10:01.965646 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"550ec18fb1967ed632ad463b992a06481679dfe9e29e9a9bb7a1c4a651630667"} err="failed to get container status \"550ec18fb1967ed632ad463b992a06481679dfe9e29e9a9bb7a1c4a651630667\": rpc error: code = NotFound desc = could not find container \"550ec18fb1967ed632ad463b992a06481679dfe9e29e9a9bb7a1c4a651630667\": container with ID starting with 550ec18fb1967ed632ad463b992a06481679dfe9e29e9a9bb7a1c4a651630667 not found: ID does not exist" Nov 22 08:10:01 crc kubenswrapper[4789]: I1122 08:10:01.965666 4789 scope.go:117] "RemoveContainer" containerID="0153cb94cea2a99ed67964df338bc6fa0590653475132e4e8739a97dcf1b835d" Nov 22 08:10:01 crc kubenswrapper[4789]: E1122 08:10:01.966131 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0153cb94cea2a99ed67964df338bc6fa0590653475132e4e8739a97dcf1b835d\": container with ID starting with 0153cb94cea2a99ed67964df338bc6fa0590653475132e4e8739a97dcf1b835d not found: ID does not exist" containerID="0153cb94cea2a99ed67964df338bc6fa0590653475132e4e8739a97dcf1b835d" Nov 22 08:10:01 crc kubenswrapper[4789]: I1122 08:10:01.966167 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0153cb94cea2a99ed67964df338bc6fa0590653475132e4e8739a97dcf1b835d"} err="failed to get container status \"0153cb94cea2a99ed67964df338bc6fa0590653475132e4e8739a97dcf1b835d\": rpc error: code = NotFound desc = could not find container \"0153cb94cea2a99ed67964df338bc6fa0590653475132e4e8739a97dcf1b835d\": container with ID starting with 0153cb94cea2a99ed67964df338bc6fa0590653475132e4e8739a97dcf1b835d not found: ID does not exist" Nov 22 08:10:01 crc kubenswrapper[4789]: I1122 08:10:01.971802 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5746cb8d-5058-45e7-9f5d-f1e90a666973" path="/var/lib/kubelet/pods/5746cb8d-5058-45e7-9f5d-f1e90a666973/volumes" Nov 22 08:10:02 crc kubenswrapper[4789]: I1122 08:10:02.881325 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-t6j77" event={"ID":"15f327f3-5cc9-4ce4-b291-efdffea58145","Type":"ContainerStarted","Data":"0211087de771db27cdc561a1cc66363aa9337e05621e69a6444e47cac5b78222"} Nov 22 08:10:02 crc kubenswrapper[4789]: I1122 08:10:02.895357 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-t6j77" podStartSLOduration=2.079389981 podStartE2EDuration="6.895341741s" podCreationTimestamp="2025-11-22 08:09:56 +0000 UTC" firstStartedPulling="2025-11-22 08:09:57.074557524 +0000 UTC m=+811.308957797" lastFinishedPulling="2025-11-22 08:10:01.890509284 +0000 UTC m=+816.124909557" observedRunningTime="2025-11-22 08:10:02.894691333 +0000 UTC m=+817.129091616" watchObservedRunningTime="2025-11-22 08:10:02.895341741 +0000 UTC m=+817.129742014" Nov 22 08:10:05 crc kubenswrapper[4789]: I1122 08:10:05.372575 4789 patch_prober.go:28] interesting pod/machine-config-daemon-pxvrm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 08:10:05 crc kubenswrapper[4789]: I1122 08:10:05.372725 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 08:10:05 crc kubenswrapper[4789]: I1122 08:10:05.372824 4789 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" Nov 22 08:10:05 crc kubenswrapper[4789]: I1122 08:10:05.373739 4789 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"447a15a0714201a673fc1d0cc8a8f355942fbc82479335216d2bf7c26b973475"} pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 22 08:10:05 crc kubenswrapper[4789]: I1122 08:10:05.373834 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" containerID="cri-o://447a15a0714201a673fc1d0cc8a8f355942fbc82479335216d2bf7c26b973475" gracePeriod=600 Nov 22 08:10:05 crc kubenswrapper[4789]: I1122 08:10:05.905357 4789 generic.go:334] "Generic (PLEG): container finished" podID="58aa8071-7ecc-4692-8789-2db04ec70647" containerID="447a15a0714201a673fc1d0cc8a8f355942fbc82479335216d2bf7c26b973475" exitCode=0 Nov 22 08:10:05 crc kubenswrapper[4789]: I1122 08:10:05.905613 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" event={"ID":"58aa8071-7ecc-4692-8789-2db04ec70647","Type":"ContainerDied","Data":"447a15a0714201a673fc1d0cc8a8f355942fbc82479335216d2bf7c26b973475"} Nov 22 08:10:05 crc kubenswrapper[4789]: I1122 08:10:05.906353 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" event={"ID":"58aa8071-7ecc-4692-8789-2db04ec70647","Type":"ContainerStarted","Data":"5067bbca7c5576cc068f8534a87953cf9b5fc5cb8f8cfddaaed99f53192975c1"} Nov 22 08:10:05 crc kubenswrapper[4789]: I1122 08:10:05.906393 4789 scope.go:117] "RemoveContainer" containerID="5afa28473bdb65dc9e867328a0ad6d3bb782e8cf253ef73d039405b797171f7a" Nov 22 08:10:06 crc kubenswrapper[4789]: I1122 08:10:06.675444 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-hccc7" Nov 22 08:10:07 crc kubenswrapper[4789]: I1122 08:10:07.078029 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-5c449cb8b8-9r629" Nov 22 08:10:07 crc kubenswrapper[4789]: I1122 08:10:07.078466 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-5c449cb8b8-9r629" Nov 22 08:10:07 crc kubenswrapper[4789]: I1122 08:10:07.085290 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-5c449cb8b8-9r629" Nov 22 08:10:07 crc kubenswrapper[4789]: I1122 08:10:07.927977 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-5c449cb8b8-9r629" Nov 22 08:10:07 crc kubenswrapper[4789]: I1122 08:10:07.988352 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-pb98x"] Nov 22 08:10:12 crc kubenswrapper[4789]: I1122 08:10:12.973422 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-nx6gk"] Nov 22 08:10:12 crc kubenswrapper[4789]: E1122 08:10:12.974265 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3009aea9-8c5a-42cc-b9d6-dca6ce289616" containerName="extract-utilities" Nov 22 08:10:12 crc kubenswrapper[4789]: I1122 08:10:12.974281 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="3009aea9-8c5a-42cc-b9d6-dca6ce289616" containerName="extract-utilities" Nov 22 08:10:12 crc kubenswrapper[4789]: E1122 08:10:12.974293 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3009aea9-8c5a-42cc-b9d6-dca6ce289616" containerName="extract-content" Nov 22 08:10:12 crc kubenswrapper[4789]: I1122 08:10:12.974300 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="3009aea9-8c5a-42cc-b9d6-dca6ce289616" containerName="extract-content" Nov 22 08:10:12 crc kubenswrapper[4789]: E1122 08:10:12.974313 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3009aea9-8c5a-42cc-b9d6-dca6ce289616" containerName="registry-server" Nov 22 08:10:12 crc kubenswrapper[4789]: I1122 08:10:12.974321 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="3009aea9-8c5a-42cc-b9d6-dca6ce289616" containerName="registry-server" Nov 22 08:10:12 crc kubenswrapper[4789]: E1122 08:10:12.974330 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5746cb8d-5058-45e7-9f5d-f1e90a666973" containerName="registry-server" Nov 22 08:10:12 crc kubenswrapper[4789]: I1122 08:10:12.974339 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="5746cb8d-5058-45e7-9f5d-f1e90a666973" containerName="registry-server" Nov 22 08:10:12 crc kubenswrapper[4789]: E1122 08:10:12.974350 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5746cb8d-5058-45e7-9f5d-f1e90a666973" containerName="extract-utilities" Nov 22 08:10:12 crc kubenswrapper[4789]: I1122 08:10:12.974357 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="5746cb8d-5058-45e7-9f5d-f1e90a666973" containerName="extract-utilities" Nov 22 08:10:12 crc kubenswrapper[4789]: E1122 08:10:12.974367 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5746cb8d-5058-45e7-9f5d-f1e90a666973" containerName="extract-content" Nov 22 08:10:12 crc kubenswrapper[4789]: I1122 08:10:12.974374 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="5746cb8d-5058-45e7-9f5d-f1e90a666973" containerName="extract-content" Nov 22 08:10:12 crc kubenswrapper[4789]: I1122 08:10:12.974515 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="5746cb8d-5058-45e7-9f5d-f1e90a666973" containerName="registry-server" Nov 22 08:10:12 crc kubenswrapper[4789]: I1122 08:10:12.974533 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="3009aea9-8c5a-42cc-b9d6-dca6ce289616" containerName="registry-server" Nov 22 08:10:12 crc kubenswrapper[4789]: I1122 08:10:12.975516 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nx6gk" Nov 22 08:10:12 crc kubenswrapper[4789]: I1122 08:10:12.999702 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nx6gk"] Nov 22 08:10:13 crc kubenswrapper[4789]: I1122 08:10:13.123796 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/089857ed-bf51-4105-8ddc-755c4c792a42-catalog-content\") pod \"certified-operators-nx6gk\" (UID: \"089857ed-bf51-4105-8ddc-755c4c792a42\") " pod="openshift-marketplace/certified-operators-nx6gk" Nov 22 08:10:13 crc kubenswrapper[4789]: I1122 08:10:13.124034 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/089857ed-bf51-4105-8ddc-755c4c792a42-utilities\") pod \"certified-operators-nx6gk\" (UID: \"089857ed-bf51-4105-8ddc-755c4c792a42\") " pod="openshift-marketplace/certified-operators-nx6gk" Nov 22 08:10:13 crc kubenswrapper[4789]: I1122 08:10:13.124080 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sp68f\" (UniqueName: \"kubernetes.io/projected/089857ed-bf51-4105-8ddc-755c4c792a42-kube-api-access-sp68f\") pod \"certified-operators-nx6gk\" (UID: \"089857ed-bf51-4105-8ddc-755c4c792a42\") " pod="openshift-marketplace/certified-operators-nx6gk" Nov 22 08:10:13 crc kubenswrapper[4789]: I1122 08:10:13.225356 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sp68f\" (UniqueName: \"kubernetes.io/projected/089857ed-bf51-4105-8ddc-755c4c792a42-kube-api-access-sp68f\") pod \"certified-operators-nx6gk\" (UID: \"089857ed-bf51-4105-8ddc-755c4c792a42\") " pod="openshift-marketplace/certified-operators-nx6gk" Nov 22 08:10:13 crc kubenswrapper[4789]: I1122 08:10:13.225721 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/089857ed-bf51-4105-8ddc-755c4c792a42-catalog-content\") pod \"certified-operators-nx6gk\" (UID: \"089857ed-bf51-4105-8ddc-755c4c792a42\") " pod="openshift-marketplace/certified-operators-nx6gk" Nov 22 08:10:13 crc kubenswrapper[4789]: I1122 08:10:13.225823 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/089857ed-bf51-4105-8ddc-755c4c792a42-utilities\") pod \"certified-operators-nx6gk\" (UID: \"089857ed-bf51-4105-8ddc-755c4c792a42\") " pod="openshift-marketplace/certified-operators-nx6gk" Nov 22 08:10:13 crc kubenswrapper[4789]: I1122 08:10:13.226260 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/089857ed-bf51-4105-8ddc-755c4c792a42-catalog-content\") pod \"certified-operators-nx6gk\" (UID: \"089857ed-bf51-4105-8ddc-755c4c792a42\") " pod="openshift-marketplace/certified-operators-nx6gk" Nov 22 08:10:13 crc kubenswrapper[4789]: I1122 08:10:13.226299 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/089857ed-bf51-4105-8ddc-755c4c792a42-utilities\") pod \"certified-operators-nx6gk\" (UID: \"089857ed-bf51-4105-8ddc-755c4c792a42\") " pod="openshift-marketplace/certified-operators-nx6gk" Nov 22 08:10:13 crc kubenswrapper[4789]: I1122 08:10:13.260030 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sp68f\" (UniqueName: \"kubernetes.io/projected/089857ed-bf51-4105-8ddc-755c4c792a42-kube-api-access-sp68f\") pod \"certified-operators-nx6gk\" (UID: \"089857ed-bf51-4105-8ddc-755c4c792a42\") " pod="openshift-marketplace/certified-operators-nx6gk" Nov 22 08:10:13 crc kubenswrapper[4789]: I1122 08:10:13.305093 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nx6gk" Nov 22 08:10:13 crc kubenswrapper[4789]: I1122 08:10:13.544132 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nx6gk"] Nov 22 08:10:13 crc kubenswrapper[4789]: I1122 08:10:13.965621 4789 generic.go:334] "Generic (PLEG): container finished" podID="089857ed-bf51-4105-8ddc-755c4c792a42" containerID="24107c82517a7b9a7c952d8eb0862869cecc46a985b6342ec5d410aca70df291" exitCode=0 Nov 22 08:10:13 crc kubenswrapper[4789]: I1122 08:10:13.973600 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nx6gk" event={"ID":"089857ed-bf51-4105-8ddc-755c4c792a42","Type":"ContainerDied","Data":"24107c82517a7b9a7c952d8eb0862869cecc46a985b6342ec5d410aca70df291"} Nov 22 08:10:13 crc kubenswrapper[4789]: I1122 08:10:13.973642 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nx6gk" event={"ID":"089857ed-bf51-4105-8ddc-755c4c792a42","Type":"ContainerStarted","Data":"eed8d5db296c4a327ba9c6c49292d6fcfff8827b2578096c155e7e934c5182cd"} Nov 22 08:10:14 crc kubenswrapper[4789]: I1122 08:10:14.973989 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nx6gk" event={"ID":"089857ed-bf51-4105-8ddc-755c4c792a42","Type":"ContainerStarted","Data":"abb39bcbb061cd1cce314a35c7877d033c22c81c2f3fec456fe1f875db453039"} Nov 22 08:10:15 crc kubenswrapper[4789]: I1122 08:10:15.982923 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nx6gk" event={"ID":"089857ed-bf51-4105-8ddc-755c4c792a42","Type":"ContainerDied","Data":"abb39bcbb061cd1cce314a35c7877d033c22c81c2f3fec456fe1f875db453039"} Nov 22 08:10:15 crc kubenswrapper[4789]: I1122 08:10:15.982607 4789 generic.go:334] "Generic (PLEG): container finished" podID="089857ed-bf51-4105-8ddc-755c4c792a42" containerID="abb39bcbb061cd1cce314a35c7877d033c22c81c2f3fec456fe1f875db453039" exitCode=0 Nov 22 08:10:16 crc kubenswrapper[4789]: I1122 08:10:16.651524 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-59p65" Nov 22 08:10:16 crc kubenswrapper[4789]: I1122 08:10:16.992543 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nx6gk" event={"ID":"089857ed-bf51-4105-8ddc-755c4c792a42","Type":"ContainerStarted","Data":"167c59852a183009f5d17b49c022fde2316310b2523cf738d1c10ee98b554ef3"} Nov 22 08:10:17 crc kubenswrapper[4789]: I1122 08:10:17.015217 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-nx6gk" podStartSLOduration=2.407389006 podStartE2EDuration="5.015201405s" podCreationTimestamp="2025-11-22 08:10:12 +0000 UTC" firstStartedPulling="2025-11-22 08:10:13.967682981 +0000 UTC m=+828.202083254" lastFinishedPulling="2025-11-22 08:10:16.57549538 +0000 UTC m=+830.809895653" observedRunningTime="2025-11-22 08:10:17.01313584 +0000 UTC m=+831.247536113" watchObservedRunningTime="2025-11-22 08:10:17.015201405 +0000 UTC m=+831.249601678" Nov 22 08:10:23 crc kubenswrapper[4789]: I1122 08:10:23.314328 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-nx6gk" Nov 22 08:10:23 crc kubenswrapper[4789]: I1122 08:10:23.314993 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-nx6gk" Nov 22 08:10:23 crc kubenswrapper[4789]: I1122 08:10:23.354105 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-nx6gk" Nov 22 08:10:24 crc kubenswrapper[4789]: I1122 08:10:24.071838 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-nx6gk" Nov 22 08:10:24 crc kubenswrapper[4789]: I1122 08:10:24.123789 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nx6gk"] Nov 22 08:10:26 crc kubenswrapper[4789]: I1122 08:10:26.045282 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-nx6gk" podUID="089857ed-bf51-4105-8ddc-755c4c792a42" containerName="registry-server" containerID="cri-o://167c59852a183009f5d17b49c022fde2316310b2523cf738d1c10ee98b554ef3" gracePeriod=2 Nov 22 08:10:26 crc kubenswrapper[4789]: I1122 08:10:26.899834 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nx6gk" Nov 22 08:10:26 crc kubenswrapper[4789]: I1122 08:10:26.903726 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/089857ed-bf51-4105-8ddc-755c4c792a42-catalog-content\") pod \"089857ed-bf51-4105-8ddc-755c4c792a42\" (UID: \"089857ed-bf51-4105-8ddc-755c4c792a42\") " Nov 22 08:10:26 crc kubenswrapper[4789]: I1122 08:10:26.903783 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sp68f\" (UniqueName: \"kubernetes.io/projected/089857ed-bf51-4105-8ddc-755c4c792a42-kube-api-access-sp68f\") pod \"089857ed-bf51-4105-8ddc-755c4c792a42\" (UID: \"089857ed-bf51-4105-8ddc-755c4c792a42\") " Nov 22 08:10:26 crc kubenswrapper[4789]: I1122 08:10:26.911722 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/089857ed-bf51-4105-8ddc-755c4c792a42-kube-api-access-sp68f" (OuterVolumeSpecName: "kube-api-access-sp68f") pod "089857ed-bf51-4105-8ddc-755c4c792a42" (UID: "089857ed-bf51-4105-8ddc-755c4c792a42"). InnerVolumeSpecName "kube-api-access-sp68f". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:10:26 crc kubenswrapper[4789]: I1122 08:10:26.958467 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/089857ed-bf51-4105-8ddc-755c4c792a42-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "089857ed-bf51-4105-8ddc-755c4c792a42" (UID: "089857ed-bf51-4105-8ddc-755c4c792a42"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:10:27 crc kubenswrapper[4789]: I1122 08:10:27.004831 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/089857ed-bf51-4105-8ddc-755c4c792a42-utilities\") pod \"089857ed-bf51-4105-8ddc-755c4c792a42\" (UID: \"089857ed-bf51-4105-8ddc-755c4c792a42\") " Nov 22 08:10:27 crc kubenswrapper[4789]: I1122 08:10:27.005454 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/089857ed-bf51-4105-8ddc-755c4c792a42-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 08:10:27 crc kubenswrapper[4789]: I1122 08:10:27.005483 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sp68f\" (UniqueName: \"kubernetes.io/projected/089857ed-bf51-4105-8ddc-755c4c792a42-kube-api-access-sp68f\") on node \"crc\" DevicePath \"\"" Nov 22 08:10:27 crc kubenswrapper[4789]: I1122 08:10:27.005639 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/089857ed-bf51-4105-8ddc-755c4c792a42-utilities" (OuterVolumeSpecName: "utilities") pod "089857ed-bf51-4105-8ddc-755c4c792a42" (UID: "089857ed-bf51-4105-8ddc-755c4c792a42"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:10:27 crc kubenswrapper[4789]: I1122 08:10:27.051676 4789 generic.go:334] "Generic (PLEG): container finished" podID="089857ed-bf51-4105-8ddc-755c4c792a42" containerID="167c59852a183009f5d17b49c022fde2316310b2523cf738d1c10ee98b554ef3" exitCode=0 Nov 22 08:10:27 crc kubenswrapper[4789]: I1122 08:10:27.051736 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nx6gk" Nov 22 08:10:27 crc kubenswrapper[4789]: I1122 08:10:27.051736 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nx6gk" event={"ID":"089857ed-bf51-4105-8ddc-755c4c792a42","Type":"ContainerDied","Data":"167c59852a183009f5d17b49c022fde2316310b2523cf738d1c10ee98b554ef3"} Nov 22 08:10:27 crc kubenswrapper[4789]: I1122 08:10:27.052885 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nx6gk" event={"ID":"089857ed-bf51-4105-8ddc-755c4c792a42","Type":"ContainerDied","Data":"eed8d5db296c4a327ba9c6c49292d6fcfff8827b2578096c155e7e934c5182cd"} Nov 22 08:10:27 crc kubenswrapper[4789]: I1122 08:10:27.052906 4789 scope.go:117] "RemoveContainer" containerID="167c59852a183009f5d17b49c022fde2316310b2523cf738d1c10ee98b554ef3" Nov 22 08:10:27 crc kubenswrapper[4789]: I1122 08:10:27.081494 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nx6gk"] Nov 22 08:10:27 crc kubenswrapper[4789]: I1122 08:10:27.081953 4789 scope.go:117] "RemoveContainer" containerID="abb39bcbb061cd1cce314a35c7877d033c22c81c2f3fec456fe1f875db453039" Nov 22 08:10:27 crc kubenswrapper[4789]: I1122 08:10:27.090343 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-nx6gk"] Nov 22 08:10:27 crc kubenswrapper[4789]: I1122 08:10:27.105364 4789 scope.go:117] "RemoveContainer" containerID="24107c82517a7b9a7c952d8eb0862869cecc46a985b6342ec5d410aca70df291" Nov 22 08:10:27 crc kubenswrapper[4789]: I1122 08:10:27.106052 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/089857ed-bf51-4105-8ddc-755c4c792a42-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 08:10:27 crc kubenswrapper[4789]: I1122 08:10:27.117797 4789 scope.go:117] "RemoveContainer" containerID="167c59852a183009f5d17b49c022fde2316310b2523cf738d1c10ee98b554ef3" Nov 22 08:10:27 crc kubenswrapper[4789]: E1122 08:10:27.118195 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"167c59852a183009f5d17b49c022fde2316310b2523cf738d1c10ee98b554ef3\": container with ID starting with 167c59852a183009f5d17b49c022fde2316310b2523cf738d1c10ee98b554ef3 not found: ID does not exist" containerID="167c59852a183009f5d17b49c022fde2316310b2523cf738d1c10ee98b554ef3" Nov 22 08:10:27 crc kubenswrapper[4789]: I1122 08:10:27.118241 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"167c59852a183009f5d17b49c022fde2316310b2523cf738d1c10ee98b554ef3"} err="failed to get container status \"167c59852a183009f5d17b49c022fde2316310b2523cf738d1c10ee98b554ef3\": rpc error: code = NotFound desc = could not find container \"167c59852a183009f5d17b49c022fde2316310b2523cf738d1c10ee98b554ef3\": container with ID starting with 167c59852a183009f5d17b49c022fde2316310b2523cf738d1c10ee98b554ef3 not found: ID does not exist" Nov 22 08:10:27 crc kubenswrapper[4789]: I1122 08:10:27.118281 4789 scope.go:117] "RemoveContainer" containerID="abb39bcbb061cd1cce314a35c7877d033c22c81c2f3fec456fe1f875db453039" Nov 22 08:10:27 crc kubenswrapper[4789]: E1122 08:10:27.118661 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"abb39bcbb061cd1cce314a35c7877d033c22c81c2f3fec456fe1f875db453039\": container with ID starting with abb39bcbb061cd1cce314a35c7877d033c22c81c2f3fec456fe1f875db453039 not found: ID does not exist" containerID="abb39bcbb061cd1cce314a35c7877d033c22c81c2f3fec456fe1f875db453039" Nov 22 08:10:27 crc kubenswrapper[4789]: I1122 08:10:27.118699 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"abb39bcbb061cd1cce314a35c7877d033c22c81c2f3fec456fe1f875db453039"} err="failed to get container status \"abb39bcbb061cd1cce314a35c7877d033c22c81c2f3fec456fe1f875db453039\": rpc error: code = NotFound desc = could not find container \"abb39bcbb061cd1cce314a35c7877d033c22c81c2f3fec456fe1f875db453039\": container with ID starting with abb39bcbb061cd1cce314a35c7877d033c22c81c2f3fec456fe1f875db453039 not found: ID does not exist" Nov 22 08:10:27 crc kubenswrapper[4789]: I1122 08:10:27.118727 4789 scope.go:117] "RemoveContainer" containerID="24107c82517a7b9a7c952d8eb0862869cecc46a985b6342ec5d410aca70df291" Nov 22 08:10:27 crc kubenswrapper[4789]: E1122 08:10:27.119100 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"24107c82517a7b9a7c952d8eb0862869cecc46a985b6342ec5d410aca70df291\": container with ID starting with 24107c82517a7b9a7c952d8eb0862869cecc46a985b6342ec5d410aca70df291 not found: ID does not exist" containerID="24107c82517a7b9a7c952d8eb0862869cecc46a985b6342ec5d410aca70df291" Nov 22 08:10:27 crc kubenswrapper[4789]: I1122 08:10:27.119125 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"24107c82517a7b9a7c952d8eb0862869cecc46a985b6342ec5d410aca70df291"} err="failed to get container status \"24107c82517a7b9a7c952d8eb0862869cecc46a985b6342ec5d410aca70df291\": rpc error: code = NotFound desc = could not find container \"24107c82517a7b9a7c952d8eb0862869cecc46a985b6342ec5d410aca70df291\": container with ID starting with 24107c82517a7b9a7c952d8eb0862869cecc46a985b6342ec5d410aca70df291 not found: ID does not exist" Nov 22 08:10:27 crc kubenswrapper[4789]: I1122 08:10:27.972656 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="089857ed-bf51-4105-8ddc-755c4c792a42" path="/var/lib/kubelet/pods/089857ed-bf51-4105-8ddc-755c4c792a42/volumes" Nov 22 08:10:29 crc kubenswrapper[4789]: I1122 08:10:29.994991 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6vdrbs"] Nov 22 08:10:29 crc kubenswrapper[4789]: E1122 08:10:29.995737 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="089857ed-bf51-4105-8ddc-755c4c792a42" containerName="extract-content" Nov 22 08:10:29 crc kubenswrapper[4789]: I1122 08:10:29.995772 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="089857ed-bf51-4105-8ddc-755c4c792a42" containerName="extract-content" Nov 22 08:10:29 crc kubenswrapper[4789]: E1122 08:10:29.995784 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="089857ed-bf51-4105-8ddc-755c4c792a42" containerName="registry-server" Nov 22 08:10:29 crc kubenswrapper[4789]: I1122 08:10:29.995792 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="089857ed-bf51-4105-8ddc-755c4c792a42" containerName="registry-server" Nov 22 08:10:29 crc kubenswrapper[4789]: E1122 08:10:29.995805 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="089857ed-bf51-4105-8ddc-755c4c792a42" containerName="extract-utilities" Nov 22 08:10:29 crc kubenswrapper[4789]: I1122 08:10:29.995813 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="089857ed-bf51-4105-8ddc-755c4c792a42" containerName="extract-utilities" Nov 22 08:10:29 crc kubenswrapper[4789]: I1122 08:10:29.995943 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="089857ed-bf51-4105-8ddc-755c4c792a42" containerName="registry-server" Nov 22 08:10:29 crc kubenswrapper[4789]: I1122 08:10:29.997005 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6vdrbs" Nov 22 08:10:30 crc kubenswrapper[4789]: I1122 08:10:30.000406 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 22 08:10:30 crc kubenswrapper[4789]: I1122 08:10:30.015836 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6vdrbs"] Nov 22 08:10:30 crc kubenswrapper[4789]: I1122 08:10:30.051409 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8f9ce070-1816-46d5-a16f-b3b8d1a1f8f3-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6vdrbs\" (UID: \"8f9ce070-1816-46d5-a16f-b3b8d1a1f8f3\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6vdrbs" Nov 22 08:10:30 crc kubenswrapper[4789]: I1122 08:10:30.051504 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f7nj4\" (UniqueName: \"kubernetes.io/projected/8f9ce070-1816-46d5-a16f-b3b8d1a1f8f3-kube-api-access-f7nj4\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6vdrbs\" (UID: \"8f9ce070-1816-46d5-a16f-b3b8d1a1f8f3\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6vdrbs" Nov 22 08:10:30 crc kubenswrapper[4789]: I1122 08:10:30.051554 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8f9ce070-1816-46d5-a16f-b3b8d1a1f8f3-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6vdrbs\" (UID: \"8f9ce070-1816-46d5-a16f-b3b8d1a1f8f3\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6vdrbs" Nov 22 08:10:30 crc kubenswrapper[4789]: I1122 08:10:30.153622 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f7nj4\" (UniqueName: \"kubernetes.io/projected/8f9ce070-1816-46d5-a16f-b3b8d1a1f8f3-kube-api-access-f7nj4\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6vdrbs\" (UID: \"8f9ce070-1816-46d5-a16f-b3b8d1a1f8f3\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6vdrbs" Nov 22 08:10:30 crc kubenswrapper[4789]: I1122 08:10:30.153697 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8f9ce070-1816-46d5-a16f-b3b8d1a1f8f3-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6vdrbs\" (UID: \"8f9ce070-1816-46d5-a16f-b3b8d1a1f8f3\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6vdrbs" Nov 22 08:10:30 crc kubenswrapper[4789]: I1122 08:10:30.153786 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8f9ce070-1816-46d5-a16f-b3b8d1a1f8f3-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6vdrbs\" (UID: \"8f9ce070-1816-46d5-a16f-b3b8d1a1f8f3\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6vdrbs" Nov 22 08:10:30 crc kubenswrapper[4789]: I1122 08:10:30.154176 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8f9ce070-1816-46d5-a16f-b3b8d1a1f8f3-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6vdrbs\" (UID: \"8f9ce070-1816-46d5-a16f-b3b8d1a1f8f3\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6vdrbs" Nov 22 08:10:30 crc kubenswrapper[4789]: I1122 08:10:30.154271 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8f9ce070-1816-46d5-a16f-b3b8d1a1f8f3-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6vdrbs\" (UID: \"8f9ce070-1816-46d5-a16f-b3b8d1a1f8f3\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6vdrbs" Nov 22 08:10:30 crc kubenswrapper[4789]: I1122 08:10:30.176884 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f7nj4\" (UniqueName: \"kubernetes.io/projected/8f9ce070-1816-46d5-a16f-b3b8d1a1f8f3-kube-api-access-f7nj4\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6vdrbs\" (UID: \"8f9ce070-1816-46d5-a16f-b3b8d1a1f8f3\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6vdrbs" Nov 22 08:10:30 crc kubenswrapper[4789]: I1122 08:10:30.310873 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6vdrbs" Nov 22 08:10:30 crc kubenswrapper[4789]: I1122 08:10:30.477294 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6vdrbs"] Nov 22 08:10:30 crc kubenswrapper[4789]: W1122 08:10:30.486251 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8f9ce070_1816_46d5_a16f_b3b8d1a1f8f3.slice/crio-377ba8bd1fcd1b7841d0cedea5e8150ae20f17dd3867231366c0d5d912815ed1 WatchSource:0}: Error finding container 377ba8bd1fcd1b7841d0cedea5e8150ae20f17dd3867231366c0d5d912815ed1: Status 404 returned error can't find the container with id 377ba8bd1fcd1b7841d0cedea5e8150ae20f17dd3867231366c0d5d912815ed1 Nov 22 08:10:31 crc kubenswrapper[4789]: I1122 08:10:31.089235 4789 generic.go:334] "Generic (PLEG): container finished" podID="8f9ce070-1816-46d5-a16f-b3b8d1a1f8f3" containerID="1b04d4755e03024a586a8e4ad3552e1d3de7b5d4439d90f45d88747e6842ffac" exitCode=0 Nov 22 08:10:31 crc kubenswrapper[4789]: I1122 08:10:31.089487 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6vdrbs" event={"ID":"8f9ce070-1816-46d5-a16f-b3b8d1a1f8f3","Type":"ContainerDied","Data":"1b04d4755e03024a586a8e4ad3552e1d3de7b5d4439d90f45d88747e6842ffac"} Nov 22 08:10:31 crc kubenswrapper[4789]: I1122 08:10:31.091902 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6vdrbs" event={"ID":"8f9ce070-1816-46d5-a16f-b3b8d1a1f8f3","Type":"ContainerStarted","Data":"377ba8bd1fcd1b7841d0cedea5e8150ae20f17dd3867231366c0d5d912815ed1"} Nov 22 08:10:33 crc kubenswrapper[4789]: I1122 08:10:33.047077 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-pb98x" podUID="9b861750-2517-4cf3-9c78-930480711af0" containerName="console" containerID="cri-o://ddcb04daed5507caf65f440b3db89e0e66a92775018aca69eac59b75a82ace6f" gracePeriod=15 Nov 22 08:10:33 crc kubenswrapper[4789]: I1122 08:10:33.103085 4789 generic.go:334] "Generic (PLEG): container finished" podID="8f9ce070-1816-46d5-a16f-b3b8d1a1f8f3" containerID="34113e686df15e7ce46e2b0bca69a398a38db6d462550d890550c6c6fe143a42" exitCode=0 Nov 22 08:10:33 crc kubenswrapper[4789]: I1122 08:10:33.103159 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6vdrbs" event={"ID":"8f9ce070-1816-46d5-a16f-b3b8d1a1f8f3","Type":"ContainerDied","Data":"34113e686df15e7ce46e2b0bca69a398a38db6d462550d890550c6c6fe143a42"} Nov 22 08:10:33 crc kubenswrapper[4789]: I1122 08:10:33.451312 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-pb98x_9b861750-2517-4cf3-9c78-930480711af0/console/0.log" Nov 22 08:10:33 crc kubenswrapper[4789]: I1122 08:10:33.451653 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-pb98x" Nov 22 08:10:33 crc kubenswrapper[4789]: I1122 08:10:33.622972 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/9b861750-2517-4cf3-9c78-930480711af0-console-oauth-config\") pod \"9b861750-2517-4cf3-9c78-930480711af0\" (UID: \"9b861750-2517-4cf3-9c78-930480711af0\") " Nov 22 08:10:33 crc kubenswrapper[4789]: I1122 08:10:33.623058 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9b861750-2517-4cf3-9c78-930480711af0-service-ca\") pod \"9b861750-2517-4cf3-9c78-930480711af0\" (UID: \"9b861750-2517-4cf3-9c78-930480711af0\") " Nov 22 08:10:33 crc kubenswrapper[4789]: I1122 08:10:33.623102 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9b861750-2517-4cf3-9c78-930480711af0-trusted-ca-bundle\") pod \"9b861750-2517-4cf3-9c78-930480711af0\" (UID: \"9b861750-2517-4cf3-9c78-930480711af0\") " Nov 22 08:10:33 crc kubenswrapper[4789]: I1122 08:10:33.623157 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xh6c5\" (UniqueName: \"kubernetes.io/projected/9b861750-2517-4cf3-9c78-930480711af0-kube-api-access-xh6c5\") pod \"9b861750-2517-4cf3-9c78-930480711af0\" (UID: \"9b861750-2517-4cf3-9c78-930480711af0\") " Nov 22 08:10:33 crc kubenswrapper[4789]: I1122 08:10:33.623196 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/9b861750-2517-4cf3-9c78-930480711af0-console-config\") pod \"9b861750-2517-4cf3-9c78-930480711af0\" (UID: \"9b861750-2517-4cf3-9c78-930480711af0\") " Nov 22 08:10:33 crc kubenswrapper[4789]: I1122 08:10:33.623227 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/9b861750-2517-4cf3-9c78-930480711af0-console-serving-cert\") pod \"9b861750-2517-4cf3-9c78-930480711af0\" (UID: \"9b861750-2517-4cf3-9c78-930480711af0\") " Nov 22 08:10:33 crc kubenswrapper[4789]: I1122 08:10:33.623255 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/9b861750-2517-4cf3-9c78-930480711af0-oauth-serving-cert\") pod \"9b861750-2517-4cf3-9c78-930480711af0\" (UID: \"9b861750-2517-4cf3-9c78-930480711af0\") " Nov 22 08:10:33 crc kubenswrapper[4789]: I1122 08:10:33.624426 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9b861750-2517-4cf3-9c78-930480711af0-service-ca" (OuterVolumeSpecName: "service-ca") pod "9b861750-2517-4cf3-9c78-930480711af0" (UID: "9b861750-2517-4cf3-9c78-930480711af0"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:10:33 crc kubenswrapper[4789]: I1122 08:10:33.624445 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9b861750-2517-4cf3-9c78-930480711af0-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "9b861750-2517-4cf3-9c78-930480711af0" (UID: "9b861750-2517-4cf3-9c78-930480711af0"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:10:33 crc kubenswrapper[4789]: I1122 08:10:33.624477 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9b861750-2517-4cf3-9c78-930480711af0-console-config" (OuterVolumeSpecName: "console-config") pod "9b861750-2517-4cf3-9c78-930480711af0" (UID: "9b861750-2517-4cf3-9c78-930480711af0"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:10:33 crc kubenswrapper[4789]: I1122 08:10:33.625522 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9b861750-2517-4cf3-9c78-930480711af0-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "9b861750-2517-4cf3-9c78-930480711af0" (UID: "9b861750-2517-4cf3-9c78-930480711af0"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:10:33 crc kubenswrapper[4789]: I1122 08:10:33.630592 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9b861750-2517-4cf3-9c78-930480711af0-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "9b861750-2517-4cf3-9c78-930480711af0" (UID: "9b861750-2517-4cf3-9c78-930480711af0"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:10:33 crc kubenswrapper[4789]: I1122 08:10:33.630664 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9b861750-2517-4cf3-9c78-930480711af0-kube-api-access-xh6c5" (OuterVolumeSpecName: "kube-api-access-xh6c5") pod "9b861750-2517-4cf3-9c78-930480711af0" (UID: "9b861750-2517-4cf3-9c78-930480711af0"). InnerVolumeSpecName "kube-api-access-xh6c5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:10:33 crc kubenswrapper[4789]: I1122 08:10:33.631611 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9b861750-2517-4cf3-9c78-930480711af0-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "9b861750-2517-4cf3-9c78-930480711af0" (UID: "9b861750-2517-4cf3-9c78-930480711af0"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:10:33 crc kubenswrapper[4789]: I1122 08:10:33.724869 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xh6c5\" (UniqueName: \"kubernetes.io/projected/9b861750-2517-4cf3-9c78-930480711af0-kube-api-access-xh6c5\") on node \"crc\" DevicePath \"\"" Nov 22 08:10:33 crc kubenswrapper[4789]: I1122 08:10:33.724910 4789 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/9b861750-2517-4cf3-9c78-930480711af0-console-config\") on node \"crc\" DevicePath \"\"" Nov 22 08:10:33 crc kubenswrapper[4789]: I1122 08:10:33.724924 4789 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/9b861750-2517-4cf3-9c78-930480711af0-console-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 08:10:33 crc kubenswrapper[4789]: I1122 08:10:33.724935 4789 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/9b861750-2517-4cf3-9c78-930480711af0-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 08:10:33 crc kubenswrapper[4789]: I1122 08:10:33.724947 4789 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/9b861750-2517-4cf3-9c78-930480711af0-console-oauth-config\") on node \"crc\" DevicePath \"\"" Nov 22 08:10:33 crc kubenswrapper[4789]: I1122 08:10:33.724960 4789 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9b861750-2517-4cf3-9c78-930480711af0-service-ca\") on node \"crc\" DevicePath \"\"" Nov 22 08:10:33 crc kubenswrapper[4789]: I1122 08:10:33.724972 4789 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9b861750-2517-4cf3-9c78-930480711af0-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:10:34 crc kubenswrapper[4789]: I1122 08:10:34.114689 4789 generic.go:334] "Generic (PLEG): container finished" podID="8f9ce070-1816-46d5-a16f-b3b8d1a1f8f3" containerID="1908088c65407506fd57ea98dd218197e0fbc837c77fbb96c9a514a01c20814c" exitCode=0 Nov 22 08:10:34 crc kubenswrapper[4789]: I1122 08:10:34.114814 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6vdrbs" event={"ID":"8f9ce070-1816-46d5-a16f-b3b8d1a1f8f3","Type":"ContainerDied","Data":"1908088c65407506fd57ea98dd218197e0fbc837c77fbb96c9a514a01c20814c"} Nov 22 08:10:34 crc kubenswrapper[4789]: I1122 08:10:34.119709 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-pb98x_9b861750-2517-4cf3-9c78-930480711af0/console/0.log" Nov 22 08:10:34 crc kubenswrapper[4789]: I1122 08:10:34.119835 4789 generic.go:334] "Generic (PLEG): container finished" podID="9b861750-2517-4cf3-9c78-930480711af0" containerID="ddcb04daed5507caf65f440b3db89e0e66a92775018aca69eac59b75a82ace6f" exitCode=2 Nov 22 08:10:34 crc kubenswrapper[4789]: I1122 08:10:34.119876 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-pb98x" event={"ID":"9b861750-2517-4cf3-9c78-930480711af0","Type":"ContainerDied","Data":"ddcb04daed5507caf65f440b3db89e0e66a92775018aca69eac59b75a82ace6f"} Nov 22 08:10:34 crc kubenswrapper[4789]: I1122 08:10:34.119908 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-pb98x" event={"ID":"9b861750-2517-4cf3-9c78-930480711af0","Type":"ContainerDied","Data":"ec47c547dee63090f313aec803697265dcce9767c0aacf0c48dddcb8f18c8b4c"} Nov 22 08:10:34 crc kubenswrapper[4789]: I1122 08:10:34.119936 4789 scope.go:117] "RemoveContainer" containerID="ddcb04daed5507caf65f440b3db89e0e66a92775018aca69eac59b75a82ace6f" Nov 22 08:10:34 crc kubenswrapper[4789]: I1122 08:10:34.120081 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-pb98x" Nov 22 08:10:34 crc kubenswrapper[4789]: I1122 08:10:34.151929 4789 scope.go:117] "RemoveContainer" containerID="ddcb04daed5507caf65f440b3db89e0e66a92775018aca69eac59b75a82ace6f" Nov 22 08:10:34 crc kubenswrapper[4789]: E1122 08:10:34.154702 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ddcb04daed5507caf65f440b3db89e0e66a92775018aca69eac59b75a82ace6f\": container with ID starting with ddcb04daed5507caf65f440b3db89e0e66a92775018aca69eac59b75a82ace6f not found: ID does not exist" containerID="ddcb04daed5507caf65f440b3db89e0e66a92775018aca69eac59b75a82ace6f" Nov 22 08:10:34 crc kubenswrapper[4789]: I1122 08:10:34.154786 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ddcb04daed5507caf65f440b3db89e0e66a92775018aca69eac59b75a82ace6f"} err="failed to get container status \"ddcb04daed5507caf65f440b3db89e0e66a92775018aca69eac59b75a82ace6f\": rpc error: code = NotFound desc = could not find container \"ddcb04daed5507caf65f440b3db89e0e66a92775018aca69eac59b75a82ace6f\": container with ID starting with ddcb04daed5507caf65f440b3db89e0e66a92775018aca69eac59b75a82ace6f not found: ID does not exist" Nov 22 08:10:34 crc kubenswrapper[4789]: I1122 08:10:34.158512 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-pb98x"] Nov 22 08:10:34 crc kubenswrapper[4789]: I1122 08:10:34.162143 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-pb98x"] Nov 22 08:10:35 crc kubenswrapper[4789]: I1122 08:10:35.351446 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6vdrbs" Nov 22 08:10:35 crc kubenswrapper[4789]: I1122 08:10:35.550908 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f7nj4\" (UniqueName: \"kubernetes.io/projected/8f9ce070-1816-46d5-a16f-b3b8d1a1f8f3-kube-api-access-f7nj4\") pod \"8f9ce070-1816-46d5-a16f-b3b8d1a1f8f3\" (UID: \"8f9ce070-1816-46d5-a16f-b3b8d1a1f8f3\") " Nov 22 08:10:35 crc kubenswrapper[4789]: I1122 08:10:35.551808 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8f9ce070-1816-46d5-a16f-b3b8d1a1f8f3-bundle\") pod \"8f9ce070-1816-46d5-a16f-b3b8d1a1f8f3\" (UID: \"8f9ce070-1816-46d5-a16f-b3b8d1a1f8f3\") " Nov 22 08:10:35 crc kubenswrapper[4789]: I1122 08:10:35.551865 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8f9ce070-1816-46d5-a16f-b3b8d1a1f8f3-util\") pod \"8f9ce070-1816-46d5-a16f-b3b8d1a1f8f3\" (UID: \"8f9ce070-1816-46d5-a16f-b3b8d1a1f8f3\") " Nov 22 08:10:35 crc kubenswrapper[4789]: I1122 08:10:35.553088 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f9ce070-1816-46d5-a16f-b3b8d1a1f8f3-bundle" (OuterVolumeSpecName: "bundle") pod "8f9ce070-1816-46d5-a16f-b3b8d1a1f8f3" (UID: "8f9ce070-1816-46d5-a16f-b3b8d1a1f8f3"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:10:35 crc kubenswrapper[4789]: I1122 08:10:35.556257 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f9ce070-1816-46d5-a16f-b3b8d1a1f8f3-kube-api-access-f7nj4" (OuterVolumeSpecName: "kube-api-access-f7nj4") pod "8f9ce070-1816-46d5-a16f-b3b8d1a1f8f3" (UID: "8f9ce070-1816-46d5-a16f-b3b8d1a1f8f3"). InnerVolumeSpecName "kube-api-access-f7nj4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:10:35 crc kubenswrapper[4789]: I1122 08:10:35.566148 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f9ce070-1816-46d5-a16f-b3b8d1a1f8f3-util" (OuterVolumeSpecName: "util") pod "8f9ce070-1816-46d5-a16f-b3b8d1a1f8f3" (UID: "8f9ce070-1816-46d5-a16f-b3b8d1a1f8f3"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:10:35 crc kubenswrapper[4789]: I1122 08:10:35.652861 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f7nj4\" (UniqueName: \"kubernetes.io/projected/8f9ce070-1816-46d5-a16f-b3b8d1a1f8f3-kube-api-access-f7nj4\") on node \"crc\" DevicePath \"\"" Nov 22 08:10:35 crc kubenswrapper[4789]: I1122 08:10:35.652912 4789 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8f9ce070-1816-46d5-a16f-b3b8d1a1f8f3-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:10:35 crc kubenswrapper[4789]: I1122 08:10:35.652928 4789 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8f9ce070-1816-46d5-a16f-b3b8d1a1f8f3-util\") on node \"crc\" DevicePath \"\"" Nov 22 08:10:35 crc kubenswrapper[4789]: I1122 08:10:35.978251 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9b861750-2517-4cf3-9c78-930480711af0" path="/var/lib/kubelet/pods/9b861750-2517-4cf3-9c78-930480711af0/volumes" Nov 22 08:10:36 crc kubenswrapper[4789]: I1122 08:10:36.133475 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6vdrbs" event={"ID":"8f9ce070-1816-46d5-a16f-b3b8d1a1f8f3","Type":"ContainerDied","Data":"377ba8bd1fcd1b7841d0cedea5e8150ae20f17dd3867231366c0d5d912815ed1"} Nov 22 08:10:36 crc kubenswrapper[4789]: I1122 08:10:36.133514 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="377ba8bd1fcd1b7841d0cedea5e8150ae20f17dd3867231366c0d5d912815ed1" Nov 22 08:10:36 crc kubenswrapper[4789]: I1122 08:10:36.133519 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6vdrbs" Nov 22 08:10:45 crc kubenswrapper[4789]: I1122 08:10:45.852979 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-686d48fbd8-6jwbz"] Nov 22 08:10:45 crc kubenswrapper[4789]: E1122 08:10:45.854052 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b861750-2517-4cf3-9c78-930480711af0" containerName="console" Nov 22 08:10:45 crc kubenswrapper[4789]: I1122 08:10:45.854071 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b861750-2517-4cf3-9c78-930480711af0" containerName="console" Nov 22 08:10:45 crc kubenswrapper[4789]: E1122 08:10:45.854099 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f9ce070-1816-46d5-a16f-b3b8d1a1f8f3" containerName="util" Nov 22 08:10:45 crc kubenswrapper[4789]: I1122 08:10:45.854107 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f9ce070-1816-46d5-a16f-b3b8d1a1f8f3" containerName="util" Nov 22 08:10:45 crc kubenswrapper[4789]: E1122 08:10:45.854118 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f9ce070-1816-46d5-a16f-b3b8d1a1f8f3" containerName="extract" Nov 22 08:10:45 crc kubenswrapper[4789]: I1122 08:10:45.854127 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f9ce070-1816-46d5-a16f-b3b8d1a1f8f3" containerName="extract" Nov 22 08:10:45 crc kubenswrapper[4789]: E1122 08:10:45.854140 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f9ce070-1816-46d5-a16f-b3b8d1a1f8f3" containerName="pull" Nov 22 08:10:45 crc kubenswrapper[4789]: I1122 08:10:45.854147 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f9ce070-1816-46d5-a16f-b3b8d1a1f8f3" containerName="pull" Nov 22 08:10:45 crc kubenswrapper[4789]: I1122 08:10:45.854264 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b861750-2517-4cf3-9c78-930480711af0" containerName="console" Nov 22 08:10:45 crc kubenswrapper[4789]: I1122 08:10:45.854283 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f9ce070-1816-46d5-a16f-b3b8d1a1f8f3" containerName="extract" Nov 22 08:10:45 crc kubenswrapper[4789]: I1122 08:10:45.854735 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-686d48fbd8-6jwbz" Nov 22 08:10:45 crc kubenswrapper[4789]: I1122 08:10:45.857654 4789 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Nov 22 08:10:45 crc kubenswrapper[4789]: I1122 08:10:45.858189 4789 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Nov 22 08:10:45 crc kubenswrapper[4789]: I1122 08:10:45.859242 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Nov 22 08:10:45 crc kubenswrapper[4789]: I1122 08:10:45.859778 4789 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-qvhrh" Nov 22 08:10:45 crc kubenswrapper[4789]: I1122 08:10:45.861308 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Nov 22 08:10:45 crc kubenswrapper[4789]: I1122 08:10:45.925502 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-686d48fbd8-6jwbz"] Nov 22 08:10:45 crc kubenswrapper[4789]: I1122 08:10:45.976047 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rq76t\" (UniqueName: \"kubernetes.io/projected/701af49e-3c52-45f0-8d1e-c32b8c606b36-kube-api-access-rq76t\") pod \"metallb-operator-controller-manager-686d48fbd8-6jwbz\" (UID: \"701af49e-3c52-45f0-8d1e-c32b8c606b36\") " pod="metallb-system/metallb-operator-controller-manager-686d48fbd8-6jwbz" Nov 22 08:10:45 crc kubenswrapper[4789]: I1122 08:10:45.976166 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/701af49e-3c52-45f0-8d1e-c32b8c606b36-webhook-cert\") pod \"metallb-operator-controller-manager-686d48fbd8-6jwbz\" (UID: \"701af49e-3c52-45f0-8d1e-c32b8c606b36\") " pod="metallb-system/metallb-operator-controller-manager-686d48fbd8-6jwbz" Nov 22 08:10:45 crc kubenswrapper[4789]: I1122 08:10:45.976274 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/701af49e-3c52-45f0-8d1e-c32b8c606b36-apiservice-cert\") pod \"metallb-operator-controller-manager-686d48fbd8-6jwbz\" (UID: \"701af49e-3c52-45f0-8d1e-c32b8c606b36\") " pod="metallb-system/metallb-operator-controller-manager-686d48fbd8-6jwbz" Nov 22 08:10:46 crc kubenswrapper[4789]: I1122 08:10:46.077700 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/701af49e-3c52-45f0-8d1e-c32b8c606b36-webhook-cert\") pod \"metallb-operator-controller-manager-686d48fbd8-6jwbz\" (UID: \"701af49e-3c52-45f0-8d1e-c32b8c606b36\") " pod="metallb-system/metallb-operator-controller-manager-686d48fbd8-6jwbz" Nov 22 08:10:46 crc kubenswrapper[4789]: I1122 08:10:46.077778 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/701af49e-3c52-45f0-8d1e-c32b8c606b36-apiservice-cert\") pod \"metallb-operator-controller-manager-686d48fbd8-6jwbz\" (UID: \"701af49e-3c52-45f0-8d1e-c32b8c606b36\") " pod="metallb-system/metallb-operator-controller-manager-686d48fbd8-6jwbz" Nov 22 08:10:46 crc kubenswrapper[4789]: I1122 08:10:46.077820 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rq76t\" (UniqueName: \"kubernetes.io/projected/701af49e-3c52-45f0-8d1e-c32b8c606b36-kube-api-access-rq76t\") pod \"metallb-operator-controller-manager-686d48fbd8-6jwbz\" (UID: \"701af49e-3c52-45f0-8d1e-c32b8c606b36\") " pod="metallb-system/metallb-operator-controller-manager-686d48fbd8-6jwbz" Nov 22 08:10:46 crc kubenswrapper[4789]: I1122 08:10:46.083496 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/701af49e-3c52-45f0-8d1e-c32b8c606b36-webhook-cert\") pod \"metallb-operator-controller-manager-686d48fbd8-6jwbz\" (UID: \"701af49e-3c52-45f0-8d1e-c32b8c606b36\") " pod="metallb-system/metallb-operator-controller-manager-686d48fbd8-6jwbz" Nov 22 08:10:46 crc kubenswrapper[4789]: I1122 08:10:46.083722 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/701af49e-3c52-45f0-8d1e-c32b8c606b36-apiservice-cert\") pod \"metallb-operator-controller-manager-686d48fbd8-6jwbz\" (UID: \"701af49e-3c52-45f0-8d1e-c32b8c606b36\") " pod="metallb-system/metallb-operator-controller-manager-686d48fbd8-6jwbz" Nov 22 08:10:46 crc kubenswrapper[4789]: I1122 08:10:46.096929 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rq76t\" (UniqueName: \"kubernetes.io/projected/701af49e-3c52-45f0-8d1e-c32b8c606b36-kube-api-access-rq76t\") pod \"metallb-operator-controller-manager-686d48fbd8-6jwbz\" (UID: \"701af49e-3c52-45f0-8d1e-c32b8c606b36\") " pod="metallb-system/metallb-operator-controller-manager-686d48fbd8-6jwbz" Nov 22 08:10:46 crc kubenswrapper[4789]: I1122 08:10:46.172315 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-686d48fbd8-6jwbz" Nov 22 08:10:46 crc kubenswrapper[4789]: I1122 08:10:46.178863 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-6f6d9fb676-r76lj"] Nov 22 08:10:46 crc kubenswrapper[4789]: I1122 08:10:46.179572 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-6f6d9fb676-r76lj" Nov 22 08:10:46 crc kubenswrapper[4789]: I1122 08:10:46.182482 4789 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Nov 22 08:10:46 crc kubenswrapper[4789]: I1122 08:10:46.182703 4789 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Nov 22 08:10:46 crc kubenswrapper[4789]: I1122 08:10:46.182853 4789 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-rdrdz" Nov 22 08:10:46 crc kubenswrapper[4789]: I1122 08:10:46.246569 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-6f6d9fb676-r76lj"] Nov 22 08:10:46 crc kubenswrapper[4789]: I1122 08:10:46.380879 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jvlcr\" (UniqueName: \"kubernetes.io/projected/39bf0884-9c38-4839-9f2c-0a7502e961f8-kube-api-access-jvlcr\") pod \"metallb-operator-webhook-server-6f6d9fb676-r76lj\" (UID: \"39bf0884-9c38-4839-9f2c-0a7502e961f8\") " pod="metallb-system/metallb-operator-webhook-server-6f6d9fb676-r76lj" Nov 22 08:10:46 crc kubenswrapper[4789]: I1122 08:10:46.381261 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/39bf0884-9c38-4839-9f2c-0a7502e961f8-webhook-cert\") pod \"metallb-operator-webhook-server-6f6d9fb676-r76lj\" (UID: \"39bf0884-9c38-4839-9f2c-0a7502e961f8\") " pod="metallb-system/metallb-operator-webhook-server-6f6d9fb676-r76lj" Nov 22 08:10:46 crc kubenswrapper[4789]: I1122 08:10:46.381330 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/39bf0884-9c38-4839-9f2c-0a7502e961f8-apiservice-cert\") pod \"metallb-operator-webhook-server-6f6d9fb676-r76lj\" (UID: \"39bf0884-9c38-4839-9f2c-0a7502e961f8\") " pod="metallb-system/metallb-operator-webhook-server-6f6d9fb676-r76lj" Nov 22 08:10:46 crc kubenswrapper[4789]: I1122 08:10:46.453402 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-686d48fbd8-6jwbz"] Nov 22 08:10:46 crc kubenswrapper[4789]: W1122 08:10:46.468929 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod701af49e_3c52_45f0_8d1e_c32b8c606b36.slice/crio-82cb8b7926b2b5ce120cf884d639caff0f337cc4d9bcb5e88d7865c0a9d80d1b WatchSource:0}: Error finding container 82cb8b7926b2b5ce120cf884d639caff0f337cc4d9bcb5e88d7865c0a9d80d1b: Status 404 returned error can't find the container with id 82cb8b7926b2b5ce120cf884d639caff0f337cc4d9bcb5e88d7865c0a9d80d1b Nov 22 08:10:46 crc kubenswrapper[4789]: I1122 08:10:46.483056 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/39bf0884-9c38-4839-9f2c-0a7502e961f8-apiservice-cert\") pod \"metallb-operator-webhook-server-6f6d9fb676-r76lj\" (UID: \"39bf0884-9c38-4839-9f2c-0a7502e961f8\") " pod="metallb-system/metallb-operator-webhook-server-6f6d9fb676-r76lj" Nov 22 08:10:46 crc kubenswrapper[4789]: I1122 08:10:46.483113 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jvlcr\" (UniqueName: \"kubernetes.io/projected/39bf0884-9c38-4839-9f2c-0a7502e961f8-kube-api-access-jvlcr\") pod \"metallb-operator-webhook-server-6f6d9fb676-r76lj\" (UID: \"39bf0884-9c38-4839-9f2c-0a7502e961f8\") " pod="metallb-system/metallb-operator-webhook-server-6f6d9fb676-r76lj" Nov 22 08:10:46 crc kubenswrapper[4789]: I1122 08:10:46.483159 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/39bf0884-9c38-4839-9f2c-0a7502e961f8-webhook-cert\") pod \"metallb-operator-webhook-server-6f6d9fb676-r76lj\" (UID: \"39bf0884-9c38-4839-9f2c-0a7502e961f8\") " pod="metallb-system/metallb-operator-webhook-server-6f6d9fb676-r76lj" Nov 22 08:10:46 crc kubenswrapper[4789]: I1122 08:10:46.488010 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/39bf0884-9c38-4839-9f2c-0a7502e961f8-webhook-cert\") pod \"metallb-operator-webhook-server-6f6d9fb676-r76lj\" (UID: \"39bf0884-9c38-4839-9f2c-0a7502e961f8\") " pod="metallb-system/metallb-operator-webhook-server-6f6d9fb676-r76lj" Nov 22 08:10:46 crc kubenswrapper[4789]: I1122 08:10:46.488053 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/39bf0884-9c38-4839-9f2c-0a7502e961f8-apiservice-cert\") pod \"metallb-operator-webhook-server-6f6d9fb676-r76lj\" (UID: \"39bf0884-9c38-4839-9f2c-0a7502e961f8\") " pod="metallb-system/metallb-operator-webhook-server-6f6d9fb676-r76lj" Nov 22 08:10:46 crc kubenswrapper[4789]: I1122 08:10:46.500348 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jvlcr\" (UniqueName: \"kubernetes.io/projected/39bf0884-9c38-4839-9f2c-0a7502e961f8-kube-api-access-jvlcr\") pod \"metallb-operator-webhook-server-6f6d9fb676-r76lj\" (UID: \"39bf0884-9c38-4839-9f2c-0a7502e961f8\") " pod="metallb-system/metallb-operator-webhook-server-6f6d9fb676-r76lj" Nov 22 08:10:46 crc kubenswrapper[4789]: I1122 08:10:46.542316 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-6f6d9fb676-r76lj" Nov 22 08:10:46 crc kubenswrapper[4789]: I1122 08:10:46.816116 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-6f6d9fb676-r76lj"] Nov 22 08:10:46 crc kubenswrapper[4789]: W1122 08:10:46.830182 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod39bf0884_9c38_4839_9f2c_0a7502e961f8.slice/crio-6f77d6f348ff2fb28a2a4143aed47dad53d8489cd93ab3d58b54410bb6b79c31 WatchSource:0}: Error finding container 6f77d6f348ff2fb28a2a4143aed47dad53d8489cd93ab3d58b54410bb6b79c31: Status 404 returned error can't find the container with id 6f77d6f348ff2fb28a2a4143aed47dad53d8489cd93ab3d58b54410bb6b79c31 Nov 22 08:10:47 crc kubenswrapper[4789]: I1122 08:10:47.200579 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-686d48fbd8-6jwbz" event={"ID":"701af49e-3c52-45f0-8d1e-c32b8c606b36","Type":"ContainerStarted","Data":"82cb8b7926b2b5ce120cf884d639caff0f337cc4d9bcb5e88d7865c0a9d80d1b"} Nov 22 08:10:47 crc kubenswrapper[4789]: I1122 08:10:47.201554 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-6f6d9fb676-r76lj" event={"ID":"39bf0884-9c38-4839-9f2c-0a7502e961f8","Type":"ContainerStarted","Data":"6f77d6f348ff2fb28a2a4143aed47dad53d8489cd93ab3d58b54410bb6b79c31"} Nov 22 08:10:52 crc kubenswrapper[4789]: I1122 08:10:52.224834 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-6f6d9fb676-r76lj" event={"ID":"39bf0884-9c38-4839-9f2c-0a7502e961f8","Type":"ContainerStarted","Data":"114f1e49f5899b5729e3eaff5510906bb36a4d0ed1b56c7fe70d6752999a5b36"} Nov 22 08:10:52 crc kubenswrapper[4789]: I1122 08:10:52.226249 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-6f6d9fb676-r76lj" Nov 22 08:10:52 crc kubenswrapper[4789]: I1122 08:10:52.226551 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-686d48fbd8-6jwbz" event={"ID":"701af49e-3c52-45f0-8d1e-c32b8c606b36","Type":"ContainerStarted","Data":"11e8b206077e173986742ac482ddf1d7c0c0432eeb9d56883cc49fd6c776e94d"} Nov 22 08:10:52 crc kubenswrapper[4789]: I1122 08:10:52.226792 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-686d48fbd8-6jwbz" Nov 22 08:10:52 crc kubenswrapper[4789]: I1122 08:10:52.243896 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-6f6d9fb676-r76lj" podStartSLOduration=1.268885988 podStartE2EDuration="6.243873035s" podCreationTimestamp="2025-11-22 08:10:46 +0000 UTC" firstStartedPulling="2025-11-22 08:10:46.833061337 +0000 UTC m=+861.067461610" lastFinishedPulling="2025-11-22 08:10:51.808048384 +0000 UTC m=+866.042448657" observedRunningTime="2025-11-22 08:10:52.240841123 +0000 UTC m=+866.475241416" watchObservedRunningTime="2025-11-22 08:10:52.243873035 +0000 UTC m=+866.478273338" Nov 22 08:10:52 crc kubenswrapper[4789]: I1122 08:10:52.265110 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-686d48fbd8-6jwbz" podStartSLOduration=1.966003809 podStartE2EDuration="7.265086882s" podCreationTimestamp="2025-11-22 08:10:45 +0000 UTC" firstStartedPulling="2025-11-22 08:10:46.472417258 +0000 UTC m=+860.706817531" lastFinishedPulling="2025-11-22 08:10:51.771500331 +0000 UTC m=+866.005900604" observedRunningTime="2025-11-22 08:10:52.260035994 +0000 UTC m=+866.494436277" watchObservedRunningTime="2025-11-22 08:10:52.265086882 +0000 UTC m=+866.499487155" Nov 22 08:11:06 crc kubenswrapper[4789]: I1122 08:11:06.547552 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-6f6d9fb676-r76lj" Nov 22 08:11:26 crc kubenswrapper[4789]: I1122 08:11:26.175188 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-686d48fbd8-6jwbz" Nov 22 08:11:26 crc kubenswrapper[4789]: I1122 08:11:26.855354 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-qwjlg"] Nov 22 08:11:26 crc kubenswrapper[4789]: I1122 08:11:26.857538 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-qwjlg" Nov 22 08:11:26 crc kubenswrapper[4789]: I1122 08:11:26.859865 4789 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-f52xq" Nov 22 08:11:26 crc kubenswrapper[4789]: I1122 08:11:26.860302 4789 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Nov 22 08:11:26 crc kubenswrapper[4789]: I1122 08:11:26.860363 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Nov 22 08:11:26 crc kubenswrapper[4789]: I1122 08:11:26.885393 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-4shz4"] Nov 22 08:11:26 crc kubenswrapper[4789]: I1122 08:11:26.886384 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-6998585d5-4shz4" Nov 22 08:11:26 crc kubenswrapper[4789]: I1122 08:11:26.888367 4789 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Nov 22 08:11:26 crc kubenswrapper[4789]: I1122 08:11:26.902206 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-4shz4"] Nov 22 08:11:26 crc kubenswrapper[4789]: I1122 08:11:26.957110 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-6tqgm"] Nov 22 08:11:26 crc kubenswrapper[4789]: I1122 08:11:26.958037 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-6tqgm" Nov 22 08:11:26 crc kubenswrapper[4789]: I1122 08:11:26.959940 4789 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Nov 22 08:11:26 crc kubenswrapper[4789]: I1122 08:11:26.960226 4789 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-xcb4t" Nov 22 08:11:26 crc kubenswrapper[4789]: I1122 08:11:26.960391 4789 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Nov 22 08:11:26 crc kubenswrapper[4789]: I1122 08:11:26.960561 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Nov 22 08:11:26 crc kubenswrapper[4789]: I1122 08:11:26.964387 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/4b633875-b147-4ac7-979c-07a76bedb3fb-frr-startup\") pod \"frr-k8s-qwjlg\" (UID: \"4b633875-b147-4ac7-979c-07a76bedb3fb\") " pod="metallb-system/frr-k8s-qwjlg" Nov 22 08:11:26 crc kubenswrapper[4789]: I1122 08:11:26.964462 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cs77z\" (UniqueName: \"kubernetes.io/projected/4b633875-b147-4ac7-979c-07a76bedb3fb-kube-api-access-cs77z\") pod \"frr-k8s-qwjlg\" (UID: \"4b633875-b147-4ac7-979c-07a76bedb3fb\") " pod="metallb-system/frr-k8s-qwjlg" Nov 22 08:11:26 crc kubenswrapper[4789]: I1122 08:11:26.964491 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/4b633875-b147-4ac7-979c-07a76bedb3fb-reloader\") pod \"frr-k8s-qwjlg\" (UID: \"4b633875-b147-4ac7-979c-07a76bedb3fb\") " pod="metallb-system/frr-k8s-qwjlg" Nov 22 08:11:26 crc kubenswrapper[4789]: I1122 08:11:26.964517 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/4b633875-b147-4ac7-979c-07a76bedb3fb-metrics\") pod \"frr-k8s-qwjlg\" (UID: \"4b633875-b147-4ac7-979c-07a76bedb3fb\") " pod="metallb-system/frr-k8s-qwjlg" Nov 22 08:11:26 crc kubenswrapper[4789]: I1122 08:11:26.964572 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/4b633875-b147-4ac7-979c-07a76bedb3fb-frr-sockets\") pod \"frr-k8s-qwjlg\" (UID: \"4b633875-b147-4ac7-979c-07a76bedb3fb\") " pod="metallb-system/frr-k8s-qwjlg" Nov 22 08:11:26 crc kubenswrapper[4789]: I1122 08:11:26.964612 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4b633875-b147-4ac7-979c-07a76bedb3fb-metrics-certs\") pod \"frr-k8s-qwjlg\" (UID: \"4b633875-b147-4ac7-979c-07a76bedb3fb\") " pod="metallb-system/frr-k8s-qwjlg" Nov 22 08:11:26 crc kubenswrapper[4789]: I1122 08:11:26.964637 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/4b633875-b147-4ac7-979c-07a76bedb3fb-frr-conf\") pod \"frr-k8s-qwjlg\" (UID: \"4b633875-b147-4ac7-979c-07a76bedb3fb\") " pod="metallb-system/frr-k8s-qwjlg" Nov 22 08:11:26 crc kubenswrapper[4789]: I1122 08:11:26.972239 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-6c7b4b5f48-wm2pz"] Nov 22 08:11:26 crc kubenswrapper[4789]: I1122 08:11:26.974526 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6c7b4b5f48-wm2pz" Nov 22 08:11:26 crc kubenswrapper[4789]: I1122 08:11:26.976572 4789 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Nov 22 08:11:26 crc kubenswrapper[4789]: I1122 08:11:26.991788 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6c7b4b5f48-wm2pz"] Nov 22 08:11:27 crc kubenswrapper[4789]: I1122 08:11:27.066528 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1f4fdc0e-ffe3-4a34-a9b4-0b4c70fccefa-metrics-certs\") pod \"controller-6c7b4b5f48-wm2pz\" (UID: \"1f4fdc0e-ffe3-4a34-a9b4-0b4c70fccefa\") " pod="metallb-system/controller-6c7b4b5f48-wm2pz" Nov 22 08:11:27 crc kubenswrapper[4789]: I1122 08:11:27.066576 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/4b633875-b147-4ac7-979c-07a76bedb3fb-frr-sockets\") pod \"frr-k8s-qwjlg\" (UID: \"4b633875-b147-4ac7-979c-07a76bedb3fb\") " pod="metallb-system/frr-k8s-qwjlg" Nov 22 08:11:27 crc kubenswrapper[4789]: I1122 08:11:27.066605 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4b633875-b147-4ac7-979c-07a76bedb3fb-metrics-certs\") pod \"frr-k8s-qwjlg\" (UID: \"4b633875-b147-4ac7-979c-07a76bedb3fb\") " pod="metallb-system/frr-k8s-qwjlg" Nov 22 08:11:27 crc kubenswrapper[4789]: I1122 08:11:27.066640 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/4b633875-b147-4ac7-979c-07a76bedb3fb-frr-conf\") pod \"frr-k8s-qwjlg\" (UID: \"4b633875-b147-4ac7-979c-07a76bedb3fb\") " pod="metallb-system/frr-k8s-qwjlg" Nov 22 08:11:27 crc kubenswrapper[4789]: I1122 08:11:27.066672 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/4b633875-b147-4ac7-979c-07a76bedb3fb-frr-startup\") pod \"frr-k8s-qwjlg\" (UID: \"4b633875-b147-4ac7-979c-07a76bedb3fb\") " pod="metallb-system/frr-k8s-qwjlg" Nov 22 08:11:27 crc kubenswrapper[4789]: I1122 08:11:27.066689 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1f4fdc0e-ffe3-4a34-a9b4-0b4c70fccefa-cert\") pod \"controller-6c7b4b5f48-wm2pz\" (UID: \"1f4fdc0e-ffe3-4a34-a9b4-0b4c70fccefa\") " pod="metallb-system/controller-6c7b4b5f48-wm2pz" Nov 22 08:11:27 crc kubenswrapper[4789]: I1122 08:11:27.066708 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xjv27\" (UniqueName: \"kubernetes.io/projected/eda37095-1c8e-49ab-b085-20c8cf688be0-kube-api-access-xjv27\") pod \"frr-k8s-webhook-server-6998585d5-4shz4\" (UID: \"eda37095-1c8e-49ab-b085-20c8cf688be0\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-4shz4" Nov 22 08:11:27 crc kubenswrapper[4789]: I1122 08:11:27.066729 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cs77z\" (UniqueName: \"kubernetes.io/projected/4b633875-b147-4ac7-979c-07a76bedb3fb-kube-api-access-cs77z\") pod \"frr-k8s-qwjlg\" (UID: \"4b633875-b147-4ac7-979c-07a76bedb3fb\") " pod="metallb-system/frr-k8s-qwjlg" Nov 22 08:11:27 crc kubenswrapper[4789]: I1122 08:11:27.066761 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/4b633875-b147-4ac7-979c-07a76bedb3fb-reloader\") pod \"frr-k8s-qwjlg\" (UID: \"4b633875-b147-4ac7-979c-07a76bedb3fb\") " pod="metallb-system/frr-k8s-qwjlg" Nov 22 08:11:27 crc kubenswrapper[4789]: I1122 08:11:27.066776 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6a534620-1ee2-419c-ae0d-cdf2495f1ea0-metrics-certs\") pod \"speaker-6tqgm\" (UID: \"6a534620-1ee2-419c-ae0d-cdf2495f1ea0\") " pod="metallb-system/speaker-6tqgm" Nov 22 08:11:27 crc kubenswrapper[4789]: I1122 08:11:27.066791 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4nflz\" (UniqueName: \"kubernetes.io/projected/6a534620-1ee2-419c-ae0d-cdf2495f1ea0-kube-api-access-4nflz\") pod \"speaker-6tqgm\" (UID: \"6a534620-1ee2-419c-ae0d-cdf2495f1ea0\") " pod="metallb-system/speaker-6tqgm" Nov 22 08:11:27 crc kubenswrapper[4789]: I1122 08:11:27.066807 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lkllf\" (UniqueName: \"kubernetes.io/projected/1f4fdc0e-ffe3-4a34-a9b4-0b4c70fccefa-kube-api-access-lkllf\") pod \"controller-6c7b4b5f48-wm2pz\" (UID: \"1f4fdc0e-ffe3-4a34-a9b4-0b4c70fccefa\") " pod="metallb-system/controller-6c7b4b5f48-wm2pz" Nov 22 08:11:27 crc kubenswrapper[4789]: I1122 08:11:27.066825 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/6a534620-1ee2-419c-ae0d-cdf2495f1ea0-memberlist\") pod \"speaker-6tqgm\" (UID: \"6a534620-1ee2-419c-ae0d-cdf2495f1ea0\") " pod="metallb-system/speaker-6tqgm" Nov 22 08:11:27 crc kubenswrapper[4789]: I1122 08:11:27.066840 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/eda37095-1c8e-49ab-b085-20c8cf688be0-cert\") pod \"frr-k8s-webhook-server-6998585d5-4shz4\" (UID: \"eda37095-1c8e-49ab-b085-20c8cf688be0\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-4shz4" Nov 22 08:11:27 crc kubenswrapper[4789]: I1122 08:11:27.067760 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/4b633875-b147-4ac7-979c-07a76bedb3fb-metrics\") pod \"frr-k8s-qwjlg\" (UID: \"4b633875-b147-4ac7-979c-07a76bedb3fb\") " pod="metallb-system/frr-k8s-qwjlg" Nov 22 08:11:27 crc kubenswrapper[4789]: I1122 08:11:27.067849 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/4b633875-b147-4ac7-979c-07a76bedb3fb-frr-sockets\") pod \"frr-k8s-qwjlg\" (UID: \"4b633875-b147-4ac7-979c-07a76bedb3fb\") " pod="metallb-system/frr-k8s-qwjlg" Nov 22 08:11:27 crc kubenswrapper[4789]: I1122 08:11:27.067923 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/6a534620-1ee2-419c-ae0d-cdf2495f1ea0-metallb-excludel2\") pod \"speaker-6tqgm\" (UID: \"6a534620-1ee2-419c-ae0d-cdf2495f1ea0\") " pod="metallb-system/speaker-6tqgm" Nov 22 08:11:27 crc kubenswrapper[4789]: I1122 08:11:27.067959 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/4b633875-b147-4ac7-979c-07a76bedb3fb-frr-conf\") pod \"frr-k8s-qwjlg\" (UID: \"4b633875-b147-4ac7-979c-07a76bedb3fb\") " pod="metallb-system/frr-k8s-qwjlg" Nov 22 08:11:27 crc kubenswrapper[4789]: I1122 08:11:27.068692 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/4b633875-b147-4ac7-979c-07a76bedb3fb-frr-startup\") pod \"frr-k8s-qwjlg\" (UID: \"4b633875-b147-4ac7-979c-07a76bedb3fb\") " pod="metallb-system/frr-k8s-qwjlg" Nov 22 08:11:27 crc kubenswrapper[4789]: I1122 08:11:27.068934 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/4b633875-b147-4ac7-979c-07a76bedb3fb-reloader\") pod \"frr-k8s-qwjlg\" (UID: \"4b633875-b147-4ac7-979c-07a76bedb3fb\") " pod="metallb-system/frr-k8s-qwjlg" Nov 22 08:11:27 crc kubenswrapper[4789]: I1122 08:11:27.069305 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/4b633875-b147-4ac7-979c-07a76bedb3fb-metrics\") pod \"frr-k8s-qwjlg\" (UID: \"4b633875-b147-4ac7-979c-07a76bedb3fb\") " pod="metallb-system/frr-k8s-qwjlg" Nov 22 08:11:27 crc kubenswrapper[4789]: I1122 08:11:27.074398 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4b633875-b147-4ac7-979c-07a76bedb3fb-metrics-certs\") pod \"frr-k8s-qwjlg\" (UID: \"4b633875-b147-4ac7-979c-07a76bedb3fb\") " pod="metallb-system/frr-k8s-qwjlg" Nov 22 08:11:27 crc kubenswrapper[4789]: I1122 08:11:27.090405 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cs77z\" (UniqueName: \"kubernetes.io/projected/4b633875-b147-4ac7-979c-07a76bedb3fb-kube-api-access-cs77z\") pod \"frr-k8s-qwjlg\" (UID: \"4b633875-b147-4ac7-979c-07a76bedb3fb\") " pod="metallb-system/frr-k8s-qwjlg" Nov 22 08:11:27 crc kubenswrapper[4789]: I1122 08:11:27.169541 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1f4fdc0e-ffe3-4a34-a9b4-0b4c70fccefa-metrics-certs\") pod \"controller-6c7b4b5f48-wm2pz\" (UID: \"1f4fdc0e-ffe3-4a34-a9b4-0b4c70fccefa\") " pod="metallb-system/controller-6c7b4b5f48-wm2pz" Nov 22 08:11:27 crc kubenswrapper[4789]: I1122 08:11:27.169632 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1f4fdc0e-ffe3-4a34-a9b4-0b4c70fccefa-cert\") pod \"controller-6c7b4b5f48-wm2pz\" (UID: \"1f4fdc0e-ffe3-4a34-a9b4-0b4c70fccefa\") " pod="metallb-system/controller-6c7b4b5f48-wm2pz" Nov 22 08:11:27 crc kubenswrapper[4789]: I1122 08:11:27.169693 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xjv27\" (UniqueName: \"kubernetes.io/projected/eda37095-1c8e-49ab-b085-20c8cf688be0-kube-api-access-xjv27\") pod \"frr-k8s-webhook-server-6998585d5-4shz4\" (UID: \"eda37095-1c8e-49ab-b085-20c8cf688be0\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-4shz4" Nov 22 08:11:27 crc kubenswrapper[4789]: I1122 08:11:27.169728 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6a534620-1ee2-419c-ae0d-cdf2495f1ea0-metrics-certs\") pod \"speaker-6tqgm\" (UID: \"6a534620-1ee2-419c-ae0d-cdf2495f1ea0\") " pod="metallb-system/speaker-6tqgm" Nov 22 08:11:27 crc kubenswrapper[4789]: I1122 08:11:27.169774 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4nflz\" (UniqueName: \"kubernetes.io/projected/6a534620-1ee2-419c-ae0d-cdf2495f1ea0-kube-api-access-4nflz\") pod \"speaker-6tqgm\" (UID: \"6a534620-1ee2-419c-ae0d-cdf2495f1ea0\") " pod="metallb-system/speaker-6tqgm" Nov 22 08:11:27 crc kubenswrapper[4789]: I1122 08:11:27.169797 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lkllf\" (UniqueName: \"kubernetes.io/projected/1f4fdc0e-ffe3-4a34-a9b4-0b4c70fccefa-kube-api-access-lkllf\") pod \"controller-6c7b4b5f48-wm2pz\" (UID: \"1f4fdc0e-ffe3-4a34-a9b4-0b4c70fccefa\") " pod="metallb-system/controller-6c7b4b5f48-wm2pz" Nov 22 08:11:27 crc kubenswrapper[4789]: I1122 08:11:27.169821 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/6a534620-1ee2-419c-ae0d-cdf2495f1ea0-memberlist\") pod \"speaker-6tqgm\" (UID: \"6a534620-1ee2-419c-ae0d-cdf2495f1ea0\") " pod="metallb-system/speaker-6tqgm" Nov 22 08:11:27 crc kubenswrapper[4789]: I1122 08:11:27.169846 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/eda37095-1c8e-49ab-b085-20c8cf688be0-cert\") pod \"frr-k8s-webhook-server-6998585d5-4shz4\" (UID: \"eda37095-1c8e-49ab-b085-20c8cf688be0\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-4shz4" Nov 22 08:11:27 crc kubenswrapper[4789]: I1122 08:11:27.169875 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/6a534620-1ee2-419c-ae0d-cdf2495f1ea0-metallb-excludel2\") pod \"speaker-6tqgm\" (UID: \"6a534620-1ee2-419c-ae0d-cdf2495f1ea0\") " pod="metallb-system/speaker-6tqgm" Nov 22 08:11:27 crc kubenswrapper[4789]: E1122 08:11:27.169816 4789 secret.go:188] Couldn't get secret metallb-system/controller-certs-secret: secret "controller-certs-secret" not found Nov 22 08:11:27 crc kubenswrapper[4789]: E1122 08:11:27.170376 4789 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Nov 22 08:11:27 crc kubenswrapper[4789]: E1122 08:11:27.170414 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1f4fdc0e-ffe3-4a34-a9b4-0b4c70fccefa-metrics-certs podName:1f4fdc0e-ffe3-4a34-a9b4-0b4c70fccefa nodeName:}" failed. No retries permitted until 2025-11-22 08:11:27.67038403 +0000 UTC m=+901.904784383 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/1f4fdc0e-ffe3-4a34-a9b4-0b4c70fccefa-metrics-certs") pod "controller-6c7b4b5f48-wm2pz" (UID: "1f4fdc0e-ffe3-4a34-a9b4-0b4c70fccefa") : secret "controller-certs-secret" not found Nov 22 08:11:27 crc kubenswrapper[4789]: E1122 08:11:27.170434 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6a534620-1ee2-419c-ae0d-cdf2495f1ea0-memberlist podName:6a534620-1ee2-419c-ae0d-cdf2495f1ea0 nodeName:}" failed. No retries permitted until 2025-11-22 08:11:27.670425721 +0000 UTC m=+901.904826124 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/6a534620-1ee2-419c-ae0d-cdf2495f1ea0-memberlist") pod "speaker-6tqgm" (UID: "6a534620-1ee2-419c-ae0d-cdf2495f1ea0") : secret "metallb-memberlist" not found Nov 22 08:11:27 crc kubenswrapper[4789]: I1122 08:11:27.170664 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/6a534620-1ee2-419c-ae0d-cdf2495f1ea0-metallb-excludel2\") pod \"speaker-6tqgm\" (UID: \"6a534620-1ee2-419c-ae0d-cdf2495f1ea0\") " pod="metallb-system/speaker-6tqgm" Nov 22 08:11:27 crc kubenswrapper[4789]: I1122 08:11:27.174245 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-qwjlg" Nov 22 08:11:27 crc kubenswrapper[4789]: I1122 08:11:27.174270 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6a534620-1ee2-419c-ae0d-cdf2495f1ea0-metrics-certs\") pod \"speaker-6tqgm\" (UID: \"6a534620-1ee2-419c-ae0d-cdf2495f1ea0\") " pod="metallb-system/speaker-6tqgm" Nov 22 08:11:27 crc kubenswrapper[4789]: I1122 08:11:27.174278 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1f4fdc0e-ffe3-4a34-a9b4-0b4c70fccefa-cert\") pod \"controller-6c7b4b5f48-wm2pz\" (UID: \"1f4fdc0e-ffe3-4a34-a9b4-0b4c70fccefa\") " pod="metallb-system/controller-6c7b4b5f48-wm2pz" Nov 22 08:11:27 crc kubenswrapper[4789]: I1122 08:11:27.174623 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/eda37095-1c8e-49ab-b085-20c8cf688be0-cert\") pod \"frr-k8s-webhook-server-6998585d5-4shz4\" (UID: \"eda37095-1c8e-49ab-b085-20c8cf688be0\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-4shz4" Nov 22 08:11:27 crc kubenswrapper[4789]: I1122 08:11:27.188618 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lkllf\" (UniqueName: \"kubernetes.io/projected/1f4fdc0e-ffe3-4a34-a9b4-0b4c70fccefa-kube-api-access-lkllf\") pod \"controller-6c7b4b5f48-wm2pz\" (UID: \"1f4fdc0e-ffe3-4a34-a9b4-0b4c70fccefa\") " pod="metallb-system/controller-6c7b4b5f48-wm2pz" Nov 22 08:11:27 crc kubenswrapper[4789]: I1122 08:11:27.198243 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xjv27\" (UniqueName: \"kubernetes.io/projected/eda37095-1c8e-49ab-b085-20c8cf688be0-kube-api-access-xjv27\") pod \"frr-k8s-webhook-server-6998585d5-4shz4\" (UID: \"eda37095-1c8e-49ab-b085-20c8cf688be0\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-4shz4" Nov 22 08:11:27 crc kubenswrapper[4789]: I1122 08:11:27.198496 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-6998585d5-4shz4" Nov 22 08:11:27 crc kubenswrapper[4789]: I1122 08:11:27.199328 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4nflz\" (UniqueName: \"kubernetes.io/projected/6a534620-1ee2-419c-ae0d-cdf2495f1ea0-kube-api-access-4nflz\") pod \"speaker-6tqgm\" (UID: \"6a534620-1ee2-419c-ae0d-cdf2495f1ea0\") " pod="metallb-system/speaker-6tqgm" Nov 22 08:11:27 crc kubenswrapper[4789]: I1122 08:11:27.389627 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-4shz4"] Nov 22 08:11:27 crc kubenswrapper[4789]: W1122 08:11:27.400213 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeda37095_1c8e_49ab_b085_20c8cf688be0.slice/crio-cc4e57c6bc0bc0242809170865747706a01271d411810f8a97a847ba82a12f8a WatchSource:0}: Error finding container cc4e57c6bc0bc0242809170865747706a01271d411810f8a97a847ba82a12f8a: Status 404 returned error can't find the container with id cc4e57c6bc0bc0242809170865747706a01271d411810f8a97a847ba82a12f8a Nov 22 08:11:27 crc kubenswrapper[4789]: I1122 08:11:27.675461 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1f4fdc0e-ffe3-4a34-a9b4-0b4c70fccefa-metrics-certs\") pod \"controller-6c7b4b5f48-wm2pz\" (UID: \"1f4fdc0e-ffe3-4a34-a9b4-0b4c70fccefa\") " pod="metallb-system/controller-6c7b4b5f48-wm2pz" Nov 22 08:11:27 crc kubenswrapper[4789]: I1122 08:11:27.676040 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/6a534620-1ee2-419c-ae0d-cdf2495f1ea0-memberlist\") pod \"speaker-6tqgm\" (UID: \"6a534620-1ee2-419c-ae0d-cdf2495f1ea0\") " pod="metallb-system/speaker-6tqgm" Nov 22 08:11:27 crc kubenswrapper[4789]: E1122 08:11:27.676220 4789 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Nov 22 08:11:27 crc kubenswrapper[4789]: E1122 08:11:27.676314 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6a534620-1ee2-419c-ae0d-cdf2495f1ea0-memberlist podName:6a534620-1ee2-419c-ae0d-cdf2495f1ea0 nodeName:}" failed. No retries permitted until 2025-11-22 08:11:28.676292256 +0000 UTC m=+902.910692529 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/6a534620-1ee2-419c-ae0d-cdf2495f1ea0-memberlist") pod "speaker-6tqgm" (UID: "6a534620-1ee2-419c-ae0d-cdf2495f1ea0") : secret "metallb-memberlist" not found Nov 22 08:11:27 crc kubenswrapper[4789]: I1122 08:11:27.683383 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1f4fdc0e-ffe3-4a34-a9b4-0b4c70fccefa-metrics-certs\") pod \"controller-6c7b4b5f48-wm2pz\" (UID: \"1f4fdc0e-ffe3-4a34-a9b4-0b4c70fccefa\") " pod="metallb-system/controller-6c7b4b5f48-wm2pz" Nov 22 08:11:27 crc kubenswrapper[4789]: I1122 08:11:27.886386 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6c7b4b5f48-wm2pz" Nov 22 08:11:27 crc kubenswrapper[4789]: I1122 08:11:27.924194 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-6998585d5-4shz4" event={"ID":"eda37095-1c8e-49ab-b085-20c8cf688be0","Type":"ContainerStarted","Data":"cc4e57c6bc0bc0242809170865747706a01271d411810f8a97a847ba82a12f8a"} Nov 22 08:11:27 crc kubenswrapper[4789]: I1122 08:11:27.925264 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-qwjlg" event={"ID":"4b633875-b147-4ac7-979c-07a76bedb3fb","Type":"ContainerStarted","Data":"c933f356f5281eeb5c3d2764c057959ed74d75f13b4aef04213acc50919d5a15"} Nov 22 08:11:28 crc kubenswrapper[4789]: I1122 08:11:28.078818 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6c7b4b5f48-wm2pz"] Nov 22 08:11:28 crc kubenswrapper[4789]: I1122 08:11:28.687556 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/6a534620-1ee2-419c-ae0d-cdf2495f1ea0-memberlist\") pod \"speaker-6tqgm\" (UID: \"6a534620-1ee2-419c-ae0d-cdf2495f1ea0\") " pod="metallb-system/speaker-6tqgm" Nov 22 08:11:28 crc kubenswrapper[4789]: I1122 08:11:28.705049 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/6a534620-1ee2-419c-ae0d-cdf2495f1ea0-memberlist\") pod \"speaker-6tqgm\" (UID: \"6a534620-1ee2-419c-ae0d-cdf2495f1ea0\") " pod="metallb-system/speaker-6tqgm" Nov 22 08:11:28 crc kubenswrapper[4789]: I1122 08:11:28.777067 4789 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-xcb4t" Nov 22 08:11:28 crc kubenswrapper[4789]: I1122 08:11:28.786102 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-6tqgm" Nov 22 08:11:28 crc kubenswrapper[4789]: I1122 08:11:28.934396 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-wm2pz" event={"ID":"1f4fdc0e-ffe3-4a34-a9b4-0b4c70fccefa","Type":"ContainerStarted","Data":"2901b1ce3c74bdc2ca62d530c00928678b6621f26ff308d2a72588a91ad3cfb5"} Nov 22 08:11:28 crc kubenswrapper[4789]: I1122 08:11:28.934452 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-wm2pz" event={"ID":"1f4fdc0e-ffe3-4a34-a9b4-0b4c70fccefa","Type":"ContainerStarted","Data":"e0881ead59ed2f9d7d91329d3534ad020c60512f19f918c93c451ab437e2085e"} Nov 22 08:11:28 crc kubenswrapper[4789]: I1122 08:11:28.934471 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-wm2pz" event={"ID":"1f4fdc0e-ffe3-4a34-a9b4-0b4c70fccefa","Type":"ContainerStarted","Data":"9da5e1a1d49ec0298618d1ae42129eca9b7f7e524213fa8fb7d2b8f523857e4e"} Nov 22 08:11:28 crc kubenswrapper[4789]: I1122 08:11:28.934537 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-6c7b4b5f48-wm2pz" Nov 22 08:11:28 crc kubenswrapper[4789]: I1122 08:11:28.935599 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-6tqgm" event={"ID":"6a534620-1ee2-419c-ae0d-cdf2495f1ea0","Type":"ContainerStarted","Data":"94dbca4567dfce80f775d4ac94426c109d5f8404fecf94df82a9fcd1b742a3aa"} Nov 22 08:11:28 crc kubenswrapper[4789]: I1122 08:11:28.953609 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-6c7b4b5f48-wm2pz" podStartSLOduration=2.953453197 podStartE2EDuration="2.953453197s" podCreationTimestamp="2025-11-22 08:11:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:11:28.949988104 +0000 UTC m=+903.184388377" watchObservedRunningTime="2025-11-22 08:11:28.953453197 +0000 UTC m=+903.187853480" Nov 22 08:11:29 crc kubenswrapper[4789]: I1122 08:11:29.981157 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-6tqgm" event={"ID":"6a534620-1ee2-419c-ae0d-cdf2495f1ea0","Type":"ContainerStarted","Data":"3cc0ec7ca3b1bcf646572e47bfff084ef6df9b85d0fa2c2bc54bf404ed3c6841"} Nov 22 08:11:29 crc kubenswrapper[4789]: I1122 08:11:29.981524 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-6tqgm" event={"ID":"6a534620-1ee2-419c-ae0d-cdf2495f1ea0","Type":"ContainerStarted","Data":"5c79620f771da3cef7f2848712c9289681db03d7ccaa65e5b863a50a6a5cdd16"} Nov 22 08:11:29 crc kubenswrapper[4789]: I1122 08:11:29.994564 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-6tqgm" podStartSLOduration=3.9945416849999997 podStartE2EDuration="3.994541685s" podCreationTimestamp="2025-11-22 08:11:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:11:29.991854472 +0000 UTC m=+904.226254765" watchObservedRunningTime="2025-11-22 08:11:29.994541685 +0000 UTC m=+904.228941978" Nov 22 08:11:30 crc kubenswrapper[4789]: I1122 08:11:30.984655 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-6tqgm" Nov 22 08:11:35 crc kubenswrapper[4789]: I1122 08:11:35.007329 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-6998585d5-4shz4" event={"ID":"eda37095-1c8e-49ab-b085-20c8cf688be0","Type":"ContainerStarted","Data":"46c7ae9c96aedf35afc456792e12d20d64d85312dd85687d3925529f10f0ea28"} Nov 22 08:11:35 crc kubenswrapper[4789]: I1122 08:11:35.008147 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-6998585d5-4shz4" Nov 22 08:11:35 crc kubenswrapper[4789]: I1122 08:11:35.009329 4789 generic.go:334] "Generic (PLEG): container finished" podID="4b633875-b147-4ac7-979c-07a76bedb3fb" containerID="4b88cce35f3a244b1dad189b61f72889fc6b84eb80fcc1d44f3bb4104f0c6e4c" exitCode=0 Nov 22 08:11:35 crc kubenswrapper[4789]: I1122 08:11:35.009385 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-qwjlg" event={"ID":"4b633875-b147-4ac7-979c-07a76bedb3fb","Type":"ContainerDied","Data":"4b88cce35f3a244b1dad189b61f72889fc6b84eb80fcc1d44f3bb4104f0c6e4c"} Nov 22 08:11:35 crc kubenswrapper[4789]: I1122 08:11:35.029453 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-6998585d5-4shz4" podStartSLOduration=2.417310021 podStartE2EDuration="9.029421929s" podCreationTimestamp="2025-11-22 08:11:26 +0000 UTC" firstStartedPulling="2025-11-22 08:11:27.402849376 +0000 UTC m=+901.637249649" lastFinishedPulling="2025-11-22 08:11:34.014961284 +0000 UTC m=+908.249361557" observedRunningTime="2025-11-22 08:11:35.026920251 +0000 UTC m=+909.261320534" watchObservedRunningTime="2025-11-22 08:11:35.029421929 +0000 UTC m=+909.263822242" Nov 22 08:11:36 crc kubenswrapper[4789]: I1122 08:11:36.020486 4789 generic.go:334] "Generic (PLEG): container finished" podID="4b633875-b147-4ac7-979c-07a76bedb3fb" containerID="d6ec4ad043d9a8fbce7a5f6aa7d4e103ee5bcec4967ced208e50e8f86d21ad99" exitCode=0 Nov 22 08:11:36 crc kubenswrapper[4789]: I1122 08:11:36.020626 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-qwjlg" event={"ID":"4b633875-b147-4ac7-979c-07a76bedb3fb","Type":"ContainerDied","Data":"d6ec4ad043d9a8fbce7a5f6aa7d4e103ee5bcec4967ced208e50e8f86d21ad99"} Nov 22 08:11:37 crc kubenswrapper[4789]: I1122 08:11:37.027932 4789 generic.go:334] "Generic (PLEG): container finished" podID="4b633875-b147-4ac7-979c-07a76bedb3fb" containerID="79fdebade1097b7aab37d296532a0bfb7f2574916e744db91d05395567f525da" exitCode=0 Nov 22 08:11:37 crc kubenswrapper[4789]: I1122 08:11:37.027979 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-qwjlg" event={"ID":"4b633875-b147-4ac7-979c-07a76bedb3fb","Type":"ContainerDied","Data":"79fdebade1097b7aab37d296532a0bfb7f2574916e744db91d05395567f525da"} Nov 22 08:11:38 crc kubenswrapper[4789]: I1122 08:11:38.038463 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-qwjlg" event={"ID":"4b633875-b147-4ac7-979c-07a76bedb3fb","Type":"ContainerStarted","Data":"337c350475c6ee2e36839910d3c0fef3bf4d17363bee59f1bbcaf751562fcde5"} Nov 22 08:11:38 crc kubenswrapper[4789]: I1122 08:11:38.038503 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-qwjlg" event={"ID":"4b633875-b147-4ac7-979c-07a76bedb3fb","Type":"ContainerStarted","Data":"0beceea161fec50e2bb11d6ba731ed3662478223d0240b7128df9a384f3be7d1"} Nov 22 08:11:38 crc kubenswrapper[4789]: I1122 08:11:38.038513 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-qwjlg" event={"ID":"4b633875-b147-4ac7-979c-07a76bedb3fb","Type":"ContainerStarted","Data":"3b9e5ec84d3d5c942c83c139ead56db6379bd36b32c2ccf5282ee1209f6c5cc2"} Nov 22 08:11:38 crc kubenswrapper[4789]: I1122 08:11:38.038523 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-qwjlg" event={"ID":"4b633875-b147-4ac7-979c-07a76bedb3fb","Type":"ContainerStarted","Data":"3e3d9f8381f3e5ebe33d4433010a08482d17f86715bc853ad5a246ef76cd0533"} Nov 22 08:11:38 crc kubenswrapper[4789]: I1122 08:11:38.038531 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-qwjlg" event={"ID":"4b633875-b147-4ac7-979c-07a76bedb3fb","Type":"ContainerStarted","Data":"49ef23b94b08ba97e15ef1470645b6dc7f1bb9662a2dd8ee6ac3b49229f647a2"} Nov 22 08:11:38 crc kubenswrapper[4789]: I1122 08:11:38.038538 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-qwjlg" event={"ID":"4b633875-b147-4ac7-979c-07a76bedb3fb","Type":"ContainerStarted","Data":"03bf66f0fddb82940e1d4dc943f2c3de5700bafbc0747545c2a9ab7be74d16e4"} Nov 22 08:11:38 crc kubenswrapper[4789]: I1122 08:11:38.038624 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-qwjlg" Nov 22 08:11:38 crc kubenswrapper[4789]: I1122 08:11:38.063963 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-qwjlg" podStartSLOduration=5.377963615 podStartE2EDuration="12.06394303s" podCreationTimestamp="2025-11-22 08:11:26 +0000 UTC" firstStartedPulling="2025-11-22 08:11:27.347348988 +0000 UTC m=+901.581749261" lastFinishedPulling="2025-11-22 08:11:34.033328403 +0000 UTC m=+908.267728676" observedRunningTime="2025-11-22 08:11:38.058977436 +0000 UTC m=+912.293377779" watchObservedRunningTime="2025-11-22 08:11:38.06394303 +0000 UTC m=+912.298343323" Nov 22 08:11:42 crc kubenswrapper[4789]: I1122 08:11:42.174817 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-qwjlg" Nov 22 08:11:42 crc kubenswrapper[4789]: I1122 08:11:42.223556 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-qwjlg" Nov 22 08:11:47 crc kubenswrapper[4789]: I1122 08:11:47.178180 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-qwjlg" Nov 22 08:11:47 crc kubenswrapper[4789]: I1122 08:11:47.204790 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-6998585d5-4shz4" Nov 22 08:11:47 crc kubenswrapper[4789]: I1122 08:11:47.890018 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-6c7b4b5f48-wm2pz" Nov 22 08:11:48 crc kubenswrapper[4789]: I1122 08:11:48.789825 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-6tqgm" Nov 22 08:11:51 crc kubenswrapper[4789]: I1122 08:11:51.728551 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-cr8n8"] Nov 22 08:11:51 crc kubenswrapper[4789]: I1122 08:11:51.729623 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-cr8n8" Nov 22 08:11:51 crc kubenswrapper[4789]: I1122 08:11:51.733120 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-xk88d" Nov 22 08:11:51 crc kubenswrapper[4789]: I1122 08:11:51.733523 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Nov 22 08:11:51 crc kubenswrapper[4789]: I1122 08:11:51.733768 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Nov 22 08:11:51 crc kubenswrapper[4789]: I1122 08:11:51.753269 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-cr8n8"] Nov 22 08:11:51 crc kubenswrapper[4789]: I1122 08:11:51.907533 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zsf7g\" (UniqueName: \"kubernetes.io/projected/f3c4c619-d25e-4379-aecd-7d855f21f7cc-kube-api-access-zsf7g\") pod \"openstack-operator-index-cr8n8\" (UID: \"f3c4c619-d25e-4379-aecd-7d855f21f7cc\") " pod="openstack-operators/openstack-operator-index-cr8n8" Nov 22 08:11:52 crc kubenswrapper[4789]: I1122 08:11:52.008738 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zsf7g\" (UniqueName: \"kubernetes.io/projected/f3c4c619-d25e-4379-aecd-7d855f21f7cc-kube-api-access-zsf7g\") pod \"openstack-operator-index-cr8n8\" (UID: \"f3c4c619-d25e-4379-aecd-7d855f21f7cc\") " pod="openstack-operators/openstack-operator-index-cr8n8" Nov 22 08:11:52 crc kubenswrapper[4789]: I1122 08:11:52.025586 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zsf7g\" (UniqueName: \"kubernetes.io/projected/f3c4c619-d25e-4379-aecd-7d855f21f7cc-kube-api-access-zsf7g\") pod \"openstack-operator-index-cr8n8\" (UID: \"f3c4c619-d25e-4379-aecd-7d855f21f7cc\") " pod="openstack-operators/openstack-operator-index-cr8n8" Nov 22 08:11:52 crc kubenswrapper[4789]: I1122 08:11:52.061133 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-cr8n8" Nov 22 08:11:52 crc kubenswrapper[4789]: I1122 08:11:52.443633 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-cr8n8"] Nov 22 08:11:53 crc kubenswrapper[4789]: I1122 08:11:53.133803 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-cr8n8" event={"ID":"f3c4c619-d25e-4379-aecd-7d855f21f7cc","Type":"ContainerStarted","Data":"3df94d2deedcbcefdea0baa8a8d6270a62294a2ffdcc41097793b458b59a9c5f"} Nov 22 08:11:54 crc kubenswrapper[4789]: I1122 08:11:54.140293 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-cr8n8" event={"ID":"f3c4c619-d25e-4379-aecd-7d855f21f7cc","Type":"ContainerStarted","Data":"f60729961bc390d6bf2767051b5bd9c80fd21a4d2acf21729a845d58e77ca6a3"} Nov 22 08:11:55 crc kubenswrapper[4789]: I1122 08:11:55.304276 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-cr8n8" podStartSLOduration=3.620154492 podStartE2EDuration="4.304258019s" podCreationTimestamp="2025-11-22 08:11:51 +0000 UTC" firstStartedPulling="2025-11-22 08:11:52.450970803 +0000 UTC m=+926.685371076" lastFinishedPulling="2025-11-22 08:11:53.13507433 +0000 UTC m=+927.369474603" observedRunningTime="2025-11-22 08:11:54.159033002 +0000 UTC m=+928.393433295" watchObservedRunningTime="2025-11-22 08:11:55.304258019 +0000 UTC m=+929.538658292" Nov 22 08:11:55 crc kubenswrapper[4789]: I1122 08:11:55.309076 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-cr8n8"] Nov 22 08:11:55 crc kubenswrapper[4789]: I1122 08:11:55.920595 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-9z5kz"] Nov 22 08:11:55 crc kubenswrapper[4789]: I1122 08:11:55.922697 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-9z5kz" Nov 22 08:11:55 crc kubenswrapper[4789]: I1122 08:11:55.936549 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-9z5kz"] Nov 22 08:11:56 crc kubenswrapper[4789]: I1122 08:11:56.063529 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mpffd\" (UniqueName: \"kubernetes.io/projected/20a926be-0d41-41ef-bb6c-70e2d19ec270-kube-api-access-mpffd\") pod \"openstack-operator-index-9z5kz\" (UID: \"20a926be-0d41-41ef-bb6c-70e2d19ec270\") " pod="openstack-operators/openstack-operator-index-9z5kz" Nov 22 08:11:56 crc kubenswrapper[4789]: I1122 08:11:56.152637 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-cr8n8" podUID="f3c4c619-d25e-4379-aecd-7d855f21f7cc" containerName="registry-server" containerID="cri-o://f60729961bc390d6bf2767051b5bd9c80fd21a4d2acf21729a845d58e77ca6a3" gracePeriod=2 Nov 22 08:11:56 crc kubenswrapper[4789]: I1122 08:11:56.164245 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mpffd\" (UniqueName: \"kubernetes.io/projected/20a926be-0d41-41ef-bb6c-70e2d19ec270-kube-api-access-mpffd\") pod \"openstack-operator-index-9z5kz\" (UID: \"20a926be-0d41-41ef-bb6c-70e2d19ec270\") " pod="openstack-operators/openstack-operator-index-9z5kz" Nov 22 08:11:56 crc kubenswrapper[4789]: I1122 08:11:56.188108 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mpffd\" (UniqueName: \"kubernetes.io/projected/20a926be-0d41-41ef-bb6c-70e2d19ec270-kube-api-access-mpffd\") pod \"openstack-operator-index-9z5kz\" (UID: \"20a926be-0d41-41ef-bb6c-70e2d19ec270\") " pod="openstack-operators/openstack-operator-index-9z5kz" Nov 22 08:11:56 crc kubenswrapper[4789]: I1122 08:11:56.254096 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-9z5kz" Nov 22 08:11:56 crc kubenswrapper[4789]: I1122 08:11:56.508875 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-cr8n8" Nov 22 08:11:56 crc kubenswrapper[4789]: I1122 08:11:56.570964 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zsf7g\" (UniqueName: \"kubernetes.io/projected/f3c4c619-d25e-4379-aecd-7d855f21f7cc-kube-api-access-zsf7g\") pod \"f3c4c619-d25e-4379-aecd-7d855f21f7cc\" (UID: \"f3c4c619-d25e-4379-aecd-7d855f21f7cc\") " Nov 22 08:11:56 crc kubenswrapper[4789]: I1122 08:11:56.574913 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f3c4c619-d25e-4379-aecd-7d855f21f7cc-kube-api-access-zsf7g" (OuterVolumeSpecName: "kube-api-access-zsf7g") pod "f3c4c619-d25e-4379-aecd-7d855f21f7cc" (UID: "f3c4c619-d25e-4379-aecd-7d855f21f7cc"). InnerVolumeSpecName "kube-api-access-zsf7g". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:11:56 crc kubenswrapper[4789]: I1122 08:11:56.666409 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-9z5kz"] Nov 22 08:11:56 crc kubenswrapper[4789]: I1122 08:11:56.673108 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zsf7g\" (UniqueName: \"kubernetes.io/projected/f3c4c619-d25e-4379-aecd-7d855f21f7cc-kube-api-access-zsf7g\") on node \"crc\" DevicePath \"\"" Nov 22 08:11:56 crc kubenswrapper[4789]: W1122 08:11:56.673279 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod20a926be_0d41_41ef_bb6c_70e2d19ec270.slice/crio-7bbd82af40df12af26a73cf89bd342f26c4a987f04aa020293a8ce7ced5a8e3e WatchSource:0}: Error finding container 7bbd82af40df12af26a73cf89bd342f26c4a987f04aa020293a8ce7ced5a8e3e: Status 404 returned error can't find the container with id 7bbd82af40df12af26a73cf89bd342f26c4a987f04aa020293a8ce7ced5a8e3e Nov 22 08:11:57 crc kubenswrapper[4789]: I1122 08:11:57.158398 4789 generic.go:334] "Generic (PLEG): container finished" podID="f3c4c619-d25e-4379-aecd-7d855f21f7cc" containerID="f60729961bc390d6bf2767051b5bd9c80fd21a4d2acf21729a845d58e77ca6a3" exitCode=0 Nov 22 08:11:57 crc kubenswrapper[4789]: I1122 08:11:57.158446 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-cr8n8" Nov 22 08:11:57 crc kubenswrapper[4789]: I1122 08:11:57.158441 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-cr8n8" event={"ID":"f3c4c619-d25e-4379-aecd-7d855f21f7cc","Type":"ContainerDied","Data":"f60729961bc390d6bf2767051b5bd9c80fd21a4d2acf21729a845d58e77ca6a3"} Nov 22 08:11:57 crc kubenswrapper[4789]: I1122 08:11:57.159018 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-cr8n8" event={"ID":"f3c4c619-d25e-4379-aecd-7d855f21f7cc","Type":"ContainerDied","Data":"3df94d2deedcbcefdea0baa8a8d6270a62294a2ffdcc41097793b458b59a9c5f"} Nov 22 08:11:57 crc kubenswrapper[4789]: I1122 08:11:57.159069 4789 scope.go:117] "RemoveContainer" containerID="f60729961bc390d6bf2767051b5bd9c80fd21a4d2acf21729a845d58e77ca6a3" Nov 22 08:11:57 crc kubenswrapper[4789]: I1122 08:11:57.160657 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-9z5kz" event={"ID":"20a926be-0d41-41ef-bb6c-70e2d19ec270","Type":"ContainerStarted","Data":"7bbd82af40df12af26a73cf89bd342f26c4a987f04aa020293a8ce7ced5a8e3e"} Nov 22 08:11:57 crc kubenswrapper[4789]: I1122 08:11:57.175383 4789 scope.go:117] "RemoveContainer" containerID="f60729961bc390d6bf2767051b5bd9c80fd21a4d2acf21729a845d58e77ca6a3" Nov 22 08:11:57 crc kubenswrapper[4789]: E1122 08:11:57.175845 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f60729961bc390d6bf2767051b5bd9c80fd21a4d2acf21729a845d58e77ca6a3\": container with ID starting with f60729961bc390d6bf2767051b5bd9c80fd21a4d2acf21729a845d58e77ca6a3 not found: ID does not exist" containerID="f60729961bc390d6bf2767051b5bd9c80fd21a4d2acf21729a845d58e77ca6a3" Nov 22 08:11:57 crc kubenswrapper[4789]: I1122 08:11:57.175880 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f60729961bc390d6bf2767051b5bd9c80fd21a4d2acf21729a845d58e77ca6a3"} err="failed to get container status \"f60729961bc390d6bf2767051b5bd9c80fd21a4d2acf21729a845d58e77ca6a3\": rpc error: code = NotFound desc = could not find container \"f60729961bc390d6bf2767051b5bd9c80fd21a4d2acf21729a845d58e77ca6a3\": container with ID starting with f60729961bc390d6bf2767051b5bd9c80fd21a4d2acf21729a845d58e77ca6a3 not found: ID does not exist" Nov 22 08:11:57 crc kubenswrapper[4789]: I1122 08:11:57.212792 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-cr8n8"] Nov 22 08:11:57 crc kubenswrapper[4789]: I1122 08:11:57.216161 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-cr8n8"] Nov 22 08:11:57 crc kubenswrapper[4789]: I1122 08:11:57.971907 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f3c4c619-d25e-4379-aecd-7d855f21f7cc" path="/var/lib/kubelet/pods/f3c4c619-d25e-4379-aecd-7d855f21f7cc/volumes" Nov 22 08:11:58 crc kubenswrapper[4789]: I1122 08:11:58.168715 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-9z5kz" event={"ID":"20a926be-0d41-41ef-bb6c-70e2d19ec270","Type":"ContainerStarted","Data":"64049cc64ac52124b012c98249fa3a5f169bd55acba1608ff05631c471bbe524"} Nov 22 08:12:05 crc kubenswrapper[4789]: I1122 08:12:05.373065 4789 patch_prober.go:28] interesting pod/machine-config-daemon-pxvrm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 08:12:05 crc kubenswrapper[4789]: I1122 08:12:05.373674 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 08:12:06 crc kubenswrapper[4789]: I1122 08:12:06.254856 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-9z5kz" Nov 22 08:12:06 crc kubenswrapper[4789]: I1122 08:12:06.255079 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-9z5kz" Nov 22 08:12:06 crc kubenswrapper[4789]: I1122 08:12:06.315156 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-9z5kz" Nov 22 08:12:06 crc kubenswrapper[4789]: I1122 08:12:06.335477 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-9z5kz" podStartSLOduration=10.898838906 podStartE2EDuration="11.33546018s" podCreationTimestamp="2025-11-22 08:11:55 +0000 UTC" firstStartedPulling="2025-11-22 08:11:56.678460598 +0000 UTC m=+930.912860871" lastFinishedPulling="2025-11-22 08:11:57.115081872 +0000 UTC m=+931.349482145" observedRunningTime="2025-11-22 08:11:58.183646116 +0000 UTC m=+932.418046399" watchObservedRunningTime="2025-11-22 08:12:06.33546018 +0000 UTC m=+940.569860453" Nov 22 08:12:07 crc kubenswrapper[4789]: I1122 08:12:07.241725 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-9z5kz" Nov 22 08:12:12 crc kubenswrapper[4789]: I1122 08:12:12.936296 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287wk9d5"] Nov 22 08:12:12 crc kubenswrapper[4789]: E1122 08:12:12.937294 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3c4c619-d25e-4379-aecd-7d855f21f7cc" containerName="registry-server" Nov 22 08:12:12 crc kubenswrapper[4789]: I1122 08:12:12.937309 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3c4c619-d25e-4379-aecd-7d855f21f7cc" containerName="registry-server" Nov 22 08:12:12 crc kubenswrapper[4789]: I1122 08:12:12.937620 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3c4c619-d25e-4379-aecd-7d855f21f7cc" containerName="registry-server" Nov 22 08:12:12 crc kubenswrapper[4789]: I1122 08:12:12.939314 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287wk9d5" Nov 22 08:12:12 crc kubenswrapper[4789]: I1122 08:12:12.941980 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-bllzg" Nov 22 08:12:12 crc kubenswrapper[4789]: I1122 08:12:12.944159 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287wk9d5"] Nov 22 08:12:13 crc kubenswrapper[4789]: I1122 08:12:13.085873 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/aeb88caf-4a0d-450a-9646-36471be5fd49-bundle\") pod \"1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287wk9d5\" (UID: \"aeb88caf-4a0d-450a-9646-36471be5fd49\") " pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287wk9d5" Nov 22 08:12:13 crc kubenswrapper[4789]: I1122 08:12:13.086011 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/aeb88caf-4a0d-450a-9646-36471be5fd49-util\") pod \"1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287wk9d5\" (UID: \"aeb88caf-4a0d-450a-9646-36471be5fd49\") " pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287wk9d5" Nov 22 08:12:13 crc kubenswrapper[4789]: I1122 08:12:13.086091 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zm56q\" (UniqueName: \"kubernetes.io/projected/aeb88caf-4a0d-450a-9646-36471be5fd49-kube-api-access-zm56q\") pod \"1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287wk9d5\" (UID: \"aeb88caf-4a0d-450a-9646-36471be5fd49\") " pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287wk9d5" Nov 22 08:12:13 crc kubenswrapper[4789]: I1122 08:12:13.187821 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/aeb88caf-4a0d-450a-9646-36471be5fd49-util\") pod \"1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287wk9d5\" (UID: \"aeb88caf-4a0d-450a-9646-36471be5fd49\") " pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287wk9d5" Nov 22 08:12:13 crc kubenswrapper[4789]: I1122 08:12:13.187893 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zm56q\" (UniqueName: \"kubernetes.io/projected/aeb88caf-4a0d-450a-9646-36471be5fd49-kube-api-access-zm56q\") pod \"1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287wk9d5\" (UID: \"aeb88caf-4a0d-450a-9646-36471be5fd49\") " pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287wk9d5" Nov 22 08:12:13 crc kubenswrapper[4789]: I1122 08:12:13.188031 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/aeb88caf-4a0d-450a-9646-36471be5fd49-bundle\") pod \"1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287wk9d5\" (UID: \"aeb88caf-4a0d-450a-9646-36471be5fd49\") " pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287wk9d5" Nov 22 08:12:13 crc kubenswrapper[4789]: I1122 08:12:13.188631 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/aeb88caf-4a0d-450a-9646-36471be5fd49-bundle\") pod \"1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287wk9d5\" (UID: \"aeb88caf-4a0d-450a-9646-36471be5fd49\") " pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287wk9d5" Nov 22 08:12:13 crc kubenswrapper[4789]: I1122 08:12:13.189035 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/aeb88caf-4a0d-450a-9646-36471be5fd49-util\") pod \"1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287wk9d5\" (UID: \"aeb88caf-4a0d-450a-9646-36471be5fd49\") " pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287wk9d5" Nov 22 08:12:13 crc kubenswrapper[4789]: I1122 08:12:13.217894 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zm56q\" (UniqueName: \"kubernetes.io/projected/aeb88caf-4a0d-450a-9646-36471be5fd49-kube-api-access-zm56q\") pod \"1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287wk9d5\" (UID: \"aeb88caf-4a0d-450a-9646-36471be5fd49\") " pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287wk9d5" Nov 22 08:12:13 crc kubenswrapper[4789]: I1122 08:12:13.266576 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287wk9d5" Nov 22 08:12:13 crc kubenswrapper[4789]: I1122 08:12:13.691512 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287wk9d5"] Nov 22 08:12:13 crc kubenswrapper[4789]: W1122 08:12:13.702647 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaeb88caf_4a0d_450a_9646_36471be5fd49.slice/crio-caa482ce4d0ba238b91a61f5efb3bd2805653ca0d23b9fb76b35ac1f3afd42e1 WatchSource:0}: Error finding container caa482ce4d0ba238b91a61f5efb3bd2805653ca0d23b9fb76b35ac1f3afd42e1: Status 404 returned error can't find the container with id caa482ce4d0ba238b91a61f5efb3bd2805653ca0d23b9fb76b35ac1f3afd42e1 Nov 22 08:12:14 crc kubenswrapper[4789]: I1122 08:12:14.263367 4789 generic.go:334] "Generic (PLEG): container finished" podID="aeb88caf-4a0d-450a-9646-36471be5fd49" containerID="ce837d0c3899ef4ccd98b6bf4d6e4ec77ee3f9fe01143492f5d8e79b9495764e" exitCode=0 Nov 22 08:12:14 crc kubenswrapper[4789]: I1122 08:12:14.263523 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287wk9d5" event={"ID":"aeb88caf-4a0d-450a-9646-36471be5fd49","Type":"ContainerDied","Data":"ce837d0c3899ef4ccd98b6bf4d6e4ec77ee3f9fe01143492f5d8e79b9495764e"} Nov 22 08:12:14 crc kubenswrapper[4789]: I1122 08:12:14.263912 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287wk9d5" event={"ID":"aeb88caf-4a0d-450a-9646-36471be5fd49","Type":"ContainerStarted","Data":"caa482ce4d0ba238b91a61f5efb3bd2805653ca0d23b9fb76b35ac1f3afd42e1"} Nov 22 08:12:15 crc kubenswrapper[4789]: I1122 08:12:15.272477 4789 generic.go:334] "Generic (PLEG): container finished" podID="aeb88caf-4a0d-450a-9646-36471be5fd49" containerID="fc38c239d21e631f8b47e45f29a49595da5561efc3c412935662efb63880852e" exitCode=0 Nov 22 08:12:15 crc kubenswrapper[4789]: I1122 08:12:15.272531 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287wk9d5" event={"ID":"aeb88caf-4a0d-450a-9646-36471be5fd49","Type":"ContainerDied","Data":"fc38c239d21e631f8b47e45f29a49595da5561efc3c412935662efb63880852e"} Nov 22 08:12:16 crc kubenswrapper[4789]: I1122 08:12:16.284634 4789 generic.go:334] "Generic (PLEG): container finished" podID="aeb88caf-4a0d-450a-9646-36471be5fd49" containerID="6964dfacf35ed0eb196326aa644234bc4956520556226caf4a7eb6a9bc5af8dc" exitCode=0 Nov 22 08:12:16 crc kubenswrapper[4789]: I1122 08:12:16.284706 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287wk9d5" event={"ID":"aeb88caf-4a0d-450a-9646-36471be5fd49","Type":"ContainerDied","Data":"6964dfacf35ed0eb196326aa644234bc4956520556226caf4a7eb6a9bc5af8dc"} Nov 22 08:12:17 crc kubenswrapper[4789]: I1122 08:12:17.529618 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287wk9d5" Nov 22 08:12:17 crc kubenswrapper[4789]: I1122 08:12:17.652001 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/aeb88caf-4a0d-450a-9646-36471be5fd49-bundle\") pod \"aeb88caf-4a0d-450a-9646-36471be5fd49\" (UID: \"aeb88caf-4a0d-450a-9646-36471be5fd49\") " Nov 22 08:12:17 crc kubenswrapper[4789]: I1122 08:12:17.652078 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/aeb88caf-4a0d-450a-9646-36471be5fd49-util\") pod \"aeb88caf-4a0d-450a-9646-36471be5fd49\" (UID: \"aeb88caf-4a0d-450a-9646-36471be5fd49\") " Nov 22 08:12:17 crc kubenswrapper[4789]: I1122 08:12:17.652150 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zm56q\" (UniqueName: \"kubernetes.io/projected/aeb88caf-4a0d-450a-9646-36471be5fd49-kube-api-access-zm56q\") pod \"aeb88caf-4a0d-450a-9646-36471be5fd49\" (UID: \"aeb88caf-4a0d-450a-9646-36471be5fd49\") " Nov 22 08:12:17 crc kubenswrapper[4789]: I1122 08:12:17.653588 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aeb88caf-4a0d-450a-9646-36471be5fd49-bundle" (OuterVolumeSpecName: "bundle") pod "aeb88caf-4a0d-450a-9646-36471be5fd49" (UID: "aeb88caf-4a0d-450a-9646-36471be5fd49"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:12:17 crc kubenswrapper[4789]: I1122 08:12:17.657660 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aeb88caf-4a0d-450a-9646-36471be5fd49-kube-api-access-zm56q" (OuterVolumeSpecName: "kube-api-access-zm56q") pod "aeb88caf-4a0d-450a-9646-36471be5fd49" (UID: "aeb88caf-4a0d-450a-9646-36471be5fd49"). InnerVolumeSpecName "kube-api-access-zm56q". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:12:17 crc kubenswrapper[4789]: I1122 08:12:17.686398 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aeb88caf-4a0d-450a-9646-36471be5fd49-util" (OuterVolumeSpecName: "util") pod "aeb88caf-4a0d-450a-9646-36471be5fd49" (UID: "aeb88caf-4a0d-450a-9646-36471be5fd49"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:12:17 crc kubenswrapper[4789]: I1122 08:12:17.753978 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zm56q\" (UniqueName: \"kubernetes.io/projected/aeb88caf-4a0d-450a-9646-36471be5fd49-kube-api-access-zm56q\") on node \"crc\" DevicePath \"\"" Nov 22 08:12:17 crc kubenswrapper[4789]: I1122 08:12:17.754058 4789 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/aeb88caf-4a0d-450a-9646-36471be5fd49-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:12:17 crc kubenswrapper[4789]: I1122 08:12:17.754093 4789 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/aeb88caf-4a0d-450a-9646-36471be5fd49-util\") on node \"crc\" DevicePath \"\"" Nov 22 08:12:18 crc kubenswrapper[4789]: I1122 08:12:18.299947 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287wk9d5" event={"ID":"aeb88caf-4a0d-450a-9646-36471be5fd49","Type":"ContainerDied","Data":"caa482ce4d0ba238b91a61f5efb3bd2805653ca0d23b9fb76b35ac1f3afd42e1"} Nov 22 08:12:18 crc kubenswrapper[4789]: I1122 08:12:18.299994 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="caa482ce4d0ba238b91a61f5efb3bd2805653ca0d23b9fb76b35ac1f3afd42e1" Nov 22 08:12:18 crc kubenswrapper[4789]: I1122 08:12:18.300311 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287wk9d5" Nov 22 08:12:25 crc kubenswrapper[4789]: I1122 08:12:25.373334 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-8486c7f98b-7vhvn"] Nov 22 08:12:25 crc kubenswrapper[4789]: E1122 08:12:25.374542 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aeb88caf-4a0d-450a-9646-36471be5fd49" containerName="pull" Nov 22 08:12:25 crc kubenswrapper[4789]: I1122 08:12:25.374558 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="aeb88caf-4a0d-450a-9646-36471be5fd49" containerName="pull" Nov 22 08:12:25 crc kubenswrapper[4789]: E1122 08:12:25.374571 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aeb88caf-4a0d-450a-9646-36471be5fd49" containerName="util" Nov 22 08:12:25 crc kubenswrapper[4789]: I1122 08:12:25.374578 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="aeb88caf-4a0d-450a-9646-36471be5fd49" containerName="util" Nov 22 08:12:25 crc kubenswrapper[4789]: E1122 08:12:25.374607 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aeb88caf-4a0d-450a-9646-36471be5fd49" containerName="extract" Nov 22 08:12:25 crc kubenswrapper[4789]: I1122 08:12:25.374613 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="aeb88caf-4a0d-450a-9646-36471be5fd49" containerName="extract" Nov 22 08:12:25 crc kubenswrapper[4789]: I1122 08:12:25.374723 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="aeb88caf-4a0d-450a-9646-36471be5fd49" containerName="extract" Nov 22 08:12:25 crc kubenswrapper[4789]: I1122 08:12:25.375527 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-8486c7f98b-7vhvn" Nov 22 08:12:25 crc kubenswrapper[4789]: I1122 08:12:25.378274 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-hmn2f" Nov 22 08:12:25 crc kubenswrapper[4789]: I1122 08:12:25.409459 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-8486c7f98b-7vhvn"] Nov 22 08:12:25 crc kubenswrapper[4789]: I1122 08:12:25.458382 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x2dqm\" (UniqueName: \"kubernetes.io/projected/7c594182-6941-4a46-aebc-0d9df403eec6-kube-api-access-x2dqm\") pod \"openstack-operator-controller-operator-8486c7f98b-7vhvn\" (UID: \"7c594182-6941-4a46-aebc-0d9df403eec6\") " pod="openstack-operators/openstack-operator-controller-operator-8486c7f98b-7vhvn" Nov 22 08:12:25 crc kubenswrapper[4789]: I1122 08:12:25.559404 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x2dqm\" (UniqueName: \"kubernetes.io/projected/7c594182-6941-4a46-aebc-0d9df403eec6-kube-api-access-x2dqm\") pod \"openstack-operator-controller-operator-8486c7f98b-7vhvn\" (UID: \"7c594182-6941-4a46-aebc-0d9df403eec6\") " pod="openstack-operators/openstack-operator-controller-operator-8486c7f98b-7vhvn" Nov 22 08:12:25 crc kubenswrapper[4789]: I1122 08:12:25.584865 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x2dqm\" (UniqueName: \"kubernetes.io/projected/7c594182-6941-4a46-aebc-0d9df403eec6-kube-api-access-x2dqm\") pod \"openstack-operator-controller-operator-8486c7f98b-7vhvn\" (UID: \"7c594182-6941-4a46-aebc-0d9df403eec6\") " pod="openstack-operators/openstack-operator-controller-operator-8486c7f98b-7vhvn" Nov 22 08:12:25 crc kubenswrapper[4789]: I1122 08:12:25.699184 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-8486c7f98b-7vhvn" Nov 22 08:12:26 crc kubenswrapper[4789]: I1122 08:12:26.112420 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-8486c7f98b-7vhvn"] Nov 22 08:12:26 crc kubenswrapper[4789]: I1122 08:12:26.346595 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-8486c7f98b-7vhvn" event={"ID":"7c594182-6941-4a46-aebc-0d9df403eec6","Type":"ContainerStarted","Data":"68124dd5ee4b58ee852f7e7890efdf3e3ba3a6e019ae7ad3dc9d4b46dbfc14b2"} Nov 22 08:12:31 crc kubenswrapper[4789]: I1122 08:12:31.379595 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-8486c7f98b-7vhvn" event={"ID":"7c594182-6941-4a46-aebc-0d9df403eec6","Type":"ContainerStarted","Data":"e3aefedff2f08c827ef8f118d9bba35b44fc182a50185b3711cc452702de7628"} Nov 22 08:12:34 crc kubenswrapper[4789]: I1122 08:12:34.399718 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-8486c7f98b-7vhvn" event={"ID":"7c594182-6941-4a46-aebc-0d9df403eec6","Type":"ContainerStarted","Data":"47f0e289a486d69c522c755790ec9b35b33b2f34d5578e503a7442c5aa0ef0cf"} Nov 22 08:12:34 crc kubenswrapper[4789]: I1122 08:12:34.401027 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-8486c7f98b-7vhvn" Nov 22 08:12:34 crc kubenswrapper[4789]: I1122 08:12:34.432149 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-8486c7f98b-7vhvn" podStartSLOduration=1.8937479499999998 podStartE2EDuration="9.432128637s" podCreationTimestamp="2025-11-22 08:12:25 +0000 UTC" firstStartedPulling="2025-11-22 08:12:26.12015044 +0000 UTC m=+960.354550713" lastFinishedPulling="2025-11-22 08:12:33.658531127 +0000 UTC m=+967.892931400" observedRunningTime="2025-11-22 08:12:34.427655966 +0000 UTC m=+968.662056239" watchObservedRunningTime="2025-11-22 08:12:34.432128637 +0000 UTC m=+968.666528900" Nov 22 08:12:35 crc kubenswrapper[4789]: I1122 08:12:35.372296 4789 patch_prober.go:28] interesting pod/machine-config-daemon-pxvrm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 08:12:35 crc kubenswrapper[4789]: I1122 08:12:35.372609 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 08:12:36 crc kubenswrapper[4789]: I1122 08:12:36.424026 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-8486c7f98b-7vhvn" Nov 22 08:13:05 crc kubenswrapper[4789]: I1122 08:13:05.623808 4789 patch_prober.go:28] interesting pod/machine-config-daemon-pxvrm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 08:13:05 crc kubenswrapper[4789]: I1122 08:13:05.624607 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 08:13:05 crc kubenswrapper[4789]: I1122 08:13:05.624659 4789 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" Nov 22 08:13:05 crc kubenswrapper[4789]: I1122 08:13:05.625289 4789 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5067bbca7c5576cc068f8534a87953cf9b5fc5cb8f8cfddaaed99f53192975c1"} pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 22 08:13:05 crc kubenswrapper[4789]: I1122 08:13:05.625346 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" containerID="cri-o://5067bbca7c5576cc068f8534a87953cf9b5fc5cb8f8cfddaaed99f53192975c1" gracePeriod=600 Nov 22 08:13:06 crc kubenswrapper[4789]: I1122 08:13:06.637827 4789 generic.go:334] "Generic (PLEG): container finished" podID="58aa8071-7ecc-4692-8789-2db04ec70647" containerID="5067bbca7c5576cc068f8534a87953cf9b5fc5cb8f8cfddaaed99f53192975c1" exitCode=0 Nov 22 08:13:06 crc kubenswrapper[4789]: I1122 08:13:06.637879 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" event={"ID":"58aa8071-7ecc-4692-8789-2db04ec70647","Type":"ContainerDied","Data":"5067bbca7c5576cc068f8534a87953cf9b5fc5cb8f8cfddaaed99f53192975c1"} Nov 22 08:13:06 crc kubenswrapper[4789]: I1122 08:13:06.638553 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" event={"ID":"58aa8071-7ecc-4692-8789-2db04ec70647","Type":"ContainerStarted","Data":"fe65d4df369fc1d7c708e571722629956fbd341c1d6b8fc8c7f64c8dfbfa797f"} Nov 22 08:13:06 crc kubenswrapper[4789]: I1122 08:13:06.638581 4789 scope.go:117] "RemoveContainer" containerID="447a15a0714201a673fc1d0cc8a8f355942fbc82479335216d2bf7c26b973475" Nov 22 08:13:13 crc kubenswrapper[4789]: I1122 08:13:13.787491 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7768f8c84f-hwbvw"] Nov 22 08:13:13 crc kubenswrapper[4789]: I1122 08:13:13.788948 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7768f8c84f-hwbvw" Nov 22 08:13:13 crc kubenswrapper[4789]: I1122 08:13:13.793988 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-mvrsc" Nov 22 08:13:13 crc kubenswrapper[4789]: I1122 08:13:13.797883 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6d8fd67bf7-2klns"] Nov 22 08:13:13 crc kubenswrapper[4789]: I1122 08:13:13.799173 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-6d8fd67bf7-2klns" Nov 22 08:13:13 crc kubenswrapper[4789]: I1122 08:13:13.801221 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-9knpb" Nov 22 08:13:13 crc kubenswrapper[4789]: I1122 08:13:13.801224 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7768f8c84f-hwbvw"] Nov 22 08:13:13 crc kubenswrapper[4789]: I1122 08:13:13.812365 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-56dfb6b67f-p4vmr"] Nov 22 08:13:13 crc kubenswrapper[4789]: I1122 08:13:13.813422 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-56dfb6b67f-p4vmr" Nov 22 08:13:13 crc kubenswrapper[4789]: I1122 08:13:13.817185 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6d8fd67bf7-2klns"] Nov 22 08:13:13 crc kubenswrapper[4789]: I1122 08:13:13.817204 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-bdvg4" Nov 22 08:13:13 crc kubenswrapper[4789]: I1122 08:13:13.832670 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-8667fbf6f6-k5kms"] Nov 22 08:13:13 crc kubenswrapper[4789]: I1122 08:13:13.833605 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-8667fbf6f6-k5kms" Nov 22 08:13:13 crc kubenswrapper[4789]: I1122 08:13:13.835268 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-chjpz" Nov 22 08:13:13 crc kubenswrapper[4789]: I1122 08:13:13.842694 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-56dfb6b67f-p4vmr"] Nov 22 08:13:13 crc kubenswrapper[4789]: I1122 08:13:13.852778 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-bf4c6585d-6shgh"] Nov 22 08:13:13 crc kubenswrapper[4789]: I1122 08:13:13.853794 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-bf4c6585d-6shgh" Nov 22 08:13:13 crc kubenswrapper[4789]: I1122 08:13:13.855948 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-dfvlj" Nov 22 08:13:13 crc kubenswrapper[4789]: I1122 08:13:13.866810 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5d86b44686-nk4zk"] Nov 22 08:13:13 crc kubenswrapper[4789]: I1122 08:13:13.867789 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-5d86b44686-nk4zk" Nov 22 08:13:13 crc kubenswrapper[4789]: I1122 08:13:13.870219 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-2hbhp" Nov 22 08:13:13 crc kubenswrapper[4789]: I1122 08:13:13.874993 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-bf4c6585d-6shgh"] Nov 22 08:13:13 crc kubenswrapper[4789]: I1122 08:13:13.890368 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tz9n8\" (UniqueName: \"kubernetes.io/projected/93c2cbf6-757a-4de1-9f83-115787d74f31-kube-api-access-tz9n8\") pod \"barbican-operator-controller-manager-7768f8c84f-hwbvw\" (UID: \"93c2cbf6-757a-4de1-9f83-115787d74f31\") " pod="openstack-operators/barbican-operator-controller-manager-7768f8c84f-hwbvw" Nov 22 08:13:13 crc kubenswrapper[4789]: I1122 08:13:13.890421 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v42n6\" (UniqueName: \"kubernetes.io/projected/55d6892d-7d96-4518-8d36-58048cf96c52-kube-api-access-v42n6\") pod \"cinder-operator-controller-manager-6d8fd67bf7-2klns\" (UID: \"55d6892d-7d96-4518-8d36-58048cf96c52\") " pod="openstack-operators/cinder-operator-controller-manager-6d8fd67bf7-2klns" Nov 22 08:13:13 crc kubenswrapper[4789]: I1122 08:13:13.890458 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7x8th\" (UniqueName: \"kubernetes.io/projected/77bd1957-6cd6-418b-9b2c-ac879bbca15e-kube-api-access-7x8th\") pod \"designate-operator-controller-manager-56dfb6b67f-p4vmr\" (UID: \"77bd1957-6cd6-418b-9b2c-ac879bbca15e\") " pod="openstack-operators/designate-operator-controller-manager-56dfb6b67f-p4vmr" Nov 22 08:13:13 crc kubenswrapper[4789]: I1122 08:13:13.890514 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-8667fbf6f6-k5kms"] Nov 22 08:13:13 crc kubenswrapper[4789]: I1122 08:13:13.920410 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5d86b44686-nk4zk"] Nov 22 08:13:13 crc kubenswrapper[4789]: I1122 08:13:13.922553 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5c75d7c94b-66wlz"] Nov 22 08:13:13 crc kubenswrapper[4789]: I1122 08:13:13.923420 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-5c75d7c94b-66wlz" Nov 22 08:13:13 crc kubenswrapper[4789]: I1122 08:13:13.928267 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-p54c2" Nov 22 08:13:13 crc kubenswrapper[4789]: I1122 08:13:13.935265 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-769d9c7585-f7bd8"] Nov 22 08:13:13 crc kubenswrapper[4789]: I1122 08:13:13.936195 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-769d9c7585-f7bd8" Nov 22 08:13:13 crc kubenswrapper[4789]: I1122 08:13:13.939170 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-skzj5" Nov 22 08:13:13 crc kubenswrapper[4789]: I1122 08:13:13.943884 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Nov 22 08:13:13 crc kubenswrapper[4789]: I1122 08:13:13.954698 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5c75d7c94b-66wlz"] Nov 22 08:13:13 crc kubenswrapper[4789]: I1122 08:13:13.960662 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-769d9c7585-f7bd8"] Nov 22 08:13:13 crc kubenswrapper[4789]: I1122 08:13:13.975812 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7879fb76fd-zn5xp"] Nov 22 08:13:13 crc kubenswrapper[4789]: I1122 08:13:13.977015 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7879fb76fd-zn5xp" Nov 22 08:13:13 crc kubenswrapper[4789]: I1122 08:13:13.980661 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-882wv" Nov 22 08:13:13 crc kubenswrapper[4789]: I1122 08:13:13.991342 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tz9n8\" (UniqueName: \"kubernetes.io/projected/93c2cbf6-757a-4de1-9f83-115787d74f31-kube-api-access-tz9n8\") pod \"barbican-operator-controller-manager-7768f8c84f-hwbvw\" (UID: \"93c2cbf6-757a-4de1-9f83-115787d74f31\") " pod="openstack-operators/barbican-operator-controller-manager-7768f8c84f-hwbvw" Nov 22 08:13:13 crc kubenswrapper[4789]: I1122 08:13:13.991407 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v42n6\" (UniqueName: \"kubernetes.io/projected/55d6892d-7d96-4518-8d36-58048cf96c52-kube-api-access-v42n6\") pod \"cinder-operator-controller-manager-6d8fd67bf7-2klns\" (UID: \"55d6892d-7d96-4518-8d36-58048cf96c52\") " pod="openstack-operators/cinder-operator-controller-manager-6d8fd67bf7-2klns" Nov 22 08:13:13 crc kubenswrapper[4789]: I1122 08:13:13.991453 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7x8th\" (UniqueName: \"kubernetes.io/projected/77bd1957-6cd6-418b-9b2c-ac879bbca15e-kube-api-access-7x8th\") pod \"designate-operator-controller-manager-56dfb6b67f-p4vmr\" (UID: \"77bd1957-6cd6-418b-9b2c-ac879bbca15e\") " pod="openstack-operators/designate-operator-controller-manager-56dfb6b67f-p4vmr" Nov 22 08:13:13 crc kubenswrapper[4789]: I1122 08:13:13.991480 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mjwlw\" (UniqueName: \"kubernetes.io/projected/5c6cfa2e-ba14-48da-a43d-6ea46a90bad1-kube-api-access-mjwlw\") pod \"glance-operator-controller-manager-8667fbf6f6-k5kms\" (UID: \"5c6cfa2e-ba14-48da-a43d-6ea46a90bad1\") " pod="openstack-operators/glance-operator-controller-manager-8667fbf6f6-k5kms" Nov 22 08:13:13 crc kubenswrapper[4789]: I1122 08:13:13.991518 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b25cr\" (UniqueName: \"kubernetes.io/projected/9e20512a-a64f-4f47-a7ef-b91f24390aef-kube-api-access-b25cr\") pod \"horizon-operator-controller-manager-5d86b44686-nk4zk\" (UID: \"9e20512a-a64f-4f47-a7ef-b91f24390aef\") " pod="openstack-operators/horizon-operator-controller-manager-5d86b44686-nk4zk" Nov 22 08:13:13 crc kubenswrapper[4789]: I1122 08:13:13.991542 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bc6d5\" (UniqueName: \"kubernetes.io/projected/0cf958d5-222e-4598-a1a7-fe887644cd24-kube-api-access-bc6d5\") pod \"heat-operator-controller-manager-bf4c6585d-6shgh\" (UID: \"0cf958d5-222e-4598-a1a7-fe887644cd24\") " pod="openstack-operators/heat-operator-controller-manager-bf4c6585d-6shgh" Nov 22 08:13:13 crc kubenswrapper[4789]: I1122 08:13:13.991586 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6rfd4\" (UniqueName: \"kubernetes.io/projected/29d2daa3-90b3-4dea-9612-edc4cc0dcc30-kube-api-access-6rfd4\") pod \"ironic-operator-controller-manager-5c75d7c94b-66wlz\" (UID: \"29d2daa3-90b3-4dea-9612-edc4cc0dcc30\") " pod="openstack-operators/ironic-operator-controller-manager-5c75d7c94b-66wlz" Nov 22 08:13:13 crc kubenswrapper[4789]: I1122 08:13:13.993840 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7879fb76fd-zn5xp"] Nov 22 08:13:13 crc kubenswrapper[4789]: I1122 08:13:13.999654 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-7bb88cb858-9g8lk"] Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.004703 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-7bb88cb858-9g8lk" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.007803 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-ppfhs" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.016906 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v42n6\" (UniqueName: \"kubernetes.io/projected/55d6892d-7d96-4518-8d36-58048cf96c52-kube-api-access-v42n6\") pod \"cinder-operator-controller-manager-6d8fd67bf7-2klns\" (UID: \"55d6892d-7d96-4518-8d36-58048cf96c52\") " pod="openstack-operators/cinder-operator-controller-manager-6d8fd67bf7-2klns" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.033192 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-7bb88cb858-9g8lk"] Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.039943 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tz9n8\" (UniqueName: \"kubernetes.io/projected/93c2cbf6-757a-4de1-9f83-115787d74f31-kube-api-access-tz9n8\") pod \"barbican-operator-controller-manager-7768f8c84f-hwbvw\" (UID: \"93c2cbf6-757a-4de1-9f83-115787d74f31\") " pod="openstack-operators/barbican-operator-controller-manager-7768f8c84f-hwbvw" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.040032 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6f8c5b86cb-rrcjv"] Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.069431 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-6f8c5b86cb-rrcjv" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.078363 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-qk77w" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.078415 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7x8th\" (UniqueName: \"kubernetes.io/projected/77bd1957-6cd6-418b-9b2c-ac879bbca15e-kube-api-access-7x8th\") pod \"designate-operator-controller-manager-56dfb6b67f-p4vmr\" (UID: \"77bd1957-6cd6-418b-9b2c-ac879bbca15e\") " pod="openstack-operators/designate-operator-controller-manager-56dfb6b67f-p4vmr" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.083326 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-66b7d6f598-hn78m"] Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.092190 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-66b7d6f598-hn78m" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.103944 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mjwlw\" (UniqueName: \"kubernetes.io/projected/5c6cfa2e-ba14-48da-a43d-6ea46a90bad1-kube-api-access-mjwlw\") pod \"glance-operator-controller-manager-8667fbf6f6-k5kms\" (UID: \"5c6cfa2e-ba14-48da-a43d-6ea46a90bad1\") " pod="openstack-operators/glance-operator-controller-manager-8667fbf6f6-k5kms" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.104027 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rckmp\" (UniqueName: \"kubernetes.io/projected/50571e3f-4e49-47b6-aa7f-2c714e8ef88f-kube-api-access-rckmp\") pod \"manila-operator-controller-manager-7bb88cb858-9g8lk\" (UID: \"50571e3f-4e49-47b6-aa7f-2c714e8ef88f\") " pod="openstack-operators/manila-operator-controller-manager-7bb88cb858-9g8lk" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.104057 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b25cr\" (UniqueName: \"kubernetes.io/projected/9e20512a-a64f-4f47-a7ef-b91f24390aef-kube-api-access-b25cr\") pod \"horizon-operator-controller-manager-5d86b44686-nk4zk\" (UID: \"9e20512a-a64f-4f47-a7ef-b91f24390aef\") " pod="openstack-operators/horizon-operator-controller-manager-5d86b44686-nk4zk" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.104081 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bc6d5\" (UniqueName: \"kubernetes.io/projected/0cf958d5-222e-4598-a1a7-fe887644cd24-kube-api-access-bc6d5\") pod \"heat-operator-controller-manager-bf4c6585d-6shgh\" (UID: \"0cf958d5-222e-4598-a1a7-fe887644cd24\") " pod="openstack-operators/heat-operator-controller-manager-bf4c6585d-6shgh" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.104100 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9hmbk\" (UniqueName: \"kubernetes.io/projected/490be79a-ce0c-4a3e-a819-7ef482fe453d-kube-api-access-9hmbk\") pod \"keystone-operator-controller-manager-7879fb76fd-zn5xp\" (UID: \"490be79a-ce0c-4a3e-a819-7ef482fe453d\") " pod="openstack-operators/keystone-operator-controller-manager-7879fb76fd-zn5xp" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.104164 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6rfd4\" (UniqueName: \"kubernetes.io/projected/29d2daa3-90b3-4dea-9612-edc4cc0dcc30-kube-api-access-6rfd4\") pod \"ironic-operator-controller-manager-5c75d7c94b-66wlz\" (UID: \"29d2daa3-90b3-4dea-9612-edc4cc0dcc30\") " pod="openstack-operators/ironic-operator-controller-manager-5c75d7c94b-66wlz" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.104205 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s8xrr\" (UniqueName: \"kubernetes.io/projected/84fafb83-aaff-4197-90d1-578c74da9299-kube-api-access-s8xrr\") pod \"infra-operator-controller-manager-769d9c7585-f7bd8\" (UID: \"84fafb83-aaff-4197-90d1-578c74da9299\") " pod="openstack-operators/infra-operator-controller-manager-769d9c7585-f7bd8" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.104292 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/84fafb83-aaff-4197-90d1-578c74da9299-cert\") pod \"infra-operator-controller-manager-769d9c7585-f7bd8\" (UID: \"84fafb83-aaff-4197-90d1-578c74da9299\") " pod="openstack-operators/infra-operator-controller-manager-769d9c7585-f7bd8" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.106860 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-h22bn" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.112141 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7768f8c84f-hwbvw" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.123397 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-6d8fd67bf7-2klns" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.130040 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-56dfb6b67f-p4vmr" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.131522 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6f8c5b86cb-rrcjv"] Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.137606 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6rfd4\" (UniqueName: \"kubernetes.io/projected/29d2daa3-90b3-4dea-9612-edc4cc0dcc30-kube-api-access-6rfd4\") pod \"ironic-operator-controller-manager-5c75d7c94b-66wlz\" (UID: \"29d2daa3-90b3-4dea-9612-edc4cc0dcc30\") " pod="openstack-operators/ironic-operator-controller-manager-5c75d7c94b-66wlz" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.143131 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-86d796d84d-4lzpn"] Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.144842 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-86d796d84d-4lzpn" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.151061 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b25cr\" (UniqueName: \"kubernetes.io/projected/9e20512a-a64f-4f47-a7ef-b91f24390aef-kube-api-access-b25cr\") pod \"horizon-operator-controller-manager-5d86b44686-nk4zk\" (UID: \"9e20512a-a64f-4f47-a7ef-b91f24390aef\") " pod="openstack-operators/horizon-operator-controller-manager-5d86b44686-nk4zk" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.151358 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-66b7d6f598-hn78m"] Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.152185 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-pgx59" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.152465 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bc6d5\" (UniqueName: \"kubernetes.io/projected/0cf958d5-222e-4598-a1a7-fe887644cd24-kube-api-access-bc6d5\") pod \"heat-operator-controller-manager-bf4c6585d-6shgh\" (UID: \"0cf958d5-222e-4598-a1a7-fe887644cd24\") " pod="openstack-operators/heat-operator-controller-manager-bf4c6585d-6shgh" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.153405 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mjwlw\" (UniqueName: \"kubernetes.io/projected/5c6cfa2e-ba14-48da-a43d-6ea46a90bad1-kube-api-access-mjwlw\") pod \"glance-operator-controller-manager-8667fbf6f6-k5kms\" (UID: \"5c6cfa2e-ba14-48da-a43d-6ea46a90bad1\") " pod="openstack-operators/glance-operator-controller-manager-8667fbf6f6-k5kms" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.164097 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-8667fbf6f6-k5kms" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.174890 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-6fdc856c5d-mccpp"] Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.183663 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-6fdc856c5d-mccpp" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.184552 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-bf4c6585d-6shgh" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.187290 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-fpkv7" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.197375 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-86d796d84d-4lzpn"] Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.199950 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-5d86b44686-nk4zk" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.205965 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wjlsk\" (UniqueName: \"kubernetes.io/projected/c4a79bb9-1d05-4c14-a668-6f63f521c98d-kube-api-access-wjlsk\") pod \"neutron-operator-controller-manager-66b7d6f598-hn78m\" (UID: \"c4a79bb9-1d05-4c14-a668-6f63f521c98d\") " pod="openstack-operators/neutron-operator-controller-manager-66b7d6f598-hn78m" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.206020 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s8xrr\" (UniqueName: \"kubernetes.io/projected/84fafb83-aaff-4197-90d1-578c74da9299-kube-api-access-s8xrr\") pod \"infra-operator-controller-manager-769d9c7585-f7bd8\" (UID: \"84fafb83-aaff-4197-90d1-578c74da9299\") " pod="openstack-operators/infra-operator-controller-manager-769d9c7585-f7bd8" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.206149 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xn5br\" (UniqueName: \"kubernetes.io/projected/c4962710-54a6-4538-90ce-be3e27062015-kube-api-access-xn5br\") pod \"mariadb-operator-controller-manager-6f8c5b86cb-rrcjv\" (UID: \"c4962710-54a6-4538-90ce-be3e27062015\") " pod="openstack-operators/mariadb-operator-controller-manager-6f8c5b86cb-rrcjv" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.206201 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/84fafb83-aaff-4197-90d1-578c74da9299-cert\") pod \"infra-operator-controller-manager-769d9c7585-f7bd8\" (UID: \"84fafb83-aaff-4197-90d1-578c74da9299\") " pod="openstack-operators/infra-operator-controller-manager-769d9c7585-f7bd8" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.206307 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rckmp\" (UniqueName: \"kubernetes.io/projected/50571e3f-4e49-47b6-aa7f-2c714e8ef88f-kube-api-access-rckmp\") pod \"manila-operator-controller-manager-7bb88cb858-9g8lk\" (UID: \"50571e3f-4e49-47b6-aa7f-2c714e8ef88f\") " pod="openstack-operators/manila-operator-controller-manager-7bb88cb858-9g8lk" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.206351 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9hmbk\" (UniqueName: \"kubernetes.io/projected/490be79a-ce0c-4a3e-a819-7ef482fe453d-kube-api-access-9hmbk\") pod \"keystone-operator-controller-manager-7879fb76fd-zn5xp\" (UID: \"490be79a-ce0c-4a3e-a819-7ef482fe453d\") " pod="openstack-operators/keystone-operator-controller-manager-7879fb76fd-zn5xp" Nov 22 08:13:14 crc kubenswrapper[4789]: E1122 08:13:14.206920 4789 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Nov 22 08:13:14 crc kubenswrapper[4789]: E1122 08:13:14.206976 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/84fafb83-aaff-4197-90d1-578c74da9299-cert podName:84fafb83-aaff-4197-90d1-578c74da9299 nodeName:}" failed. No retries permitted until 2025-11-22 08:13:14.706957854 +0000 UTC m=+1008.941358127 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/84fafb83-aaff-4197-90d1-578c74da9299-cert") pod "infra-operator-controller-manager-769d9c7585-f7bd8" (UID: "84fafb83-aaff-4197-90d1-578c74da9299") : secret "infra-operator-webhook-server-cert" not found Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.229800 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9hmbk\" (UniqueName: \"kubernetes.io/projected/490be79a-ce0c-4a3e-a819-7ef482fe453d-kube-api-access-9hmbk\") pod \"keystone-operator-controller-manager-7879fb76fd-zn5xp\" (UID: \"490be79a-ce0c-4a3e-a819-7ef482fe453d\") " pod="openstack-operators/keystone-operator-controller-manager-7879fb76fd-zn5xp" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.230041 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-6fdc856c5d-mccpp"] Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.235243 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s8xrr\" (UniqueName: \"kubernetes.io/projected/84fafb83-aaff-4197-90d1-578c74da9299-kube-api-access-s8xrr\") pod \"infra-operator-controller-manager-769d9c7585-f7bd8\" (UID: \"84fafb83-aaff-4197-90d1-578c74da9299\") " pod="openstack-operators/infra-operator-controller-manager-769d9c7585-f7bd8" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.236895 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rckmp\" (UniqueName: \"kubernetes.io/projected/50571e3f-4e49-47b6-aa7f-2c714e8ef88f-kube-api-access-rckmp\") pod \"manila-operator-controller-manager-7bb88cb858-9g8lk\" (UID: \"50571e3f-4e49-47b6-aa7f-2c714e8ef88f\") " pod="openstack-operators/manila-operator-controller-manager-7bb88cb858-9g8lk" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.240734 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-5bdf4f7f7f-29chq"] Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.243073 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-5bdf4f7f7f-29chq" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.245472 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-6scbn" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.245786 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44kf66v"] Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.248259 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44kf66v" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.251085 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-x58pp" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.251188 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.255449 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-5c75d7c94b-66wlz" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.260280 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-5bdf4f7f7f-29chq"] Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.297652 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7879fb76fd-zn5xp" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.303579 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-6dc664666c-p26lg"] Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.305465 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-6dc664666c-p26lg" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.307716 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-64n2k" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.313349 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hcb4g\" (UniqueName: \"kubernetes.io/projected/2ec2f4f5-6ca9-4dae-ae2e-655ef780f3dd-kube-api-access-hcb4g\") pod \"nova-operator-controller-manager-86d796d84d-4lzpn\" (UID: \"2ec2f4f5-6ca9-4dae-ae2e-655ef780f3dd\") " pod="openstack-operators/nova-operator-controller-manager-86d796d84d-4lzpn" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.313420 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xn5br\" (UniqueName: \"kubernetes.io/projected/c4962710-54a6-4538-90ce-be3e27062015-kube-api-access-xn5br\") pod \"mariadb-operator-controller-manager-6f8c5b86cb-rrcjv\" (UID: \"c4962710-54a6-4538-90ce-be3e27062015\") " pod="openstack-operators/mariadb-operator-controller-manager-6f8c5b86cb-rrcjv" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.313496 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xp7x7\" (UniqueName: \"kubernetes.io/projected/bc429681-3e2a-4318-870d-ef62aef7d55b-kube-api-access-xp7x7\") pod \"octavia-operator-controller-manager-6fdc856c5d-mccpp\" (UID: \"bc429681-3e2a-4318-870d-ef62aef7d55b\") " pod="openstack-operators/octavia-operator-controller-manager-6fdc856c5d-mccpp" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.313675 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wjlsk\" (UniqueName: \"kubernetes.io/projected/c4a79bb9-1d05-4c14-a668-6f63f521c98d-kube-api-access-wjlsk\") pod \"neutron-operator-controller-manager-66b7d6f598-hn78m\" (UID: \"c4a79bb9-1d05-4c14-a668-6f63f521c98d\") " pod="openstack-operators/neutron-operator-controller-manager-66b7d6f598-hn78m" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.313719 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6s4mg\" (UniqueName: \"kubernetes.io/projected/4cbc0d4d-5b2e-4678-84e9-7eb69b3742e0-kube-api-access-6s4mg\") pod \"ovn-operator-controller-manager-5bdf4f7f7f-29chq\" (UID: \"4cbc0d4d-5b2e-4678-84e9-7eb69b3742e0\") " pod="openstack-operators/ovn-operator-controller-manager-5bdf4f7f7f-29chq" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.331206 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44kf66v"] Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.335111 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wjlsk\" (UniqueName: \"kubernetes.io/projected/c4a79bb9-1d05-4c14-a668-6f63f521c98d-kube-api-access-wjlsk\") pod \"neutron-operator-controller-manager-66b7d6f598-hn78m\" (UID: \"c4a79bb9-1d05-4c14-a668-6f63f521c98d\") " pod="openstack-operators/neutron-operator-controller-manager-66b7d6f598-hn78m" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.337631 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-6dc664666c-p26lg"] Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.339147 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xn5br\" (UniqueName: \"kubernetes.io/projected/c4962710-54a6-4538-90ce-be3e27062015-kube-api-access-xn5br\") pod \"mariadb-operator-controller-manager-6f8c5b86cb-rrcjv\" (UID: \"c4962710-54a6-4538-90ce-be3e27062015\") " pod="openstack-operators/mariadb-operator-controller-manager-6f8c5b86cb-rrcjv" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.362839 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-799cb6ffd6-h272x"] Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.364172 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-799cb6ffd6-h272x" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.374513 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-x99rs" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.381938 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-799cb6ffd6-h272x"] Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.406923 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-7bb88cb858-9g8lk" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.414851 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bb7wx\" (UniqueName: \"kubernetes.io/projected/d42b84d7-0575-401f-b556-fdaa161588c9-kube-api-access-bb7wx\") pod \"openstack-baremetal-operator-controller-manager-79d88dcd44kf66v\" (UID: \"d42b84d7-0575-401f-b556-fdaa161588c9\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44kf66v" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.414912 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6s4mg\" (UniqueName: \"kubernetes.io/projected/4cbc0d4d-5b2e-4678-84e9-7eb69b3742e0-kube-api-access-6s4mg\") pod \"ovn-operator-controller-manager-5bdf4f7f7f-29chq\" (UID: \"4cbc0d4d-5b2e-4678-84e9-7eb69b3742e0\") " pod="openstack-operators/ovn-operator-controller-manager-5bdf4f7f7f-29chq" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.414947 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-krldn\" (UniqueName: \"kubernetes.io/projected/a41e302b-afc7-45bc-8694-29e5ea87edc8-kube-api-access-krldn\") pod \"placement-operator-controller-manager-6dc664666c-p26lg\" (UID: \"a41e302b-afc7-45bc-8694-29e5ea87edc8\") " pod="openstack-operators/placement-operator-controller-manager-6dc664666c-p26lg" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.414966 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d42b84d7-0575-401f-b556-fdaa161588c9-cert\") pod \"openstack-baremetal-operator-controller-manager-79d88dcd44kf66v\" (UID: \"d42b84d7-0575-401f-b556-fdaa161588c9\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44kf66v" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.414994 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hcb4g\" (UniqueName: \"kubernetes.io/projected/2ec2f4f5-6ca9-4dae-ae2e-655ef780f3dd-kube-api-access-hcb4g\") pod \"nova-operator-controller-manager-86d796d84d-4lzpn\" (UID: \"2ec2f4f5-6ca9-4dae-ae2e-655ef780f3dd\") " pod="openstack-operators/nova-operator-controller-manager-86d796d84d-4lzpn" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.415037 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xp7x7\" (UniqueName: \"kubernetes.io/projected/bc429681-3e2a-4318-870d-ef62aef7d55b-kube-api-access-xp7x7\") pod \"octavia-operator-controller-manager-6fdc856c5d-mccpp\" (UID: \"bc429681-3e2a-4318-870d-ef62aef7d55b\") " pod="openstack-operators/octavia-operator-controller-manager-6fdc856c5d-mccpp" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.437566 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6s4mg\" (UniqueName: \"kubernetes.io/projected/4cbc0d4d-5b2e-4678-84e9-7eb69b3742e0-kube-api-access-6s4mg\") pod \"ovn-operator-controller-manager-5bdf4f7f7f-29chq\" (UID: \"4cbc0d4d-5b2e-4678-84e9-7eb69b3742e0\") " pod="openstack-operators/ovn-operator-controller-manager-5bdf4f7f7f-29chq" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.445047 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-7798859c74-j9vct"] Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.445626 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hcb4g\" (UniqueName: \"kubernetes.io/projected/2ec2f4f5-6ca9-4dae-ae2e-655ef780f3dd-kube-api-access-hcb4g\") pod \"nova-operator-controller-manager-86d796d84d-4lzpn\" (UID: \"2ec2f4f5-6ca9-4dae-ae2e-655ef780f3dd\") " pod="openstack-operators/nova-operator-controller-manager-86d796d84d-4lzpn" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.446185 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-6f8c5b86cb-rrcjv" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.446579 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-7798859c74-j9vct" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.448992 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-87knf" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.457035 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xp7x7\" (UniqueName: \"kubernetes.io/projected/bc429681-3e2a-4318-870d-ef62aef7d55b-kube-api-access-xp7x7\") pod \"octavia-operator-controller-manager-6fdc856c5d-mccpp\" (UID: \"bc429681-3e2a-4318-870d-ef62aef7d55b\") " pod="openstack-operators/octavia-operator-controller-manager-6fdc856c5d-mccpp" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.498391 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-66b7d6f598-hn78m" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.507678 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-7798859c74-j9vct"] Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.518570 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zzp5w\" (UniqueName: \"kubernetes.io/projected/e79f2bb5-95ac-4fbc-a065-da8067f25cb4-kube-api-access-zzp5w\") pod \"telemetry-operator-controller-manager-7798859c74-j9vct\" (UID: \"e79f2bb5-95ac-4fbc-a065-da8067f25cb4\") " pod="openstack-operators/telemetry-operator-controller-manager-7798859c74-j9vct" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.518650 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bb7wx\" (UniqueName: \"kubernetes.io/projected/d42b84d7-0575-401f-b556-fdaa161588c9-kube-api-access-bb7wx\") pod \"openstack-baremetal-operator-controller-manager-79d88dcd44kf66v\" (UID: \"d42b84d7-0575-401f-b556-fdaa161588c9\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44kf66v" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.518683 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ttgqf\" (UniqueName: \"kubernetes.io/projected/0478f782-f1c5-412e-9964-749ee7f83c58-kube-api-access-ttgqf\") pod \"swift-operator-controller-manager-799cb6ffd6-h272x\" (UID: \"0478f782-f1c5-412e-9964-749ee7f83c58\") " pod="openstack-operators/swift-operator-controller-manager-799cb6ffd6-h272x" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.518708 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-krldn\" (UniqueName: \"kubernetes.io/projected/a41e302b-afc7-45bc-8694-29e5ea87edc8-kube-api-access-krldn\") pod \"placement-operator-controller-manager-6dc664666c-p26lg\" (UID: \"a41e302b-afc7-45bc-8694-29e5ea87edc8\") " pod="openstack-operators/placement-operator-controller-manager-6dc664666c-p26lg" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.518730 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d42b84d7-0575-401f-b556-fdaa161588c9-cert\") pod \"openstack-baremetal-operator-controller-manager-79d88dcd44kf66v\" (UID: \"d42b84d7-0575-401f-b556-fdaa161588c9\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44kf66v" Nov 22 08:13:14 crc kubenswrapper[4789]: E1122 08:13:14.518896 4789 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 22 08:13:14 crc kubenswrapper[4789]: E1122 08:13:14.518942 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d42b84d7-0575-401f-b556-fdaa161588c9-cert podName:d42b84d7-0575-401f-b556-fdaa161588c9 nodeName:}" failed. No retries permitted until 2025-11-22 08:13:15.018927678 +0000 UTC m=+1009.253327951 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/d42b84d7-0575-401f-b556-fdaa161588c9-cert") pod "openstack-baremetal-operator-controller-manager-79d88dcd44kf66v" (UID: "d42b84d7-0575-401f-b556-fdaa161588c9") : secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.524559 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-86d796d84d-4lzpn" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.538736 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-8464cf66df-hc8lc"] Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.539105 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-6fdc856c5d-mccpp" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.539959 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-8464cf66df-hc8lc" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.542081 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-grzkb" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.554793 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-krldn\" (UniqueName: \"kubernetes.io/projected/a41e302b-afc7-45bc-8694-29e5ea87edc8-kube-api-access-krldn\") pod \"placement-operator-controller-manager-6dc664666c-p26lg\" (UID: \"a41e302b-afc7-45bc-8694-29e5ea87edc8\") " pod="openstack-operators/placement-operator-controller-manager-6dc664666c-p26lg" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.563632 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-8464cf66df-hc8lc"] Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.571235 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-5bdf4f7f7f-29chq" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.574387 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-7cd4fb6f79-kcq2g"] Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.575684 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-7cd4fb6f79-kcq2g" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.579467 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-klwsw" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.581024 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bb7wx\" (UniqueName: \"kubernetes.io/projected/d42b84d7-0575-401f-b556-fdaa161588c9-kube-api-access-bb7wx\") pod \"openstack-baremetal-operator-controller-manager-79d88dcd44kf66v\" (UID: \"d42b84d7-0575-401f-b556-fdaa161588c9\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44kf66v" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.591279 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-7cd4fb6f79-kcq2g"] Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.621055 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ttgqf\" (UniqueName: \"kubernetes.io/projected/0478f782-f1c5-412e-9964-749ee7f83c58-kube-api-access-ttgqf\") pod \"swift-operator-controller-manager-799cb6ffd6-h272x\" (UID: \"0478f782-f1c5-412e-9964-749ee7f83c58\") " pod="openstack-operators/swift-operator-controller-manager-799cb6ffd6-h272x" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.621181 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pwkjl\" (UniqueName: \"kubernetes.io/projected/50a6fd25-a186-4125-99c1-427d6214b39a-kube-api-access-pwkjl\") pod \"test-operator-controller-manager-8464cf66df-hc8lc\" (UID: \"50a6fd25-a186-4125-99c1-427d6214b39a\") " pod="openstack-operators/test-operator-controller-manager-8464cf66df-hc8lc" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.621262 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zzp5w\" (UniqueName: \"kubernetes.io/projected/e79f2bb5-95ac-4fbc-a065-da8067f25cb4-kube-api-access-zzp5w\") pod \"telemetry-operator-controller-manager-7798859c74-j9vct\" (UID: \"e79f2bb5-95ac-4fbc-a065-da8067f25cb4\") " pod="openstack-operators/telemetry-operator-controller-manager-7798859c74-j9vct" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.629908 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-6dc664666c-p26lg" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.656653 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ttgqf\" (UniqueName: \"kubernetes.io/projected/0478f782-f1c5-412e-9964-749ee7f83c58-kube-api-access-ttgqf\") pod \"swift-operator-controller-manager-799cb6ffd6-h272x\" (UID: \"0478f782-f1c5-412e-9964-749ee7f83c58\") " pod="openstack-operators/swift-operator-controller-manager-799cb6ffd6-h272x" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.669287 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-fzmzt"] Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.670554 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-fzmzt" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.671361 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-fzmzt"] Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.677324 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.677338 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-xbddv" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.678494 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zzp5w\" (UniqueName: \"kubernetes.io/projected/e79f2bb5-95ac-4fbc-a065-da8067f25cb4-kube-api-access-zzp5w\") pod \"telemetry-operator-controller-manager-7798859c74-j9vct\" (UID: \"e79f2bb5-95ac-4fbc-a065-da8067f25cb4\") " pod="openstack-operators/telemetry-operator-controller-manager-7798859c74-j9vct" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.698977 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-77vzm"] Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.700409 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-77vzm" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.702361 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-msjl9" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.708987 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-799cb6ffd6-h272x" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.715006 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-77vzm"] Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.726355 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pwkjl\" (UniqueName: \"kubernetes.io/projected/50a6fd25-a186-4125-99c1-427d6214b39a-kube-api-access-pwkjl\") pod \"test-operator-controller-manager-8464cf66df-hc8lc\" (UID: \"50a6fd25-a186-4125-99c1-427d6214b39a\") " pod="openstack-operators/test-operator-controller-manager-8464cf66df-hc8lc" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.726422 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/84fafb83-aaff-4197-90d1-578c74da9299-cert\") pod \"infra-operator-controller-manager-769d9c7585-f7bd8\" (UID: \"84fafb83-aaff-4197-90d1-578c74da9299\") " pod="openstack-operators/infra-operator-controller-manager-769d9c7585-f7bd8" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.726465 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j82kq\" (UniqueName: \"kubernetes.io/projected/ca7faed1-038d-4ec8-800d-35848427a921-kube-api-access-j82kq\") pod \"watcher-operator-controller-manager-7cd4fb6f79-kcq2g\" (UID: \"ca7faed1-038d-4ec8-800d-35848427a921\") " pod="openstack-operators/watcher-operator-controller-manager-7cd4fb6f79-kcq2g" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.751292 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pwkjl\" (UniqueName: \"kubernetes.io/projected/50a6fd25-a186-4125-99c1-427d6214b39a-kube-api-access-pwkjl\") pod \"test-operator-controller-manager-8464cf66df-hc8lc\" (UID: \"50a6fd25-a186-4125-99c1-427d6214b39a\") " pod="openstack-operators/test-operator-controller-manager-8464cf66df-hc8lc" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.755480 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/84fafb83-aaff-4197-90d1-578c74da9299-cert\") pod \"infra-operator-controller-manager-769d9c7585-f7bd8\" (UID: \"84fafb83-aaff-4197-90d1-578c74da9299\") " pod="openstack-operators/infra-operator-controller-manager-769d9c7585-f7bd8" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.784187 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-7798859c74-j9vct" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.827427 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-8464cf66df-hc8lc" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.828011 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/03652378-7433-48c7-a601-aea6d67f1a14-cert\") pod \"openstack-operator-controller-manager-6cb9dc54f8-fzmzt\" (UID: \"03652378-7433-48c7-a601-aea6d67f1a14\") " pod="openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-fzmzt" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.828067 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rt9ls\" (UniqueName: \"kubernetes.io/projected/03652378-7433-48c7-a601-aea6d67f1a14-kube-api-access-rt9ls\") pod \"openstack-operator-controller-manager-6cb9dc54f8-fzmzt\" (UID: \"03652378-7433-48c7-a601-aea6d67f1a14\") " pod="openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-fzmzt" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.828097 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nhfq8\" (UniqueName: \"kubernetes.io/projected/c94020a6-11c7-4ad5-90bb-8dd40e08ffd2-kube-api-access-nhfq8\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-77vzm\" (UID: \"c94020a6-11c7-4ad5-90bb-8dd40e08ffd2\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-77vzm" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.828144 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j82kq\" (UniqueName: \"kubernetes.io/projected/ca7faed1-038d-4ec8-800d-35848427a921-kube-api-access-j82kq\") pod \"watcher-operator-controller-manager-7cd4fb6f79-kcq2g\" (UID: \"ca7faed1-038d-4ec8-800d-35848427a921\") " pod="openstack-operators/watcher-operator-controller-manager-7cd4fb6f79-kcq2g" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.855195 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j82kq\" (UniqueName: \"kubernetes.io/projected/ca7faed1-038d-4ec8-800d-35848427a921-kube-api-access-j82kq\") pod \"watcher-operator-controller-manager-7cd4fb6f79-kcq2g\" (UID: \"ca7faed1-038d-4ec8-800d-35848427a921\") " pod="openstack-operators/watcher-operator-controller-manager-7cd4fb6f79-kcq2g" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.882221 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-769d9c7585-f7bd8" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.930386 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rt9ls\" (UniqueName: \"kubernetes.io/projected/03652378-7433-48c7-a601-aea6d67f1a14-kube-api-access-rt9ls\") pod \"openstack-operator-controller-manager-6cb9dc54f8-fzmzt\" (UID: \"03652378-7433-48c7-a601-aea6d67f1a14\") " pod="openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-fzmzt" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.932778 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nhfq8\" (UniqueName: \"kubernetes.io/projected/c94020a6-11c7-4ad5-90bb-8dd40e08ffd2-kube-api-access-nhfq8\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-77vzm\" (UID: \"c94020a6-11c7-4ad5-90bb-8dd40e08ffd2\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-77vzm" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.932969 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/03652378-7433-48c7-a601-aea6d67f1a14-cert\") pod \"openstack-operator-controller-manager-6cb9dc54f8-fzmzt\" (UID: \"03652378-7433-48c7-a601-aea6d67f1a14\") " pod="openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-fzmzt" Nov 22 08:13:14 crc kubenswrapper[4789]: E1122 08:13:14.935311 4789 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Nov 22 08:13:14 crc kubenswrapper[4789]: E1122 08:13:14.935386 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/03652378-7433-48c7-a601-aea6d67f1a14-cert podName:03652378-7433-48c7-a601-aea6d67f1a14 nodeName:}" failed. No retries permitted until 2025-11-22 08:13:15.435370335 +0000 UTC m=+1009.669770608 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/03652378-7433-48c7-a601-aea6d67f1a14-cert") pod "openstack-operator-controller-manager-6cb9dc54f8-fzmzt" (UID: "03652378-7433-48c7-a601-aea6d67f1a14") : secret "webhook-server-cert" not found Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.947923 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rt9ls\" (UniqueName: \"kubernetes.io/projected/03652378-7433-48c7-a601-aea6d67f1a14-kube-api-access-rt9ls\") pod \"openstack-operator-controller-manager-6cb9dc54f8-fzmzt\" (UID: \"03652378-7433-48c7-a601-aea6d67f1a14\") " pod="openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-fzmzt" Nov 22 08:13:14 crc kubenswrapper[4789]: I1122 08:13:14.966920 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nhfq8\" (UniqueName: \"kubernetes.io/projected/c94020a6-11c7-4ad5-90bb-8dd40e08ffd2-kube-api-access-nhfq8\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-77vzm\" (UID: \"c94020a6-11c7-4ad5-90bb-8dd40e08ffd2\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-77vzm" Nov 22 08:13:15 crc kubenswrapper[4789]: I1122 08:13:15.034111 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d42b84d7-0575-401f-b556-fdaa161588c9-cert\") pod \"openstack-baremetal-operator-controller-manager-79d88dcd44kf66v\" (UID: \"d42b84d7-0575-401f-b556-fdaa161588c9\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44kf66v" Nov 22 08:13:15 crc kubenswrapper[4789]: I1122 08:13:15.039629 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d42b84d7-0575-401f-b556-fdaa161588c9-cert\") pod \"openstack-baremetal-operator-controller-manager-79d88dcd44kf66v\" (UID: \"d42b84d7-0575-401f-b556-fdaa161588c9\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44kf66v" Nov 22 08:13:15 crc kubenswrapper[4789]: I1122 08:13:15.155045 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-7cd4fb6f79-kcq2g" Nov 22 08:13:15 crc kubenswrapper[4789]: I1122 08:13:15.161256 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7768f8c84f-hwbvw"] Nov 22 08:13:15 crc kubenswrapper[4789]: W1122 08:13:15.165653 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod93c2cbf6_757a_4de1_9f83_115787d74f31.slice/crio-b5408cf3edf0ab667517b346658eeaf9d1a78f2e37b4e2a445a1b9c61edaf89d WatchSource:0}: Error finding container b5408cf3edf0ab667517b346658eeaf9d1a78f2e37b4e2a445a1b9c61edaf89d: Status 404 returned error can't find the container with id b5408cf3edf0ab667517b346658eeaf9d1a78f2e37b4e2a445a1b9c61edaf89d Nov 22 08:13:15 crc kubenswrapper[4789]: I1122 08:13:15.174828 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-77vzm" Nov 22 08:13:15 crc kubenswrapper[4789]: I1122 08:13:15.189897 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44kf66v" Nov 22 08:13:15 crc kubenswrapper[4789]: I1122 08:13:15.455540 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/03652378-7433-48c7-a601-aea6d67f1a14-cert\") pod \"openstack-operator-controller-manager-6cb9dc54f8-fzmzt\" (UID: \"03652378-7433-48c7-a601-aea6d67f1a14\") " pod="openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-fzmzt" Nov 22 08:13:15 crc kubenswrapper[4789]: I1122 08:13:15.463773 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/03652378-7433-48c7-a601-aea6d67f1a14-cert\") pod \"openstack-operator-controller-manager-6cb9dc54f8-fzmzt\" (UID: \"03652378-7433-48c7-a601-aea6d67f1a14\") " pod="openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-fzmzt" Nov 22 08:13:15 crc kubenswrapper[4789]: I1122 08:13:15.484946 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5c75d7c94b-66wlz"] Nov 22 08:13:15 crc kubenswrapper[4789]: W1122 08:13:15.495005 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod29d2daa3_90b3_4dea_9612_edc4cc0dcc30.slice/crio-c70221042a62c8555c2c8a201fef5c56a07a6ac6763cf9ce1299e432c9ba429b WatchSource:0}: Error finding container c70221042a62c8555c2c8a201fef5c56a07a6ac6763cf9ce1299e432c9ba429b: Status 404 returned error can't find the container with id c70221042a62c8555c2c8a201fef5c56a07a6ac6763cf9ce1299e432c9ba429b Nov 22 08:13:15 crc kubenswrapper[4789]: I1122 08:13:15.495044 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5d86b44686-nk4zk"] Nov 22 08:13:15 crc kubenswrapper[4789]: W1122 08:13:15.527441 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod77bd1957_6cd6_418b_9b2c_ac879bbca15e.slice/crio-92531ae80a0f9d40b21722a330373949d3a2d40c9b3f204246847ca87c1203cf WatchSource:0}: Error finding container 92531ae80a0f9d40b21722a330373949d3a2d40c9b3f204246847ca87c1203cf: Status 404 returned error can't find the container with id 92531ae80a0f9d40b21722a330373949d3a2d40c9b3f204246847ca87c1203cf Nov 22 08:13:15 crc kubenswrapper[4789]: I1122 08:13:15.535609 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6d8fd67bf7-2klns"] Nov 22 08:13:15 crc kubenswrapper[4789]: I1122 08:13:15.554484 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-56dfb6b67f-p4vmr"] Nov 22 08:13:15 crc kubenswrapper[4789]: I1122 08:13:15.712795 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7768f8c84f-hwbvw" event={"ID":"93c2cbf6-757a-4de1-9f83-115787d74f31","Type":"ContainerStarted","Data":"b5408cf3edf0ab667517b346658eeaf9d1a78f2e37b4e2a445a1b9c61edaf89d"} Nov 22 08:13:15 crc kubenswrapper[4789]: I1122 08:13:15.713986 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6d8fd67bf7-2klns" event={"ID":"55d6892d-7d96-4518-8d36-58048cf96c52","Type":"ContainerStarted","Data":"df8722329a3aff29ac5bd774018f562c410bd622b892271923b8520eb6c79958"} Nov 22 08:13:15 crc kubenswrapper[4789]: I1122 08:13:15.716958 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-5d86b44686-nk4zk" event={"ID":"9e20512a-a64f-4f47-a7ef-b91f24390aef","Type":"ContainerStarted","Data":"ae19d0334b9d909f9ec7343b0c1d77c8ff117d910ce60a57f5ec37356d8cba42"} Nov 22 08:13:15 crc kubenswrapper[4789]: I1122 08:13:15.718804 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-56dfb6b67f-p4vmr" event={"ID":"77bd1957-6cd6-418b-9b2c-ac879bbca15e","Type":"ContainerStarted","Data":"92531ae80a0f9d40b21722a330373949d3a2d40c9b3f204246847ca87c1203cf"} Nov 22 08:13:15 crc kubenswrapper[4789]: I1122 08:13:15.719849 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5c75d7c94b-66wlz" event={"ID":"29d2daa3-90b3-4dea-9612-edc4cc0dcc30","Type":"ContainerStarted","Data":"c70221042a62c8555c2c8a201fef5c56a07a6ac6763cf9ce1299e432c9ba429b"} Nov 22 08:13:15 crc kubenswrapper[4789]: I1122 08:13:15.756450 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-fzmzt" Nov 22 08:13:15 crc kubenswrapper[4789]: I1122 08:13:15.773925 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-5bdf4f7f7f-29chq"] Nov 22 08:13:15 crc kubenswrapper[4789]: I1122 08:13:15.775993 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-86d796d84d-4lzpn"] Nov 22 08:13:15 crc kubenswrapper[4789]: I1122 08:13:15.791057 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-8667fbf6f6-k5kms"] Nov 22 08:13:15 crc kubenswrapper[4789]: I1122 08:13:15.809398 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-7798859c74-j9vct"] Nov 22 08:13:15 crc kubenswrapper[4789]: W1122 08:13:15.813575 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2ec2f4f5_6ca9_4dae_ae2e_655ef780f3dd.slice/crio-ddf865273d20ac7f0a9afadcd48f8248bdae5dcf673ba8256d2386e13b96d163 WatchSource:0}: Error finding container ddf865273d20ac7f0a9afadcd48f8248bdae5dcf673ba8256d2386e13b96d163: Status 404 returned error can't find the container with id ddf865273d20ac7f0a9afadcd48f8248bdae5dcf673ba8256d2386e13b96d163 Nov 22 08:13:15 crc kubenswrapper[4789]: I1122 08:13:15.819391 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-6dc664666c-p26lg"] Nov 22 08:13:15 crc kubenswrapper[4789]: W1122 08:13:15.830420 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode79f2bb5_95ac_4fbc_a065_da8067f25cb4.slice/crio-9fceaa5ed64e7dec74fc8c76902cc469918f506882f988bd9cb64f215a160282 WatchSource:0}: Error finding container 9fceaa5ed64e7dec74fc8c76902cc469918f506882f988bd9cb64f215a160282: Status 404 returned error can't find the container with id 9fceaa5ed64e7dec74fc8c76902cc469918f506882f988bd9cb64f215a160282 Nov 22 08:13:15 crc kubenswrapper[4789]: W1122 08:13:15.833032 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc4962710_54a6_4538_90ce_be3e27062015.slice/crio-156a4ee2521938812a90dfc5f2b9d67225c144c80a24bdbbfdfa21aa3a60fe90 WatchSource:0}: Error finding container 156a4ee2521938812a90dfc5f2b9d67225c144c80a24bdbbfdfa21aa3a60fe90: Status 404 returned error can't find the container with id 156a4ee2521938812a90dfc5f2b9d67225c144c80a24bdbbfdfa21aa3a60fe90 Nov 22 08:13:15 crc kubenswrapper[4789]: W1122 08:13:15.835591 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda41e302b_afc7_45bc_8694_29e5ea87edc8.slice/crio-2a9857a1cafc9ca6992987debd9f6ded9a8f2aa3b454a616ab0cf978dbd3b9ae WatchSource:0}: Error finding container 2a9857a1cafc9ca6992987debd9f6ded9a8f2aa3b454a616ab0cf978dbd3b9ae: Status 404 returned error can't find the container with id 2a9857a1cafc9ca6992987debd9f6ded9a8f2aa3b454a616ab0cf978dbd3b9ae Nov 22 08:13:15 crc kubenswrapper[4789]: I1122 08:13:15.836866 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-bf4c6585d-6shgh"] Nov 22 08:13:15 crc kubenswrapper[4789]: I1122 08:13:15.849114 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-7bb88cb858-9g8lk"] Nov 22 08:13:15 crc kubenswrapper[4789]: I1122 08:13:15.853006 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6f8c5b86cb-rrcjv"] Nov 22 08:13:15 crc kubenswrapper[4789]: I1122 08:13:15.859054 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7879fb76fd-zn5xp"] Nov 22 08:13:15 crc kubenswrapper[4789]: E1122 08:13:15.861376 4789 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:c0b5f124a37c1538042c0e63f0978429572e2a851d7f3a6eb80de09b86d755a0,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-ttgqf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-799cb6ffd6-h272x_openstack-operators(0478f782-f1c5-412e-9964-749ee7f83c58): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 22 08:13:15 crc kubenswrapper[4789]: I1122 08:13:15.862885 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-799cb6ffd6-h272x"] Nov 22 08:13:15 crc kubenswrapper[4789]: I1122 08:13:15.868371 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-66b7d6f598-hn78m"] Nov 22 08:13:15 crc kubenswrapper[4789]: W1122 08:13:15.869299 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0cf958d5_222e_4598_a1a7_fe887644cd24.slice/crio-6ba9216723546adf91f8b3ef5f285813068becf305cfd575d42d6cc6f4b7392f WatchSource:0}: Error finding container 6ba9216723546adf91f8b3ef5f285813068becf305cfd575d42d6cc6f4b7392f: Status 404 returned error can't find the container with id 6ba9216723546adf91f8b3ef5f285813068becf305cfd575d42d6cc6f4b7392f Nov 22 08:13:15 crc kubenswrapper[4789]: E1122 08:13:15.870716 4789 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:3ef72bbd7cce89ff54d850ff44ca6d7b2360834a502da3d561aeb6fd3d9af50a,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9hmbk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-7879fb76fd-zn5xp_openstack-operators(490be79a-ce0c-4a3e-a819-7ef482fe453d): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 22 08:13:15 crc kubenswrapper[4789]: W1122 08:13:15.873321 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod50a6fd25_a186_4125_99c1_427d6214b39a.slice/crio-6c28791a936835e3dee64c24283ca9b86f7a7e771c346802cf472cf75057407f WatchSource:0}: Error finding container 6c28791a936835e3dee64c24283ca9b86f7a7e771c346802cf472cf75057407f: Status 404 returned error can't find the container with id 6c28791a936835e3dee64c24283ca9b86f7a7e771c346802cf472cf75057407f Nov 22 08:13:15 crc kubenswrapper[4789]: I1122 08:13:15.874213 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-8464cf66df-hc8lc"] Nov 22 08:13:15 crc kubenswrapper[4789]: E1122 08:13:15.874762 4789 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/heat-operator@sha256:5edd825a235f5784d9a65892763c5388c39df1731d0fcbf4ee33408b8c83ac96,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-bc6d5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-operator-controller-manager-bf4c6585d-6shgh_openstack-operators(0cf958d5-222e-4598-a1a7-fe887644cd24): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 22 08:13:15 crc kubenswrapper[4789]: W1122 08:13:15.875324 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbc429681_3e2a_4318_870d_ef62aef7d55b.slice/crio-ffa42c160099dfc2e95f65912cdef590f45e5fe8492b525e66b5a6b1d683c37a WatchSource:0}: Error finding container ffa42c160099dfc2e95f65912cdef590f45e5fe8492b525e66b5a6b1d683c37a: Status 404 returned error can't find the container with id ffa42c160099dfc2e95f65912cdef590f45e5fe8492b525e66b5a6b1d683c37a Nov 22 08:13:15 crc kubenswrapper[4789]: W1122 08:13:15.878037 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc4a79bb9_1d05_4c14_a668_6f63f521c98d.slice/crio-806cbf437702e93135fde0b46d84848b30eec519cf2ab54cf8b1940724abe49d WatchSource:0}: Error finding container 806cbf437702e93135fde0b46d84848b30eec519cf2ab54cf8b1940724abe49d: Status 404 returned error can't find the container with id 806cbf437702e93135fde0b46d84848b30eec519cf2ab54cf8b1940724abe49d Nov 22 08:13:15 crc kubenswrapper[4789]: I1122 08:13:15.880087 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-6fdc856c5d-mccpp"] Nov 22 08:13:15 crc kubenswrapper[4789]: E1122 08:13:15.891161 4789 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:82207e753574d4be246f86c4b074500d66cf20214aa80f0a8525cf3287a35e6d,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-pwkjl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-8464cf66df-hc8lc_openstack-operators(50a6fd25-a186-4125-99c1-427d6214b39a): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 22 08:13:15 crc kubenswrapper[4789]: E1122 08:13:15.891375 4789 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:442c269d79163f8da75505019c02e9f0815837aaadcaddacb8e6c12df297ca13,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-xp7x7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-6fdc856c5d-mccpp_openstack-operators(bc429681-3e2a-4318-870d-ef62aef7d55b): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 22 08:13:15 crc kubenswrapper[4789]: E1122 08:13:15.891475 4789 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:207578cb433471cc1a79c21a808c8a15489d1d3c9fa77e29f3f697c33917fec6,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-wjlsk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-66b7d6f598-hn78m_openstack-operators(c4a79bb9-1d05-4c14-a668-6f63f521c98d): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 22 08:13:15 crc kubenswrapper[4789]: I1122 08:13:15.954012 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-769d9c7585-f7bd8"] Nov 22 08:13:16 crc kubenswrapper[4789]: I1122 08:13:16.046103 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-7cd4fb6f79-kcq2g"] Nov 22 08:13:16 crc kubenswrapper[4789]: I1122 08:13:16.046553 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44kf66v"] Nov 22 08:13:16 crc kubenswrapper[4789]: I1122 08:13:16.063436 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-77vzm"] Nov 22 08:13:16 crc kubenswrapper[4789]: E1122 08:13:16.087419 4789 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:4838402d41d42c56613d43dc5041aae475a2b18e6172491d6c4d4a78a580697f,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-j82kq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-7cd4fb6f79-kcq2g_openstack-operators(ca7faed1-038d-4ec8-800d-35848427a921): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 22 08:13:16 crc kubenswrapper[4789]: E1122 08:13:16.088524 4789 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:78852f8ba332a5756c1551c126157f735279101a0fc3277ba4aa4db3478789dd,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-baremetal-operator-agent@sha256:7dbadf7b98f2f305f9f1382f55a084c8ca404f4263f76b28e56bd0dc437e2192,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_ANSIBLEEE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-ansibleee-runner@sha256:0473ff9eec0da231e2d0a10bf1abbe1dfa1a0f95b8f619e3a07605386951449a,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-api@sha256:c8101c77a82eae4407e41e1fd766dfc6e1b7f9ed1679e3efb6f91ff97a1557b2,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_EVALUATOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-evaluator@sha256:eb9743b21bbadca6f7cb9ac4fc46b5d58c51c674073c7e1121f4474a71304071,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-listener@sha256:3d81f839b98c2e2a5bf0da79f2f9a92dff7d0a3c5a830b0e95c89dad8cf98a6a,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_NOTIFIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-notifier@sha256:d19ac99249b47dd8ea16cd6aaa5756346aa8a2f119ee50819c15c5366efb417d,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_APACHE_IMAGE_URL_DEFAULT,Value:registry.redhat.io/ubi9/httpd-24@sha256:8536169e5537fe6c330eba814248abdcf39cdd8f7e7336034d74e6fda9544050,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-api@sha256:4c93a5cccb9971e24f05daf93b3aa11ba71752bc3469a1a1a2c4906f92f69645,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_KEYSTONE_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-keystone-listener@sha256:4f1fa337760e82bfd67cdd142a97c121146dd7e621daac161940dd5e4ddb80dc,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-worker@sha256:3613b345d5baed98effd906f8b0242d863e14c97078ea473ef01fe1b0afc46f3,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-central@sha256:d375d370be5ead0dac71109af644849e5795f535f9ad8eeacea261d77ae6f140,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-compute@sha256:9f9f367ed4c85efb16c3a74a4bb707ff0db271d7bc5abc70a71e984b55f43003,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_IPMI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-ipmi@sha256:b73ad22b4955b06d584bce81742556d8c0c7828c495494f8ea7c99391c61b70f,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_MYSQLD_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/mysqld-exporter@sha256:7211a617ec657701ca819aa0ba28e1d5750f5bf2c1391b755cc4a48cc360b0fa,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_NOTIFICATION_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-notification@sha256:aa1d3aaf6b394621ed4089a98e0a82b763f467e8b5c5db772f9fdf99fc86e333,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_SGCORE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/sg-core@sha256:09b5017c95d7697e66b9c64846bc48ef5826a009cba89b956ec54561e5f4a2d1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:37d64e0a00c54e71a4c1fcbbbf7e832f6886ffd03c9a02b6ee3ca48fabc30879,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_BACKUP_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-backup@sha256:d6661053141b6df421288a7c9968a155ab82e478c1d75ab41f2cebe2f0ca02d2,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-scheduler@sha256:ce2d63258cb4e7d0d1c07234de6889c5434464190906798019311a1c7cf6387f,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_VOLUME_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-volume@sha256:0485ef9e5b4437f7cd2ba54034a87722ce4669ee86b3773c6b0c037ed8000e91,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CLOUDKITTY_API_IMAGE_URL_DEFAULT,Value:quay.rdoproject.org/podified-master-centos10/openstack-cloudkitty-api@sha256:962c004551d0503779364b767b9bf0cecdf78dbba8809b2ca8b073f58e1f4e5d,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CLOUDKITTY_PROC_IMAGE_URL_DEFAULT,Value:quay.rdoproject.org/podified-master-centos10/openstack-cloudkitty-processor@sha256:0ebf4c465fb6cc7dad9e6cb2da0ff54874c9acbcb40d62234a629ec2c12cdd62,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-api@sha256:ff0c553ceeb2e0f44b010e37dc6d0db8a251797b88e56468b7cf7f05253e4232,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_BACKENDBIND9_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-backend-bind9@sha256:624f553f073af7493d34828b074adc9981cce403edd8e71482c7307008479fd9,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-central@sha256:e3874936a518c8560339db8f840fc5461885819f6050b5de8d3ab9199bea5094,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_MDNS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-mdns@sha256:1cea25f1d2a45affc80c46fb9d427749d3f06b61590ac6070a2910e3ec8a4e5d,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_PRODUCER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-producer@sha256:e36d5b9a65194f12f7b01c6422ba3ed52a687fd1695fbb21f4986c67d9f9317f,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_UNBOUND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-unbound@sha256:8b21bec527d54cd766e277889df6bcccd2baeaa946274606b986c0c3b7ca689f,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-worker@sha256:45aceca77f8fcf61127f0da650bdfdf11ede9b0944c78b63fab819d03283f96b,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_FRR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-frr@sha256:709ac58998927dd61786821ae1e63343fd97ccf5763aac5edb4583eea9401d22,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_ISCSID_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-iscsid@sha256:867d4ef7c21f75e6030a685b5762ab4d84b671316ed6b98d75200076e93342cd,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_KEPLER_IMAGE_URL_DEFAULT,Value:quay.io/sustainable_computing_io/kepler@sha256:581b65b646301e0fcb07582150ba63438f1353a85bf9acf1eb2acb4ce71c58bd,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_LOGROTATE_CROND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cron@sha256:2b90da93550b99d2fcfa95bd819f3363aa68346a416f8dc7baac3e9c5f487761,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_MULTIPATHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-multipathd@sha256:6f86db36d668348be8c5b46dcda8b1fa23d34bfdc07164fbcbe7a6327fb4de24,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_DHCP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent@sha256:8cde52cef8795d1c91983b100d86541c7718160ec260fe0f97b96add4c2c8ee8,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_METADATA_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn@sha256:a9583cb3baf440d2358ef041373833afbeae60da8159dd031502379901141620,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_OVN_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-ovn-agent@sha256:835ebed082fe1c45bd799d1d5357595ce63efeb05ca876f26b08443facb9c164,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_SRIOV_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-sriov-agent@sha256:011d682241db724bc40736c9b54d2ea450ea7e6be095b1ff5fa28c8007466775,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NODE_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_OVN_BGP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-bgp-agent@sha256:2025da90cff8f563deb08bee71efe16d4078edc2a767b2e225cca5c77f1aa2f9,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_PODMAN_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_GLANCE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-glance-api@sha256:26bd7b0bd6070856aefef6fe754c547d55c056396ea30d879d34c2d49b5a1d29,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api@sha256:ff46cd5e0e13d105c4629e78c2734a50835f06b6a1e31da9e0462981d10c4be3,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_CFNAPI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api-cfn@sha256:5b4fd0c2b76fa5539f74687b11c5882d77bd31352452322b37ff51fa18f12a61,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-engine@sha256:5e03376bd895346dc8f627ca15ded942526ed8b5e92872f453ce272e694d18d4,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HORIZON_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-horizon@sha256:65b94ff9fcd486845fb0544583bf2a973246a61a0ad32340fb92d632285f1057,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_MEMCACHED_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-memcached@sha256:36a0fb31978aee0ded2483de311631e64a644d0b0685b5b055f65ede7eb8e8a2,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_REDIS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-redis@sha256:5f6045841aff0fde6f684a34cdf49f8dc7b2c3bcbdeab201f1058971e0c5f79e,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-api@sha256:448f4e1b740c30936e340bd6e8534d78c83357bf373a4223950aa64d3484f007,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-conductor@sha256:b68e3615af8a0eb0ef6bf9ceeef59540a6f4a9a85f6078a3620be115c73a7db8,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_INSPECTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-inspector@sha256:7eae01cf60383e523c9cd94d158a9162120a7370829a1dad20fdea6b0fd660bd,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_NEUTRON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-neutron-agent@sha256:28cc10501788081eb61b5a1af35546191a92741f4f109df54c74e2b19439d0f9,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PXE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-pxe@sha256:9a616e37acfd120612f78043237a8541266ba34883833c9beb43f3da313661ad,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PYTHON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/ironic-python-agent@sha256:6b1be6cd94a0942259bca5d5d2c30cc7de4a33276b61f8ae3940226772106256,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KEYSTONE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-keystone@sha256:02d2c22d15401574941fbe057095442dee0d6f7a0a9341de35d25e6a12a3fe4b,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KSM_IMAGE_URL_DEFAULT,Value:registry.k8s.io/kube-state-metrics/kube-state-metrics@sha256:db384bf43222b066c378e77027a675d4cd9911107adba46c2922b3a55e10d6fb,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-api@sha256:fc3b3a36b74fd653946723c54b208072d52200635850b531e9d595a7aaea5a01,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-scheduler@sha256:7850ccbff320bf9a1c9c769c1c70777eb97117dd8cd5ae4435be9b4622cf807a,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SHARE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-share@sha256:397dac7e39cf40d14a986e6ec4a60fb698ca35c197d0db315b1318514cc6d1d4,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MARIADB_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-mariadb@sha256:10452e2144368e2f128c8fb8ef9e54880b06ef1d71d9f084a0217dcb099c51ce,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NET_UTILS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-netutils@sha256:1c95142a36276686e720f86423ee171dc9adcc1e89879f627545b7c906ccd9bd,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NEUTRON_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:18f8463fe46fe6081d5682009e92bbcb3df33282b83b0a2857abaece795cf1ba,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-api@sha256:e331a8fde6638e5ba154c4f0b38772a9a424f60656f2777245975fb1fa02f07d,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-compute@sha256:b6e1e8a249d36ef36c6ac4170af1e043dda1ccc0f9672832d3ff151bf3533076,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-conductor@sha256:cd3cf7a34053e850b4d4f9f4ea4c74953a54a42fd18e47d7c01d44a88923e925,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_NOVNC_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-novncproxy@sha256:aee28476344fc0cc148fbe97daf9b1bfcedc22001550bba4bdc4e84be7b6989d,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-scheduler@sha256:cfa0b92c976603ee2a937d34013a238fcd8aa75f998e50642e33489f14124633,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-api@sha256:73c2f2d6eecf88acf4e45b133c8373d9bb006b530e0aff0b28f3b7420620a874,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HEALTHMANAGER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-health-manager@sha256:927b405cc04abe5ff716186e8d35e2dc5fad1c8430194659ee6617d74e4e055d,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HOUSEKEEPING_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-housekeeping@sha256:6154d7cebd7c339afa5b86330262156171743aa5b79c2b78f9a2f378005ed8fb,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_RSYSLOG_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rsyslog@sha256:e2db2f4af8d3d0be7868c6efef0189f3a2c74a8f96ae10e3f991cdf83feaef29,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-worker@sha256:c773629df257726a6d3cacc24a6e4df0babcd7d37df04e6d14676a8da028b9c9,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_CLIENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-openstackclient@sha256:776211111e2e6493706dbc49a3ba44f31d1b947919313ed3a0f35810e304ec52,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_MUST_GATHER_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-must-gather@sha256:0a98e8f5c83522ca6c8e40c5e9561f6628d2d5e69f0e8a64279c541c989d3d8b,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_NETWORK_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OS_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/edpm-hardened-uefi@sha256:7cccf24ad0a152f90ca39893064f48a1656950ee8142685a5d482c71f0bdc9f5,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-controller@sha256:af46761060c7987e1dee5f14c06d85b46f12ad8e09c83d4246ab4e3a65dfda3e,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_OVS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-base@sha256:05450b48f6b5352b2686a26e933e8727748edae2ae9652d9164b7d7a1817c55a,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server@sha256:fc9c99eeef91523482bd8f92661b393287e1f2a24ad2ba9e33191f8de9af74cf,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NORTHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-northd@sha256:3e4ecc02b4b5e0860482a93599ba9ca598c5ce26c093c46e701f96fe51acb208,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_SB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-sb-db-server@sha256:2346037e064861c7892690d2e8b3e1eea1a26ce3c3a11fda0b41301965bc828c,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_PLACEMENT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-placement-api@sha256:7dd2e0dbb6bb5a6cecd1763e43479ca8cb6a0c502534e83c8795c0da2b50e099,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_RABBITMQ_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rabbitmq@sha256:95d67f51dfedd5bd3ec785b488425295b2d8c41feae3e6386ef471615381809b,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_ACCOUNT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-account@sha256:c26c3ff9cabe3593ceb10006e782bf9391ac14785768ce9eec4f938c2d3cf228,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-container@sha256:273fe8c27d08d0f62773a02f8cef6a761a7768116ee1a4be611f93bbf63f2b75,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_OBJECT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-object@sha256:daa45220bb1c47922d0917aa8fe423bb82b03a01429f1c9e37635e701e352d71,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_PROXY_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-proxy-server@sha256:a80a074e227d3238bb6f285788a9e886ae7a5909ccbc5c19c93c369bdfe5b3b8,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_TEST_TEMPEST_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-tempest-all@sha256:58ac66ca1be01fe0157977bd79a26cde4d0de153edfaf4162367c924826b2ef4,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-api@sha256:99a63770d80cc7c3afa1118b400972fb0e6bff5284a2eae781b12582ad79c29c,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_APPLIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-applier@sha256:9ee4d84529394afcd860f1a1186484560f02f08c15c37cac42a22473b7116d5f,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_DECISION_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-decision-engine@sha256:ea15fadda7b0439ec637edfaf6ea5dbf3e35fb3be012c7c5a31e722c90becb11,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bb7wx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-baremetal-operator-controller-manager-79d88dcd44kf66v_openstack-operators(d42b84d7-0575-401f-b556-fdaa161588c9): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 22 08:13:16 crc kubenswrapper[4789]: E1122 08:13:16.104562 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/heat-operator-controller-manager-bf4c6585d-6shgh" podUID="0cf958d5-222e-4598-a1a7-fe887644cd24" Nov 22 08:13:16 crc kubenswrapper[4789]: W1122 08:13:16.119556 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc94020a6_11c7_4ad5_90bb_8dd40e08ffd2.slice/crio-3561ca00da344c3672a8e0f225784e8bb3f416792b143020b967159a1a538d30 WatchSource:0}: Error finding container 3561ca00da344c3672a8e0f225784e8bb3f416792b143020b967159a1a538d30: Status 404 returned error can't find the container with id 3561ca00da344c3672a8e0f225784e8bb3f416792b143020b967159a1a538d30 Nov 22 08:13:16 crc kubenswrapper[4789]: E1122 08:13:16.125967 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/keystone-operator-controller-manager-7879fb76fd-zn5xp" podUID="490be79a-ce0c-4a3e-a819-7ef482fe453d" Nov 22 08:13:16 crc kubenswrapper[4789]: I1122 08:13:16.156871 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-fzmzt"] Nov 22 08:13:16 crc kubenswrapper[4789]: E1122 08:13:16.411081 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/swift-operator-controller-manager-799cb6ffd6-h272x" podUID="0478f782-f1c5-412e-9964-749ee7f83c58" Nov 22 08:13:16 crc kubenswrapper[4789]: E1122 08:13:16.457644 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44kf66v" podUID="d42b84d7-0575-401f-b556-fdaa161588c9" Nov 22 08:13:16 crc kubenswrapper[4789]: E1122 08:13:16.478906 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/test-operator-controller-manager-8464cf66df-hc8lc" podUID="50a6fd25-a186-4125-99c1-427d6214b39a" Nov 22 08:13:16 crc kubenswrapper[4789]: E1122 08:13:16.511376 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/neutron-operator-controller-manager-66b7d6f598-hn78m" podUID="c4a79bb9-1d05-4c14-a668-6f63f521c98d" Nov 22 08:13:16 crc kubenswrapper[4789]: E1122 08:13:16.511491 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/octavia-operator-controller-manager-6fdc856c5d-mccpp" podUID="bc429681-3e2a-4318-870d-ef62aef7d55b" Nov 22 08:13:16 crc kubenswrapper[4789]: E1122 08:13:16.540317 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/watcher-operator-controller-manager-7cd4fb6f79-kcq2g" podUID="ca7faed1-038d-4ec8-800d-35848427a921" Nov 22 08:13:16 crc kubenswrapper[4789]: I1122 08:13:16.728695 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-7798859c74-j9vct" event={"ID":"e79f2bb5-95ac-4fbc-a065-da8067f25cb4","Type":"ContainerStarted","Data":"9fceaa5ed64e7dec74fc8c76902cc469918f506882f988bd9cb64f215a160282"} Nov 22 08:13:16 crc kubenswrapper[4789]: I1122 08:13:16.734278 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-6dc664666c-p26lg" event={"ID":"a41e302b-afc7-45bc-8694-29e5ea87edc8","Type":"ContainerStarted","Data":"2a9857a1cafc9ca6992987debd9f6ded9a8f2aa3b454a616ab0cf978dbd3b9ae"} Nov 22 08:13:16 crc kubenswrapper[4789]: I1122 08:13:16.735837 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-8667fbf6f6-k5kms" event={"ID":"5c6cfa2e-ba14-48da-a43d-6ea46a90bad1","Type":"ContainerStarted","Data":"b3c6b8f5d45c49c758af2323e8a8a4e494425f27124ea9bd012a71ec5b760b1c"} Nov 22 08:13:16 crc kubenswrapper[4789]: I1122 08:13:16.748237 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7879fb76fd-zn5xp" event={"ID":"490be79a-ce0c-4a3e-a819-7ef482fe453d","Type":"ContainerStarted","Data":"b44a30365cece20694037c10d7ef8165616f62d17108f38a487d3282b946a8b4"} Nov 22 08:13:16 crc kubenswrapper[4789]: I1122 08:13:16.748284 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7879fb76fd-zn5xp" event={"ID":"490be79a-ce0c-4a3e-a819-7ef482fe453d","Type":"ContainerStarted","Data":"bd96e678f9c97000eba0045985a50a40c51c26e59c651e2e42dd68689bc3c7b0"} Nov 22 08:13:16 crc kubenswrapper[4789]: E1122 08:13:16.750344 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:3ef72bbd7cce89ff54d850ff44ca6d7b2360834a502da3d561aeb6fd3d9af50a\\\"\"" pod="openstack-operators/keystone-operator-controller-manager-7879fb76fd-zn5xp" podUID="490be79a-ce0c-4a3e-a819-7ef482fe453d" Nov 22 08:13:16 crc kubenswrapper[4789]: I1122 08:13:16.757846 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-66b7d6f598-hn78m" event={"ID":"c4a79bb9-1d05-4c14-a668-6f63f521c98d","Type":"ContainerStarted","Data":"23bde6a9bf0e11f7077397b2c39978194b8ec74bf89e82e9e762add59897ecd8"} Nov 22 08:13:16 crc kubenswrapper[4789]: I1122 08:13:16.757911 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-66b7d6f598-hn78m" event={"ID":"c4a79bb9-1d05-4c14-a668-6f63f521c98d","Type":"ContainerStarted","Data":"806cbf437702e93135fde0b46d84848b30eec519cf2ab54cf8b1940724abe49d"} Nov 22 08:13:16 crc kubenswrapper[4789]: E1122 08:13:16.760800 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:207578cb433471cc1a79c21a808c8a15489d1d3c9fa77e29f3f697c33917fec6\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-66b7d6f598-hn78m" podUID="c4a79bb9-1d05-4c14-a668-6f63f521c98d" Nov 22 08:13:16 crc kubenswrapper[4789]: I1122 08:13:16.762523 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-5bdf4f7f7f-29chq" event={"ID":"4cbc0d4d-5b2e-4678-84e9-7eb69b3742e0","Type":"ContainerStarted","Data":"67b650741a16992f714758739b45e6291ca898bca1d86cbaaee1e24966e59aae"} Nov 22 08:13:16 crc kubenswrapper[4789]: I1122 08:13:16.767194 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-77vzm" event={"ID":"c94020a6-11c7-4ad5-90bb-8dd40e08ffd2","Type":"ContainerStarted","Data":"3561ca00da344c3672a8e0f225784e8bb3f416792b143020b967159a1a538d30"} Nov 22 08:13:16 crc kubenswrapper[4789]: I1122 08:13:16.779057 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-fzmzt" event={"ID":"03652378-7433-48c7-a601-aea6d67f1a14","Type":"ContainerStarted","Data":"58e3e48455d578bb042fa6a52c29799e60617db3769e0ee3f1503bd90c0d05f6"} Nov 22 08:13:16 crc kubenswrapper[4789]: I1122 08:13:16.779102 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-fzmzt" event={"ID":"03652378-7433-48c7-a601-aea6d67f1a14","Type":"ContainerStarted","Data":"c8500d6e6790c1a2f0a11441855f2a5426053a1b2752b7dfb57202bd155ae91b"} Nov 22 08:13:16 crc kubenswrapper[4789]: I1122 08:13:16.779282 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-fzmzt" Nov 22 08:13:16 crc kubenswrapper[4789]: I1122 08:13:16.783528 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-bf4c6585d-6shgh" event={"ID":"0cf958d5-222e-4598-a1a7-fe887644cd24","Type":"ContainerStarted","Data":"1d09cb8d15c7917dc1b29ed01619208fbe3893ffa5cb516b6bcd7324d96c51e7"} Nov 22 08:13:16 crc kubenswrapper[4789]: I1122 08:13:16.783565 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-bf4c6585d-6shgh" event={"ID":"0cf958d5-222e-4598-a1a7-fe887644cd24","Type":"ContainerStarted","Data":"6ba9216723546adf91f8b3ef5f285813068becf305cfd575d42d6cc6f4b7392f"} Nov 22 08:13:16 crc kubenswrapper[4789]: I1122 08:13:16.784817 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-86d796d84d-4lzpn" event={"ID":"2ec2f4f5-6ca9-4dae-ae2e-655ef780f3dd","Type":"ContainerStarted","Data":"ddf865273d20ac7f0a9afadcd48f8248bdae5dcf673ba8256d2386e13b96d163"} Nov 22 08:13:16 crc kubenswrapper[4789]: E1122 08:13:16.784822 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/heat-operator@sha256:5edd825a235f5784d9a65892763c5388c39df1731d0fcbf4ee33408b8c83ac96\\\"\"" pod="openstack-operators/heat-operator-controller-manager-bf4c6585d-6shgh" podUID="0cf958d5-222e-4598-a1a7-fe887644cd24" Nov 22 08:13:16 crc kubenswrapper[4789]: I1122 08:13:16.786955 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-6f8c5b86cb-rrcjv" event={"ID":"c4962710-54a6-4538-90ce-be3e27062015","Type":"ContainerStarted","Data":"156a4ee2521938812a90dfc5f2b9d67225c144c80a24bdbbfdfa21aa3a60fe90"} Nov 22 08:13:16 crc kubenswrapper[4789]: I1122 08:13:16.796766 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-799cb6ffd6-h272x" event={"ID":"0478f782-f1c5-412e-9964-749ee7f83c58","Type":"ContainerStarted","Data":"c63d7f7f3a0d54d6cc20dd8f039785c8d2e49190fa75bffe1a965ff20e8e54c8"} Nov 22 08:13:16 crc kubenswrapper[4789]: I1122 08:13:16.796808 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-799cb6ffd6-h272x" event={"ID":"0478f782-f1c5-412e-9964-749ee7f83c58","Type":"ContainerStarted","Data":"4b3a8242b36111b50ae798ee97fbb46c8f5bbd4a81719ca654fccc2880ef8921"} Nov 22 08:13:16 crc kubenswrapper[4789]: E1122 08:13:16.801550 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:c0b5f124a37c1538042c0e63f0978429572e2a851d7f3a6eb80de09b86d755a0\\\"\"" pod="openstack-operators/swift-operator-controller-manager-799cb6ffd6-h272x" podUID="0478f782-f1c5-412e-9964-749ee7f83c58" Nov 22 08:13:16 crc kubenswrapper[4789]: I1122 08:13:16.810766 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7bb88cb858-9g8lk" event={"ID":"50571e3f-4e49-47b6-aa7f-2c714e8ef88f","Type":"ContainerStarted","Data":"2aa047583be875a72278087d3b45d9cf66998fbc6b9385f47a564f2d47011a4a"} Nov 22 08:13:16 crc kubenswrapper[4789]: I1122 08:13:16.830206 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44kf66v" event={"ID":"d42b84d7-0575-401f-b556-fdaa161588c9","Type":"ContainerStarted","Data":"819c6e4ff09ccb7ed5f010e1a2f74d5e361a12e8a59838d9492934833c3b0bbb"} Nov 22 08:13:16 crc kubenswrapper[4789]: I1122 08:13:16.830251 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44kf66v" event={"ID":"d42b84d7-0575-401f-b556-fdaa161588c9","Type":"ContainerStarted","Data":"9ccdbb59f204768124a9939830be92c696373087177ff79225726c4cfaf16a18"} Nov 22 08:13:16 crc kubenswrapper[4789]: E1122 08:13:16.836851 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:78852f8ba332a5756c1551c126157f735279101a0fc3277ba4aa4db3478789dd\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44kf66v" podUID="d42b84d7-0575-401f-b556-fdaa161588c9" Nov 22 08:13:16 crc kubenswrapper[4789]: I1122 08:13:16.841036 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-8464cf66df-hc8lc" event={"ID":"50a6fd25-a186-4125-99c1-427d6214b39a","Type":"ContainerStarted","Data":"0a32a6879db01bf9780e16610fa1359f977fc0147852bffb86708470546ce377"} Nov 22 08:13:16 crc kubenswrapper[4789]: I1122 08:13:16.841082 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-8464cf66df-hc8lc" event={"ID":"50a6fd25-a186-4125-99c1-427d6214b39a","Type":"ContainerStarted","Data":"6c28791a936835e3dee64c24283ca9b86f7a7e771c346802cf472cf75057407f"} Nov 22 08:13:16 crc kubenswrapper[4789]: E1122 08:13:16.845406 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:82207e753574d4be246f86c4b074500d66cf20214aa80f0a8525cf3287a35e6d\\\"\"" pod="openstack-operators/test-operator-controller-manager-8464cf66df-hc8lc" podUID="50a6fd25-a186-4125-99c1-427d6214b39a" Nov 22 08:13:16 crc kubenswrapper[4789]: I1122 08:13:16.854410 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-769d9c7585-f7bd8" event={"ID":"84fafb83-aaff-4197-90d1-578c74da9299","Type":"ContainerStarted","Data":"8b0c99572d6c49a7a8db0c955f8c7bc81181358bafb6c2bce629736af69425f4"} Nov 22 08:13:16 crc kubenswrapper[4789]: I1122 08:13:16.855943 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-6fdc856c5d-mccpp" event={"ID":"bc429681-3e2a-4318-870d-ef62aef7d55b","Type":"ContainerStarted","Data":"09b105df130599fc333a1c64f418b28716dd2cf56ae6a4849774bc1be5101b09"} Nov 22 08:13:16 crc kubenswrapper[4789]: I1122 08:13:16.855963 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-6fdc856c5d-mccpp" event={"ID":"bc429681-3e2a-4318-870d-ef62aef7d55b","Type":"ContainerStarted","Data":"ffa42c160099dfc2e95f65912cdef590f45e5fe8492b525e66b5a6b1d683c37a"} Nov 22 08:13:16 crc kubenswrapper[4789]: E1122 08:13:16.857461 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:442c269d79163f8da75505019c02e9f0815837aaadcaddacb8e6c12df297ca13\\\"\"" pod="openstack-operators/octavia-operator-controller-manager-6fdc856c5d-mccpp" podUID="bc429681-3e2a-4318-870d-ef62aef7d55b" Nov 22 08:13:16 crc kubenswrapper[4789]: I1122 08:13:16.860890 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-7cd4fb6f79-kcq2g" event={"ID":"ca7faed1-038d-4ec8-800d-35848427a921","Type":"ContainerStarted","Data":"cf02bd05c5bbe4d180109d27b3bd66dcc8a11fe9de8d1e85f0f189935f844bb4"} Nov 22 08:13:16 crc kubenswrapper[4789]: I1122 08:13:16.860922 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-7cd4fb6f79-kcq2g" event={"ID":"ca7faed1-038d-4ec8-800d-35848427a921","Type":"ContainerStarted","Data":"68f7c138cec858b1c5722c36b13d45dc157998607ffc5b4670008953f9d5554d"} Nov 22 08:13:16 crc kubenswrapper[4789]: E1122 08:13:16.875142 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:4838402d41d42c56613d43dc5041aae475a2b18e6172491d6c4d4a78a580697f\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-7cd4fb6f79-kcq2g" podUID="ca7faed1-038d-4ec8-800d-35848427a921" Nov 22 08:13:16 crc kubenswrapper[4789]: I1122 08:13:16.894003 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-fzmzt" podStartSLOduration=2.893986716 podStartE2EDuration="2.893986716s" podCreationTimestamp="2025-11-22 08:13:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:13:16.892025492 +0000 UTC m=+1011.126425765" watchObservedRunningTime="2025-11-22 08:13:16.893986716 +0000 UTC m=+1011.128386989" Nov 22 08:13:17 crc kubenswrapper[4789]: I1122 08:13:17.874886 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-fzmzt" event={"ID":"03652378-7433-48c7-a601-aea6d67f1a14","Type":"ContainerStarted","Data":"acea833d2278395cc545eeba3af544e564bb39f8808adb201aa816a6274a7399"} Nov 22 08:13:17 crc kubenswrapper[4789]: E1122 08:13:17.877068 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:442c269d79163f8da75505019c02e9f0815837aaadcaddacb8e6c12df297ca13\\\"\"" pod="openstack-operators/octavia-operator-controller-manager-6fdc856c5d-mccpp" podUID="bc429681-3e2a-4318-870d-ef62aef7d55b" Nov 22 08:13:17 crc kubenswrapper[4789]: E1122 08:13:17.878500 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/heat-operator@sha256:5edd825a235f5784d9a65892763c5388c39df1731d0fcbf4ee33408b8c83ac96\\\"\"" pod="openstack-operators/heat-operator-controller-manager-bf4c6585d-6shgh" podUID="0cf958d5-222e-4598-a1a7-fe887644cd24" Nov 22 08:13:17 crc kubenswrapper[4789]: E1122 08:13:17.878644 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:78852f8ba332a5756c1551c126157f735279101a0fc3277ba4aa4db3478789dd\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44kf66v" podUID="d42b84d7-0575-401f-b556-fdaa161588c9" Nov 22 08:13:17 crc kubenswrapper[4789]: E1122 08:13:17.878710 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:c0b5f124a37c1538042c0e63f0978429572e2a851d7f3a6eb80de09b86d755a0\\\"\"" pod="openstack-operators/swift-operator-controller-manager-799cb6ffd6-h272x" podUID="0478f782-f1c5-412e-9964-749ee7f83c58" Nov 22 08:13:17 crc kubenswrapper[4789]: E1122 08:13:17.878891 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:4838402d41d42c56613d43dc5041aae475a2b18e6172491d6c4d4a78a580697f\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-7cd4fb6f79-kcq2g" podUID="ca7faed1-038d-4ec8-800d-35848427a921" Nov 22 08:13:17 crc kubenswrapper[4789]: E1122 08:13:17.878998 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:207578cb433471cc1a79c21a808c8a15489d1d3c9fa77e29f3f697c33917fec6\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-66b7d6f598-hn78m" podUID="c4a79bb9-1d05-4c14-a668-6f63f521c98d" Nov 22 08:13:17 crc kubenswrapper[4789]: E1122 08:13:17.879048 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:3ef72bbd7cce89ff54d850ff44ca6d7b2360834a502da3d561aeb6fd3d9af50a\\\"\"" pod="openstack-operators/keystone-operator-controller-manager-7879fb76fd-zn5xp" podUID="490be79a-ce0c-4a3e-a819-7ef482fe453d" Nov 22 08:13:17 crc kubenswrapper[4789]: E1122 08:13:17.879309 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:82207e753574d4be246f86c4b074500d66cf20214aa80f0a8525cf3287a35e6d\\\"\"" pod="openstack-operators/test-operator-controller-manager-8464cf66df-hc8lc" podUID="50a6fd25-a186-4125-99c1-427d6214b39a" Nov 22 08:13:25 crc kubenswrapper[4789]: I1122 08:13:25.764038 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-fzmzt" Nov 22 08:13:27 crc kubenswrapper[4789]: I1122 08:13:27.945352 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-5bdf4f7f7f-29chq" event={"ID":"4cbc0d4d-5b2e-4678-84e9-7eb69b3742e0","Type":"ContainerStarted","Data":"bb70aabd7595cb0d5bb3432556b48f7a814a2d808e0e561da7159cf19805b4be"} Nov 22 08:13:27 crc kubenswrapper[4789]: I1122 08:13:27.949645 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-77vzm" event={"ID":"c94020a6-11c7-4ad5-90bb-8dd40e08ffd2","Type":"ContainerStarted","Data":"456f2259a6bbb2ba284b6c58e744e1512675c2ad6935537ec9f63ad9a2c0432a"} Nov 22 08:13:27 crc kubenswrapper[4789]: I1122 08:13:27.960577 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-86d796d84d-4lzpn" event={"ID":"2ec2f4f5-6ca9-4dae-ae2e-655ef780f3dd","Type":"ContainerStarted","Data":"576eb37b11ace3e7c63d58095e1996d12cfefb06b7ae2576a505fefe7b3d343d"} Nov 22 08:13:27 crc kubenswrapper[4789]: I1122 08:13:27.960626 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-86d796d84d-4lzpn" event={"ID":"2ec2f4f5-6ca9-4dae-ae2e-655ef780f3dd","Type":"ContainerStarted","Data":"e4a3b4afb58c292f0544059dd08128582d3ba5f4c8ae0a9d1ae420769a0558d5"} Nov 22 08:13:27 crc kubenswrapper[4789]: I1122 08:13:27.961486 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-86d796d84d-4lzpn" Nov 22 08:13:27 crc kubenswrapper[4789]: I1122 08:13:27.980021 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-77vzm" podStartSLOduration=3.453924394 podStartE2EDuration="13.980002415s" podCreationTimestamp="2025-11-22 08:13:14 +0000 UTC" firstStartedPulling="2025-11-22 08:13:16.143466163 +0000 UTC m=+1010.377866436" lastFinishedPulling="2025-11-22 08:13:26.669544184 +0000 UTC m=+1020.903944457" observedRunningTime="2025-11-22 08:13:27.973054977 +0000 UTC m=+1022.207455260" watchObservedRunningTime="2025-11-22 08:13:27.980002415 +0000 UTC m=+1022.214402688" Nov 22 08:13:28 crc kubenswrapper[4789]: I1122 08:13:28.020684 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-86d796d84d-4lzpn" podStartSLOduration=4.200283376 podStartE2EDuration="15.016716825s" podCreationTimestamp="2025-11-22 08:13:13 +0000 UTC" firstStartedPulling="2025-11-22 08:13:15.825309789 +0000 UTC m=+1010.059710062" lastFinishedPulling="2025-11-22 08:13:26.641743238 +0000 UTC m=+1020.876143511" observedRunningTime="2025-11-22 08:13:28.01100896 +0000 UTC m=+1022.245409233" watchObservedRunningTime="2025-11-22 08:13:28.016716825 +0000 UTC m=+1022.251117098" Nov 22 08:13:28 crc kubenswrapper[4789]: I1122 08:13:28.022453 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-7768f8c84f-hwbvw" Nov 22 08:13:28 crc kubenswrapper[4789]: I1122 08:13:28.022588 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-6f8c5b86cb-rrcjv" Nov 22 08:13:28 crc kubenswrapper[4789]: I1122 08:13:28.022706 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-6dc664666c-p26lg" event={"ID":"a41e302b-afc7-45bc-8694-29e5ea87edc8","Type":"ContainerStarted","Data":"837608d55df1c49fe106068e496413e2cf0f4d92e757c68cafcf1c005f320586"} Nov 22 08:13:28 crc kubenswrapper[4789]: I1122 08:13:28.022837 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-8667fbf6f6-k5kms" event={"ID":"5c6cfa2e-ba14-48da-a43d-6ea46a90bad1","Type":"ContainerStarted","Data":"79f201b1311725769a09b981dcba8fa5e092f4629b6d748c3e8940a3045e33e0"} Nov 22 08:13:28 crc kubenswrapper[4789]: I1122 08:13:28.022935 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7bb88cb858-9g8lk" event={"ID":"50571e3f-4e49-47b6-aa7f-2c714e8ef88f","Type":"ContainerStarted","Data":"682a83f4adf6e9b6c3eb6c16a54ca9a6c1b260a031977a20fe05d49c5866f5bd"} Nov 22 08:13:28 crc kubenswrapper[4789]: I1122 08:13:28.023020 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-56dfb6b67f-p4vmr" event={"ID":"77bd1957-6cd6-418b-9b2c-ac879bbca15e","Type":"ContainerStarted","Data":"d36f8cc0ede72a2a55f98de300dfd3e012ebcac6805c6db567789bd75a597bde"} Nov 22 08:13:28 crc kubenswrapper[4789]: I1122 08:13:28.023097 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7768f8c84f-hwbvw" event={"ID":"93c2cbf6-757a-4de1-9f83-115787d74f31","Type":"ContainerStarted","Data":"b83d67acf008160bef3f950fb59109e118bcfb950c26766570b38f64df20c981"} Nov 22 08:13:28 crc kubenswrapper[4789]: I1122 08:13:28.023232 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7768f8c84f-hwbvw" event={"ID":"93c2cbf6-757a-4de1-9f83-115787d74f31","Type":"ContainerStarted","Data":"9f60177bd7110547c206d2bca8bf99fae2b2d679bda9ede120c6df28196f16c3"} Nov 22 08:13:28 crc kubenswrapper[4789]: I1122 08:13:28.023311 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-6f8c5b86cb-rrcjv" event={"ID":"c4962710-54a6-4538-90ce-be3e27062015","Type":"ContainerStarted","Data":"12b2950b5b8f140492f5f53ac1444dab70e53bc5224c2bdf19970a08a1ea9f57"} Nov 22 08:13:28 crc kubenswrapper[4789]: I1122 08:13:28.023393 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-6f8c5b86cb-rrcjv" event={"ID":"c4962710-54a6-4538-90ce-be3e27062015","Type":"ContainerStarted","Data":"7a1cd9387977d4b4465d166e3be80f71baab459060c84046ebf0562d8dfd1c2a"} Nov 22 08:13:28 crc kubenswrapper[4789]: I1122 08:13:28.044723 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5c75d7c94b-66wlz" event={"ID":"29d2daa3-90b3-4dea-9612-edc4cc0dcc30","Type":"ContainerStarted","Data":"823536eeb715f86d421ce6d6fe98a7bdfa7a225fb7e517b5d377f93e3ca06c2c"} Nov 22 08:13:28 crc kubenswrapper[4789]: I1122 08:13:28.045280 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-5c75d7c94b-66wlz" Nov 22 08:13:28 crc kubenswrapper[4789]: I1122 08:13:28.057111 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-7798859c74-j9vct" event={"ID":"e79f2bb5-95ac-4fbc-a065-da8067f25cb4","Type":"ContainerStarted","Data":"5ce65478ee7eec7890df27d0ef4c48fbd4d083dab317babfdf121f3f644587c4"} Nov 22 08:13:28 crc kubenswrapper[4789]: I1122 08:13:28.057172 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-7798859c74-j9vct" event={"ID":"e79f2bb5-95ac-4fbc-a065-da8067f25cb4","Type":"ContainerStarted","Data":"7d3f596ad045e028b839a9804d8583f37ad25ec72cbf4e048b5510978459e733"} Nov 22 08:13:28 crc kubenswrapper[4789]: I1122 08:13:28.057189 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-7798859c74-j9vct" Nov 22 08:13:28 crc kubenswrapper[4789]: I1122 08:13:28.062449 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-5d86b44686-nk4zk" event={"ID":"9e20512a-a64f-4f47-a7ef-b91f24390aef","Type":"ContainerStarted","Data":"90d98c0741a8aecdc88fe92dd72fb243e0bb8c45acabc3afe2315bfef40588cc"} Nov 22 08:13:28 crc kubenswrapper[4789]: I1122 08:13:28.062491 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-5d86b44686-nk4zk" event={"ID":"9e20512a-a64f-4f47-a7ef-b91f24390aef","Type":"ContainerStarted","Data":"cb7a664f64996d2db06e360a4d23c1a37d5f8ee5ff4b8daa12f4189e39e36edd"} Nov 22 08:13:28 crc kubenswrapper[4789]: I1122 08:13:28.063166 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-5d86b44686-nk4zk" Nov 22 08:13:28 crc kubenswrapper[4789]: I1122 08:13:28.068635 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-6f8c5b86cb-rrcjv" podStartSLOduration=4.336711318 podStartE2EDuration="15.068615596s" podCreationTimestamp="2025-11-22 08:13:13 +0000 UTC" firstStartedPulling="2025-11-22 08:13:15.835355133 +0000 UTC m=+1010.069755406" lastFinishedPulling="2025-11-22 08:13:26.567259411 +0000 UTC m=+1020.801659684" observedRunningTime="2025-11-22 08:13:28.050693239 +0000 UTC m=+1022.285093522" watchObservedRunningTime="2025-11-22 08:13:28.068615596 +0000 UTC m=+1022.303015869" Nov 22 08:13:28 crc kubenswrapper[4789]: I1122 08:13:28.071944 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6d8fd67bf7-2klns" event={"ID":"55d6892d-7d96-4518-8d36-58048cf96c52","Type":"ContainerStarted","Data":"a758fae11ce10b391dd6b70c5d8a254bdc7ec640e678f91f1299fdb4fd258e8d"} Nov 22 08:13:28 crc kubenswrapper[4789]: I1122 08:13:28.080243 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-7768f8c84f-hwbvw" podStartSLOduration=3.70216833 podStartE2EDuration="15.080223122s" podCreationTimestamp="2025-11-22 08:13:13 +0000 UTC" firstStartedPulling="2025-11-22 08:13:15.170331046 +0000 UTC m=+1009.404731319" lastFinishedPulling="2025-11-22 08:13:26.548385838 +0000 UTC m=+1020.782786111" observedRunningTime="2025-11-22 08:13:28.076637804 +0000 UTC m=+1022.311038087" watchObservedRunningTime="2025-11-22 08:13:28.080223122 +0000 UTC m=+1022.314623395" Nov 22 08:13:28 crc kubenswrapper[4789]: I1122 08:13:28.095105 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-769d9c7585-f7bd8" event={"ID":"84fafb83-aaff-4197-90d1-578c74da9299","Type":"ContainerStarted","Data":"d49be7326c5b072f3b0cd4d2432550c0c69f0e93a637c7978c9e490de4b4d6a0"} Nov 22 08:13:28 crc kubenswrapper[4789]: I1122 08:13:28.095162 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-769d9c7585-f7bd8" event={"ID":"84fafb83-aaff-4197-90d1-578c74da9299","Type":"ContainerStarted","Data":"71dd9b91544227e9a822dc64eeb99cf29591f7fd3de9e70d4002b48657a43150"} Nov 22 08:13:28 crc kubenswrapper[4789]: I1122 08:13:28.095925 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-769d9c7585-f7bd8" Nov 22 08:13:28 crc kubenswrapper[4789]: I1122 08:13:28.102324 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-5d86b44686-nk4zk" podStartSLOduration=4.062950862 podStartE2EDuration="15.102307143s" podCreationTimestamp="2025-11-22 08:13:13 +0000 UTC" firstStartedPulling="2025-11-22 08:13:15.508946825 +0000 UTC m=+1009.743347098" lastFinishedPulling="2025-11-22 08:13:26.548303106 +0000 UTC m=+1020.782703379" observedRunningTime="2025-11-22 08:13:28.099876456 +0000 UTC m=+1022.334276739" watchObservedRunningTime="2025-11-22 08:13:28.102307143 +0000 UTC m=+1022.336707416" Nov 22 08:13:28 crc kubenswrapper[4789]: I1122 08:13:28.128846 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-7798859c74-j9vct" podStartSLOduration=3.322203434 podStartE2EDuration="14.128826204s" podCreationTimestamp="2025-11-22 08:13:14 +0000 UTC" firstStartedPulling="2025-11-22 08:13:15.833293937 +0000 UTC m=+1010.067694210" lastFinishedPulling="2025-11-22 08:13:26.639916707 +0000 UTC m=+1020.874316980" observedRunningTime="2025-11-22 08:13:28.125771581 +0000 UTC m=+1022.360171864" watchObservedRunningTime="2025-11-22 08:13:28.128826204 +0000 UTC m=+1022.363226487" Nov 22 08:13:28 crc kubenswrapper[4789]: I1122 08:13:28.165467 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-5c75d7c94b-66wlz" podStartSLOduration=4.031609659 podStartE2EDuration="15.16544211s" podCreationTimestamp="2025-11-22 08:13:13 +0000 UTC" firstStartedPulling="2025-11-22 08:13:15.506662313 +0000 UTC m=+1009.741062586" lastFinishedPulling="2025-11-22 08:13:26.640494764 +0000 UTC m=+1020.874895037" observedRunningTime="2025-11-22 08:13:28.149354352 +0000 UTC m=+1022.383754645" watchObservedRunningTime="2025-11-22 08:13:28.16544211 +0000 UTC m=+1022.399842383" Nov 22 08:13:28 crc kubenswrapper[4789]: I1122 08:13:28.184298 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-769d9c7585-f7bd8" podStartSLOduration=4.593541001 podStartE2EDuration="15.184279271s" podCreationTimestamp="2025-11-22 08:13:13 +0000 UTC" firstStartedPulling="2025-11-22 08:13:15.982273779 +0000 UTC m=+1010.216674042" lastFinishedPulling="2025-11-22 08:13:26.573012039 +0000 UTC m=+1020.807412312" observedRunningTime="2025-11-22 08:13:28.180159399 +0000 UTC m=+1022.414559672" watchObservedRunningTime="2025-11-22 08:13:28.184279271 +0000 UTC m=+1022.418679544" Nov 22 08:13:29 crc kubenswrapper[4789]: I1122 08:13:29.104373 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-8667fbf6f6-k5kms" event={"ID":"5c6cfa2e-ba14-48da-a43d-6ea46a90bad1","Type":"ContainerStarted","Data":"a3af124de76788ca64c3fd827758acd3d28e9defd7e421fc414bbbcbe4d9c3fd"} Nov 22 08:13:29 crc kubenswrapper[4789]: I1122 08:13:29.105281 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-8667fbf6f6-k5kms" Nov 22 08:13:29 crc kubenswrapper[4789]: I1122 08:13:29.107038 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6d8fd67bf7-2klns" event={"ID":"55d6892d-7d96-4518-8d36-58048cf96c52","Type":"ContainerStarted","Data":"b80166db6b524f3cec4e61d120cab8d61deed38af07b15dd9e386809076fefa8"} Nov 22 08:13:29 crc kubenswrapper[4789]: I1122 08:13:29.107314 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-6d8fd67bf7-2klns" Nov 22 08:13:29 crc kubenswrapper[4789]: I1122 08:13:29.109160 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7bb88cb858-9g8lk" event={"ID":"50571e3f-4e49-47b6-aa7f-2c714e8ef88f","Type":"ContainerStarted","Data":"54091d0e047f24f0235043f38c93ee1e52dc9cabd6d8e9cf8a10dbd8eb5ce1d7"} Nov 22 08:13:29 crc kubenswrapper[4789]: I1122 08:13:29.109375 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-7bb88cb858-9g8lk" Nov 22 08:13:29 crc kubenswrapper[4789]: I1122 08:13:29.111116 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5c75d7c94b-66wlz" event={"ID":"29d2daa3-90b3-4dea-9612-edc4cc0dcc30","Type":"ContainerStarted","Data":"b1fbde87d90d21f47795153db013017122da5f04cfbd1a033aa32eb65ba0daa8"} Nov 22 08:13:29 crc kubenswrapper[4789]: I1122 08:13:29.113197 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-5bdf4f7f7f-29chq" event={"ID":"4cbc0d4d-5b2e-4678-84e9-7eb69b3742e0","Type":"ContainerStarted","Data":"52413f23adbdb5d673b99d563feb775f9848c8d9d6cee6c3bfb732cc5b6f5a41"} Nov 22 08:13:29 crc kubenswrapper[4789]: I1122 08:13:29.113465 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-5bdf4f7f7f-29chq" Nov 22 08:13:29 crc kubenswrapper[4789]: I1122 08:13:29.115176 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-56dfb6b67f-p4vmr" event={"ID":"77bd1957-6cd6-418b-9b2c-ac879bbca15e","Type":"ContainerStarted","Data":"3f2ba11208dfac8afb9abd032e8978b4ec9aeb32c4d63809c55d0b03661ec8b9"} Nov 22 08:13:29 crc kubenswrapper[4789]: I1122 08:13:29.115310 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-56dfb6b67f-p4vmr" Nov 22 08:13:29 crc kubenswrapper[4789]: I1122 08:13:29.117299 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-6dc664666c-p26lg" event={"ID":"a41e302b-afc7-45bc-8694-29e5ea87edc8","Type":"ContainerStarted","Data":"b6c80e1fc549d03ab299f12f58d7645969f5c2340caca9da4f7cc0e303ec0b0d"} Nov 22 08:13:29 crc kubenswrapper[4789]: I1122 08:13:29.136498 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-8667fbf6f6-k5kms" podStartSLOduration=5.404135079 podStartE2EDuration="16.136472529s" podCreationTimestamp="2025-11-22 08:13:13 +0000 UTC" firstStartedPulling="2025-11-22 08:13:15.816986184 +0000 UTC m=+1010.051386457" lastFinishedPulling="2025-11-22 08:13:26.549323624 +0000 UTC m=+1020.783723907" observedRunningTime="2025-11-22 08:13:29.129382517 +0000 UTC m=+1023.363782800" watchObservedRunningTime="2025-11-22 08:13:29.136472529 +0000 UTC m=+1023.370872802" Nov 22 08:13:29 crc kubenswrapper[4789]: I1122 08:13:29.150125 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-6d8fd67bf7-2klns" podStartSLOduration=5.01333596 podStartE2EDuration="16.15010145s" podCreationTimestamp="2025-11-22 08:13:13 +0000 UTC" firstStartedPulling="2025-11-22 08:13:15.508681308 +0000 UTC m=+1009.743081581" lastFinishedPulling="2025-11-22 08:13:26.645446798 +0000 UTC m=+1020.879847071" observedRunningTime="2025-11-22 08:13:29.147231172 +0000 UTC m=+1023.381631445" watchObservedRunningTime="2025-11-22 08:13:29.15010145 +0000 UTC m=+1023.384501723" Nov 22 08:13:29 crc kubenswrapper[4789]: I1122 08:13:29.169977 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-7bb88cb858-9g8lk" podStartSLOduration=5.444190809 podStartE2EDuration="16.16995959s" podCreationTimestamp="2025-11-22 08:13:13 +0000 UTC" firstStartedPulling="2025-11-22 08:13:15.856774256 +0000 UTC m=+1010.091174529" lastFinishedPulling="2025-11-22 08:13:26.582543037 +0000 UTC m=+1020.816943310" observedRunningTime="2025-11-22 08:13:29.162920329 +0000 UTC m=+1023.397320622" watchObservedRunningTime="2025-11-22 08:13:29.16995959 +0000 UTC m=+1023.404359853" Nov 22 08:13:29 crc kubenswrapper[4789]: I1122 08:13:29.187133 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-6dc664666c-p26lg" podStartSLOduration=4.43687235 podStartE2EDuration="15.187108707s" podCreationTimestamp="2025-11-22 08:13:14 +0000 UTC" firstStartedPulling="2025-11-22 08:13:15.840878144 +0000 UTC m=+1010.075278417" lastFinishedPulling="2025-11-22 08:13:26.591114501 +0000 UTC m=+1020.825514774" observedRunningTime="2025-11-22 08:13:29.178778441 +0000 UTC m=+1023.413178734" watchObservedRunningTime="2025-11-22 08:13:29.187108707 +0000 UTC m=+1023.421508980" Nov 22 08:13:29 crc kubenswrapper[4789]: I1122 08:13:29.200311 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-5bdf4f7f7f-29chq" podStartSLOduration=4.374686178 podStartE2EDuration="15.200290255s" podCreationTimestamp="2025-11-22 08:13:14 +0000 UTC" firstStartedPulling="2025-11-22 08:13:15.815436681 +0000 UTC m=+1010.049836954" lastFinishedPulling="2025-11-22 08:13:26.641040758 +0000 UTC m=+1020.875441031" observedRunningTime="2025-11-22 08:13:29.199919056 +0000 UTC m=+1023.434319349" watchObservedRunningTime="2025-11-22 08:13:29.200290255 +0000 UTC m=+1023.434690548" Nov 22 08:13:29 crc kubenswrapper[4789]: I1122 08:13:29.220224 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-56dfb6b67f-p4vmr" podStartSLOduration=5.110827201 podStartE2EDuration="16.220203647s" podCreationTimestamp="2025-11-22 08:13:13 +0000 UTC" firstStartedPulling="2025-11-22 08:13:15.529459803 +0000 UTC m=+1009.763860076" lastFinishedPulling="2025-11-22 08:13:26.638836249 +0000 UTC m=+1020.873236522" observedRunningTime="2025-11-22 08:13:29.214384519 +0000 UTC m=+1023.448784792" watchObservedRunningTime="2025-11-22 08:13:29.220203647 +0000 UTC m=+1023.454603920" Nov 22 08:13:30 crc kubenswrapper[4789]: I1122 08:13:30.133262 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-6dc664666c-p26lg" Nov 22 08:13:34 crc kubenswrapper[4789]: I1122 08:13:34.116430 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-7768f8c84f-hwbvw" Nov 22 08:13:34 crc kubenswrapper[4789]: I1122 08:13:34.126548 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-6d8fd67bf7-2klns" Nov 22 08:13:34 crc kubenswrapper[4789]: I1122 08:13:34.133990 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-56dfb6b67f-p4vmr" Nov 22 08:13:34 crc kubenswrapper[4789]: I1122 08:13:34.169066 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-8667fbf6f6-k5kms" Nov 22 08:13:34 crc kubenswrapper[4789]: I1122 08:13:34.208209 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-5d86b44686-nk4zk" Nov 22 08:13:34 crc kubenswrapper[4789]: I1122 08:13:34.258492 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-5c75d7c94b-66wlz" Nov 22 08:13:34 crc kubenswrapper[4789]: I1122 08:13:34.410811 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-7bb88cb858-9g8lk" Nov 22 08:13:34 crc kubenswrapper[4789]: I1122 08:13:34.450136 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-6f8c5b86cb-rrcjv" Nov 22 08:13:34 crc kubenswrapper[4789]: I1122 08:13:34.528827 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-86d796d84d-4lzpn" Nov 22 08:13:34 crc kubenswrapper[4789]: I1122 08:13:34.575105 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-5bdf4f7f7f-29chq" Nov 22 08:13:34 crc kubenswrapper[4789]: I1122 08:13:34.633290 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-6dc664666c-p26lg" Nov 22 08:13:34 crc kubenswrapper[4789]: I1122 08:13:34.789609 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-7798859c74-j9vct" Nov 22 08:13:34 crc kubenswrapper[4789]: I1122 08:13:34.894901 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-769d9c7585-f7bd8" Nov 22 08:13:42 crc kubenswrapper[4789]: I1122 08:13:42.247425 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-6fdc856c5d-mccpp" event={"ID":"bc429681-3e2a-4318-870d-ef62aef7d55b","Type":"ContainerStarted","Data":"820296b85bca4a04119489c624a894e32950496ff9edd567016c78ee8bf73273"} Nov 22 08:13:42 crc kubenswrapper[4789]: I1122 08:13:42.248060 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-6fdc856c5d-mccpp" Nov 22 08:13:42 crc kubenswrapper[4789]: I1122 08:13:42.250436 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44kf66v" event={"ID":"d42b84d7-0575-401f-b556-fdaa161588c9","Type":"ContainerStarted","Data":"6805ae0de738ab627d98fb7d955d4c1986649ab57d3ef664da70a9d28801e70e"} Nov 22 08:13:42 crc kubenswrapper[4789]: I1122 08:13:42.250615 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44kf66v" Nov 22 08:13:42 crc kubenswrapper[4789]: I1122 08:13:42.252724 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-8464cf66df-hc8lc" event={"ID":"50a6fd25-a186-4125-99c1-427d6214b39a","Type":"ContainerStarted","Data":"91b0dd2990b97f4ac5d5a357acb0573fd640d2da6303f19ca22866fb509567c8"} Nov 22 08:13:42 crc kubenswrapper[4789]: I1122 08:13:42.252940 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-8464cf66df-hc8lc" Nov 22 08:13:42 crc kubenswrapper[4789]: I1122 08:13:42.254314 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-7cd4fb6f79-kcq2g" event={"ID":"ca7faed1-038d-4ec8-800d-35848427a921","Type":"ContainerStarted","Data":"295577efaccdb51834d9a215d89756413cfd34c48fe9f95850f4385c97bb15d9"} Nov 22 08:13:42 crc kubenswrapper[4789]: I1122 08:13:42.254458 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-7cd4fb6f79-kcq2g" Nov 22 08:13:42 crc kubenswrapper[4789]: I1122 08:13:42.255694 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-bf4c6585d-6shgh" event={"ID":"0cf958d5-222e-4598-a1a7-fe887644cd24","Type":"ContainerStarted","Data":"1d58ee2b5851cb231dd88308d7a42074df93523a084ee54d976491a856ff8f04"} Nov 22 08:13:42 crc kubenswrapper[4789]: I1122 08:13:42.255862 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-bf4c6585d-6shgh" Nov 22 08:13:42 crc kubenswrapper[4789]: I1122 08:13:42.257295 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7879fb76fd-zn5xp" event={"ID":"490be79a-ce0c-4a3e-a819-7ef482fe453d","Type":"ContainerStarted","Data":"2f810b694c53d30163949700daae5664b48a083f3519a15955830cc364f2e1fe"} Nov 22 08:13:42 crc kubenswrapper[4789]: I1122 08:13:42.257427 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-7879fb76fd-zn5xp" Nov 22 08:13:42 crc kubenswrapper[4789]: I1122 08:13:42.259154 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-66b7d6f598-hn78m" event={"ID":"c4a79bb9-1d05-4c14-a668-6f63f521c98d","Type":"ContainerStarted","Data":"308436a8374b0f1c270b9f7c8c6d9666a3c21df073d6ce1fedee6d62dfeea9c8"} Nov 22 08:13:42 crc kubenswrapper[4789]: I1122 08:13:42.259317 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-66b7d6f598-hn78m" Nov 22 08:13:42 crc kubenswrapper[4789]: I1122 08:13:42.260816 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-799cb6ffd6-h272x" event={"ID":"0478f782-f1c5-412e-9964-749ee7f83c58","Type":"ContainerStarted","Data":"d8a661fbbc9b780a6941f860088ac589c80c72ac5cd6fb70243fca3da3f0c30a"} Nov 22 08:13:42 crc kubenswrapper[4789]: I1122 08:13:42.261001 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-799cb6ffd6-h272x" Nov 22 08:13:42 crc kubenswrapper[4789]: I1122 08:13:42.282436 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-6fdc856c5d-mccpp" podStartSLOduration=3.543841573 podStartE2EDuration="29.282414126s" podCreationTimestamp="2025-11-22 08:13:13 +0000 UTC" firstStartedPulling="2025-11-22 08:13:15.891232103 +0000 UTC m=+1010.125632376" lastFinishedPulling="2025-11-22 08:13:41.629804656 +0000 UTC m=+1035.864204929" observedRunningTime="2025-11-22 08:13:42.276107115 +0000 UTC m=+1036.510507388" watchObservedRunningTime="2025-11-22 08:13:42.282414126 +0000 UTC m=+1036.516814399" Nov 22 08:13:42 crc kubenswrapper[4789]: I1122 08:13:42.298556 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-8464cf66df-hc8lc" podStartSLOduration=2.559786506 podStartE2EDuration="28.298535705s" podCreationTimestamp="2025-11-22 08:13:14 +0000 UTC" firstStartedPulling="2025-11-22 08:13:15.891058197 +0000 UTC m=+1010.125458470" lastFinishedPulling="2025-11-22 08:13:41.629807396 +0000 UTC m=+1035.864207669" observedRunningTime="2025-11-22 08:13:42.295578024 +0000 UTC m=+1036.529978297" watchObservedRunningTime="2025-11-22 08:13:42.298535705 +0000 UTC m=+1036.532935978" Nov 22 08:13:42 crc kubenswrapper[4789]: I1122 08:13:42.350917 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44kf66v" podStartSLOduration=2.808912362 podStartE2EDuration="28.350900089s" podCreationTimestamp="2025-11-22 08:13:14 +0000 UTC" firstStartedPulling="2025-11-22 08:13:16.086307678 +0000 UTC m=+1010.320707951" lastFinishedPulling="2025-11-22 08:13:41.628295405 +0000 UTC m=+1035.862695678" observedRunningTime="2025-11-22 08:13:42.350388255 +0000 UTC m=+1036.584788528" watchObservedRunningTime="2025-11-22 08:13:42.350900089 +0000 UTC m=+1036.585300352" Nov 22 08:13:42 crc kubenswrapper[4789]: I1122 08:13:42.442148 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-7879fb76fd-zn5xp" podStartSLOduration=3.680415837 podStartE2EDuration="29.4421256s" podCreationTimestamp="2025-11-22 08:13:13 +0000 UTC" firstStartedPulling="2025-11-22 08:13:15.870364455 +0000 UTC m=+1010.104764728" lastFinishedPulling="2025-11-22 08:13:41.632074218 +0000 UTC m=+1035.866474491" observedRunningTime="2025-11-22 08:13:42.401767103 +0000 UTC m=+1036.636167386" watchObservedRunningTime="2025-11-22 08:13:42.4421256 +0000 UTC m=+1036.676525873" Nov 22 08:13:42 crc kubenswrapper[4789]: I1122 08:13:42.446817 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-799cb6ffd6-h272x" podStartSLOduration=2.679931414 podStartE2EDuration="28.446797507s" podCreationTimestamp="2025-11-22 08:13:14 +0000 UTC" firstStartedPulling="2025-11-22 08:13:15.861236757 +0000 UTC m=+1010.095637030" lastFinishedPulling="2025-11-22 08:13:41.62810285 +0000 UTC m=+1035.862503123" observedRunningTime="2025-11-22 08:13:42.439150049 +0000 UTC m=+1036.673550352" watchObservedRunningTime="2025-11-22 08:13:42.446797507 +0000 UTC m=+1036.681197790" Nov 22 08:13:42 crc kubenswrapper[4789]: I1122 08:13:42.467339 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-7cd4fb6f79-kcq2g" podStartSLOduration=2.7838033490000003 podStartE2EDuration="28.467324165s" podCreationTimestamp="2025-11-22 08:13:14 +0000 UTC" firstStartedPulling="2025-11-22 08:13:16.087247334 +0000 UTC m=+1010.321647607" lastFinishedPulling="2025-11-22 08:13:41.77076814 +0000 UTC m=+1036.005168423" observedRunningTime="2025-11-22 08:13:42.464169809 +0000 UTC m=+1036.698570102" watchObservedRunningTime="2025-11-22 08:13:42.467324165 +0000 UTC m=+1036.701724428" Nov 22 08:13:42 crc kubenswrapper[4789]: I1122 08:13:42.498481 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-bf4c6585d-6shgh" podStartSLOduration=3.745780896 podStartE2EDuration="29.498466823s" podCreationTimestamp="2025-11-22 08:13:13 +0000 UTC" firstStartedPulling="2025-11-22 08:13:15.874604411 +0000 UTC m=+1010.109004684" lastFinishedPulling="2025-11-22 08:13:41.627290338 +0000 UTC m=+1035.861690611" observedRunningTime="2025-11-22 08:13:42.49212264 +0000 UTC m=+1036.726522913" watchObservedRunningTime="2025-11-22 08:13:42.498466823 +0000 UTC m=+1036.732867096" Nov 22 08:13:42 crc kubenswrapper[4789]: I1122 08:13:42.537582 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-66b7d6f598-hn78m" podStartSLOduration=3.799196059 podStartE2EDuration="29.537563626s" podCreationTimestamp="2025-11-22 08:13:13 +0000 UTC" firstStartedPulling="2025-11-22 08:13:15.891409578 +0000 UTC m=+1010.125809851" lastFinishedPulling="2025-11-22 08:13:41.629777135 +0000 UTC m=+1035.864177418" observedRunningTime="2025-11-22 08:13:42.536931478 +0000 UTC m=+1036.771331751" watchObservedRunningTime="2025-11-22 08:13:42.537563626 +0000 UTC m=+1036.771963899" Nov 22 08:13:54 crc kubenswrapper[4789]: I1122 08:13:54.188829 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-bf4c6585d-6shgh" Nov 22 08:13:54 crc kubenswrapper[4789]: I1122 08:13:54.300067 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-7879fb76fd-zn5xp" Nov 22 08:13:54 crc kubenswrapper[4789]: I1122 08:13:54.507401 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-66b7d6f598-hn78m" Nov 22 08:13:54 crc kubenswrapper[4789]: I1122 08:13:54.542396 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-6fdc856c5d-mccpp" Nov 22 08:13:54 crc kubenswrapper[4789]: I1122 08:13:54.711698 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-799cb6ffd6-h272x" Nov 22 08:13:54 crc kubenswrapper[4789]: I1122 08:13:54.831156 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-8464cf66df-hc8lc" Nov 22 08:13:55 crc kubenswrapper[4789]: I1122 08:13:55.158511 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-7cd4fb6f79-kcq2g" Nov 22 08:13:55 crc kubenswrapper[4789]: I1122 08:13:55.202396 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44kf66v" Nov 22 08:14:10 crc kubenswrapper[4789]: I1122 08:14:10.283743 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7bdd77c89-nq45j"] Nov 22 08:14:10 crc kubenswrapper[4789]: I1122 08:14:10.285565 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bdd77c89-nq45j" Nov 22 08:14:10 crc kubenswrapper[4789]: I1122 08:14:10.288023 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Nov 22 08:14:10 crc kubenswrapper[4789]: I1122 08:14:10.288180 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Nov 22 08:14:10 crc kubenswrapper[4789]: I1122 08:14:10.288280 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Nov 22 08:14:10 crc kubenswrapper[4789]: I1122 08:14:10.288774 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-9bmdf" Nov 22 08:14:10 crc kubenswrapper[4789]: I1122 08:14:10.299812 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7bdd77c89-nq45j"] Nov 22 08:14:10 crc kubenswrapper[4789]: I1122 08:14:10.315611 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b4x4q\" (UniqueName: \"kubernetes.io/projected/f235f4f5-490c-411c-9a1d-452bae11b2eb-kube-api-access-b4x4q\") pod \"dnsmasq-dns-7bdd77c89-nq45j\" (UID: \"f235f4f5-490c-411c-9a1d-452bae11b2eb\") " pod="openstack/dnsmasq-dns-7bdd77c89-nq45j" Nov 22 08:14:10 crc kubenswrapper[4789]: I1122 08:14:10.315805 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f235f4f5-490c-411c-9a1d-452bae11b2eb-config\") pod \"dnsmasq-dns-7bdd77c89-nq45j\" (UID: \"f235f4f5-490c-411c-9a1d-452bae11b2eb\") " pod="openstack/dnsmasq-dns-7bdd77c89-nq45j" Nov 22 08:14:10 crc kubenswrapper[4789]: I1122 08:14:10.335002 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6584b49599-wzdnt"] Nov 22 08:14:10 crc kubenswrapper[4789]: I1122 08:14:10.336258 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6584b49599-wzdnt" Nov 22 08:14:10 crc kubenswrapper[4789]: I1122 08:14:10.342916 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Nov 22 08:14:10 crc kubenswrapper[4789]: I1122 08:14:10.353255 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6584b49599-wzdnt"] Nov 22 08:14:10 crc kubenswrapper[4789]: I1122 08:14:10.417538 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/79936918-3fbc-41b6-9c9c-31757e4a6fc6-dns-svc\") pod \"dnsmasq-dns-6584b49599-wzdnt\" (UID: \"79936918-3fbc-41b6-9c9c-31757e4a6fc6\") " pod="openstack/dnsmasq-dns-6584b49599-wzdnt" Nov 22 08:14:10 crc kubenswrapper[4789]: I1122 08:14:10.417604 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b4x4q\" (UniqueName: \"kubernetes.io/projected/f235f4f5-490c-411c-9a1d-452bae11b2eb-kube-api-access-b4x4q\") pod \"dnsmasq-dns-7bdd77c89-nq45j\" (UID: \"f235f4f5-490c-411c-9a1d-452bae11b2eb\") " pod="openstack/dnsmasq-dns-7bdd77c89-nq45j" Nov 22 08:14:10 crc kubenswrapper[4789]: I1122 08:14:10.417726 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f235f4f5-490c-411c-9a1d-452bae11b2eb-config\") pod \"dnsmasq-dns-7bdd77c89-nq45j\" (UID: \"f235f4f5-490c-411c-9a1d-452bae11b2eb\") " pod="openstack/dnsmasq-dns-7bdd77c89-nq45j" Nov 22 08:14:10 crc kubenswrapper[4789]: I1122 08:14:10.417821 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/79936918-3fbc-41b6-9c9c-31757e4a6fc6-config\") pod \"dnsmasq-dns-6584b49599-wzdnt\" (UID: \"79936918-3fbc-41b6-9c9c-31757e4a6fc6\") " pod="openstack/dnsmasq-dns-6584b49599-wzdnt" Nov 22 08:14:10 crc kubenswrapper[4789]: I1122 08:14:10.417867 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c6zvz\" (UniqueName: \"kubernetes.io/projected/79936918-3fbc-41b6-9c9c-31757e4a6fc6-kube-api-access-c6zvz\") pod \"dnsmasq-dns-6584b49599-wzdnt\" (UID: \"79936918-3fbc-41b6-9c9c-31757e4a6fc6\") " pod="openstack/dnsmasq-dns-6584b49599-wzdnt" Nov 22 08:14:10 crc kubenswrapper[4789]: I1122 08:14:10.418701 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f235f4f5-490c-411c-9a1d-452bae11b2eb-config\") pod \"dnsmasq-dns-7bdd77c89-nq45j\" (UID: \"f235f4f5-490c-411c-9a1d-452bae11b2eb\") " pod="openstack/dnsmasq-dns-7bdd77c89-nq45j" Nov 22 08:14:10 crc kubenswrapper[4789]: I1122 08:14:10.435623 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b4x4q\" (UniqueName: \"kubernetes.io/projected/f235f4f5-490c-411c-9a1d-452bae11b2eb-kube-api-access-b4x4q\") pod \"dnsmasq-dns-7bdd77c89-nq45j\" (UID: \"f235f4f5-490c-411c-9a1d-452bae11b2eb\") " pod="openstack/dnsmasq-dns-7bdd77c89-nq45j" Nov 22 08:14:10 crc kubenswrapper[4789]: I1122 08:14:10.518875 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/79936918-3fbc-41b6-9c9c-31757e4a6fc6-config\") pod \"dnsmasq-dns-6584b49599-wzdnt\" (UID: \"79936918-3fbc-41b6-9c9c-31757e4a6fc6\") " pod="openstack/dnsmasq-dns-6584b49599-wzdnt" Nov 22 08:14:10 crc kubenswrapper[4789]: I1122 08:14:10.518928 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c6zvz\" (UniqueName: \"kubernetes.io/projected/79936918-3fbc-41b6-9c9c-31757e4a6fc6-kube-api-access-c6zvz\") pod \"dnsmasq-dns-6584b49599-wzdnt\" (UID: \"79936918-3fbc-41b6-9c9c-31757e4a6fc6\") " pod="openstack/dnsmasq-dns-6584b49599-wzdnt" Nov 22 08:14:10 crc kubenswrapper[4789]: I1122 08:14:10.518992 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/79936918-3fbc-41b6-9c9c-31757e4a6fc6-dns-svc\") pod \"dnsmasq-dns-6584b49599-wzdnt\" (UID: \"79936918-3fbc-41b6-9c9c-31757e4a6fc6\") " pod="openstack/dnsmasq-dns-6584b49599-wzdnt" Nov 22 08:14:10 crc kubenswrapper[4789]: I1122 08:14:10.519775 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/79936918-3fbc-41b6-9c9c-31757e4a6fc6-config\") pod \"dnsmasq-dns-6584b49599-wzdnt\" (UID: \"79936918-3fbc-41b6-9c9c-31757e4a6fc6\") " pod="openstack/dnsmasq-dns-6584b49599-wzdnt" Nov 22 08:14:10 crc kubenswrapper[4789]: I1122 08:14:10.519902 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/79936918-3fbc-41b6-9c9c-31757e4a6fc6-dns-svc\") pod \"dnsmasq-dns-6584b49599-wzdnt\" (UID: \"79936918-3fbc-41b6-9c9c-31757e4a6fc6\") " pod="openstack/dnsmasq-dns-6584b49599-wzdnt" Nov 22 08:14:10 crc kubenswrapper[4789]: I1122 08:14:10.538945 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c6zvz\" (UniqueName: \"kubernetes.io/projected/79936918-3fbc-41b6-9c9c-31757e4a6fc6-kube-api-access-c6zvz\") pod \"dnsmasq-dns-6584b49599-wzdnt\" (UID: \"79936918-3fbc-41b6-9c9c-31757e4a6fc6\") " pod="openstack/dnsmasq-dns-6584b49599-wzdnt" Nov 22 08:14:10 crc kubenswrapper[4789]: I1122 08:14:10.603556 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bdd77c89-nq45j" Nov 22 08:14:10 crc kubenswrapper[4789]: I1122 08:14:10.650497 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6584b49599-wzdnt" Nov 22 08:14:11 crc kubenswrapper[4789]: I1122 08:14:11.055602 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7bdd77c89-nq45j"] Nov 22 08:14:11 crc kubenswrapper[4789]: I1122 08:14:11.066623 4789 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 22 08:14:11 crc kubenswrapper[4789]: I1122 08:14:11.132804 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6584b49599-wzdnt"] Nov 22 08:14:11 crc kubenswrapper[4789]: I1122 08:14:11.458647 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6584b49599-wzdnt" event={"ID":"79936918-3fbc-41b6-9c9c-31757e4a6fc6","Type":"ContainerStarted","Data":"bb243ce6a40860b4c430b1230bf4b32fa8e8c774a4f42356768a69b1b12a6d7a"} Nov 22 08:14:11 crc kubenswrapper[4789]: I1122 08:14:11.460141 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bdd77c89-nq45j" event={"ID":"f235f4f5-490c-411c-9a1d-452bae11b2eb","Type":"ContainerStarted","Data":"e53dd5d61e2512f60bcff635d2a631f5e0eddf6a64deca3b00f3780d16962654"} Nov 22 08:14:13 crc kubenswrapper[4789]: I1122 08:14:13.181790 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6584b49599-wzdnt"] Nov 22 08:14:13 crc kubenswrapper[4789]: I1122 08:14:13.205955 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7c6d9948dc-2dpcz"] Nov 22 08:14:13 crc kubenswrapper[4789]: I1122 08:14:13.207287 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c6d9948dc-2dpcz" Nov 22 08:14:13 crc kubenswrapper[4789]: I1122 08:14:13.220703 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7c6d9948dc-2dpcz"] Nov 22 08:14:13 crc kubenswrapper[4789]: I1122 08:14:13.354668 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e32fe7bc-c0c8-443b-af3e-d9cc58db4f24-dns-svc\") pod \"dnsmasq-dns-7c6d9948dc-2dpcz\" (UID: \"e32fe7bc-c0c8-443b-af3e-d9cc58db4f24\") " pod="openstack/dnsmasq-dns-7c6d9948dc-2dpcz" Nov 22 08:14:13 crc kubenswrapper[4789]: I1122 08:14:13.355025 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-llqn6\" (UniqueName: \"kubernetes.io/projected/e32fe7bc-c0c8-443b-af3e-d9cc58db4f24-kube-api-access-llqn6\") pod \"dnsmasq-dns-7c6d9948dc-2dpcz\" (UID: \"e32fe7bc-c0c8-443b-af3e-d9cc58db4f24\") " pod="openstack/dnsmasq-dns-7c6d9948dc-2dpcz" Nov 22 08:14:13 crc kubenswrapper[4789]: I1122 08:14:13.355078 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e32fe7bc-c0c8-443b-af3e-d9cc58db4f24-config\") pod \"dnsmasq-dns-7c6d9948dc-2dpcz\" (UID: \"e32fe7bc-c0c8-443b-af3e-d9cc58db4f24\") " pod="openstack/dnsmasq-dns-7c6d9948dc-2dpcz" Nov 22 08:14:13 crc kubenswrapper[4789]: I1122 08:14:13.456581 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e32fe7bc-c0c8-443b-af3e-d9cc58db4f24-config\") pod \"dnsmasq-dns-7c6d9948dc-2dpcz\" (UID: \"e32fe7bc-c0c8-443b-af3e-d9cc58db4f24\") " pod="openstack/dnsmasq-dns-7c6d9948dc-2dpcz" Nov 22 08:14:13 crc kubenswrapper[4789]: I1122 08:14:13.456693 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e32fe7bc-c0c8-443b-af3e-d9cc58db4f24-dns-svc\") pod \"dnsmasq-dns-7c6d9948dc-2dpcz\" (UID: \"e32fe7bc-c0c8-443b-af3e-d9cc58db4f24\") " pod="openstack/dnsmasq-dns-7c6d9948dc-2dpcz" Nov 22 08:14:13 crc kubenswrapper[4789]: I1122 08:14:13.456725 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-llqn6\" (UniqueName: \"kubernetes.io/projected/e32fe7bc-c0c8-443b-af3e-d9cc58db4f24-kube-api-access-llqn6\") pod \"dnsmasq-dns-7c6d9948dc-2dpcz\" (UID: \"e32fe7bc-c0c8-443b-af3e-d9cc58db4f24\") " pod="openstack/dnsmasq-dns-7c6d9948dc-2dpcz" Nov 22 08:14:13 crc kubenswrapper[4789]: I1122 08:14:13.457669 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e32fe7bc-c0c8-443b-af3e-d9cc58db4f24-dns-svc\") pod \"dnsmasq-dns-7c6d9948dc-2dpcz\" (UID: \"e32fe7bc-c0c8-443b-af3e-d9cc58db4f24\") " pod="openstack/dnsmasq-dns-7c6d9948dc-2dpcz" Nov 22 08:14:13 crc kubenswrapper[4789]: I1122 08:14:13.457840 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e32fe7bc-c0c8-443b-af3e-d9cc58db4f24-config\") pod \"dnsmasq-dns-7c6d9948dc-2dpcz\" (UID: \"e32fe7bc-c0c8-443b-af3e-d9cc58db4f24\") " pod="openstack/dnsmasq-dns-7c6d9948dc-2dpcz" Nov 22 08:14:13 crc kubenswrapper[4789]: I1122 08:14:13.490706 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-llqn6\" (UniqueName: \"kubernetes.io/projected/e32fe7bc-c0c8-443b-af3e-d9cc58db4f24-kube-api-access-llqn6\") pod \"dnsmasq-dns-7c6d9948dc-2dpcz\" (UID: \"e32fe7bc-c0c8-443b-af3e-d9cc58db4f24\") " pod="openstack/dnsmasq-dns-7c6d9948dc-2dpcz" Nov 22 08:14:13 crc kubenswrapper[4789]: I1122 08:14:13.492074 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7bdd77c89-nq45j"] Nov 22 08:14:13 crc kubenswrapper[4789]: I1122 08:14:13.523615 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6486446b9f-2w9gc"] Nov 22 08:14:13 crc kubenswrapper[4789]: I1122 08:14:13.526305 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c6d9948dc-2dpcz" Nov 22 08:14:13 crc kubenswrapper[4789]: I1122 08:14:13.527484 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6486446b9f-2w9gc" Nov 22 08:14:13 crc kubenswrapper[4789]: I1122 08:14:13.547355 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6486446b9f-2w9gc"] Nov 22 08:14:13 crc kubenswrapper[4789]: I1122 08:14:13.562610 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/66b9aef5-3c42-4449-987c-9087653553b5-config\") pod \"dnsmasq-dns-6486446b9f-2w9gc\" (UID: \"66b9aef5-3c42-4449-987c-9087653553b5\") " pod="openstack/dnsmasq-dns-6486446b9f-2w9gc" Nov 22 08:14:13 crc kubenswrapper[4789]: I1122 08:14:13.562720 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j9nj2\" (UniqueName: \"kubernetes.io/projected/66b9aef5-3c42-4449-987c-9087653553b5-kube-api-access-j9nj2\") pod \"dnsmasq-dns-6486446b9f-2w9gc\" (UID: \"66b9aef5-3c42-4449-987c-9087653553b5\") " pod="openstack/dnsmasq-dns-6486446b9f-2w9gc" Nov 22 08:14:13 crc kubenswrapper[4789]: I1122 08:14:13.562782 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/66b9aef5-3c42-4449-987c-9087653553b5-dns-svc\") pod \"dnsmasq-dns-6486446b9f-2w9gc\" (UID: \"66b9aef5-3c42-4449-987c-9087653553b5\") " pod="openstack/dnsmasq-dns-6486446b9f-2w9gc" Nov 22 08:14:13 crc kubenswrapper[4789]: I1122 08:14:13.664694 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/66b9aef5-3c42-4449-987c-9087653553b5-config\") pod \"dnsmasq-dns-6486446b9f-2w9gc\" (UID: \"66b9aef5-3c42-4449-987c-9087653553b5\") " pod="openstack/dnsmasq-dns-6486446b9f-2w9gc" Nov 22 08:14:13 crc kubenswrapper[4789]: I1122 08:14:13.664774 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j9nj2\" (UniqueName: \"kubernetes.io/projected/66b9aef5-3c42-4449-987c-9087653553b5-kube-api-access-j9nj2\") pod \"dnsmasq-dns-6486446b9f-2w9gc\" (UID: \"66b9aef5-3c42-4449-987c-9087653553b5\") " pod="openstack/dnsmasq-dns-6486446b9f-2w9gc" Nov 22 08:14:13 crc kubenswrapper[4789]: I1122 08:14:13.664798 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/66b9aef5-3c42-4449-987c-9087653553b5-dns-svc\") pod \"dnsmasq-dns-6486446b9f-2w9gc\" (UID: \"66b9aef5-3c42-4449-987c-9087653553b5\") " pod="openstack/dnsmasq-dns-6486446b9f-2w9gc" Nov 22 08:14:13 crc kubenswrapper[4789]: I1122 08:14:13.665701 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/66b9aef5-3c42-4449-987c-9087653553b5-dns-svc\") pod \"dnsmasq-dns-6486446b9f-2w9gc\" (UID: \"66b9aef5-3c42-4449-987c-9087653553b5\") " pod="openstack/dnsmasq-dns-6486446b9f-2w9gc" Nov 22 08:14:13 crc kubenswrapper[4789]: I1122 08:14:13.666227 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/66b9aef5-3c42-4449-987c-9087653553b5-config\") pod \"dnsmasq-dns-6486446b9f-2w9gc\" (UID: \"66b9aef5-3c42-4449-987c-9087653553b5\") " pod="openstack/dnsmasq-dns-6486446b9f-2w9gc" Nov 22 08:14:13 crc kubenswrapper[4789]: I1122 08:14:13.700524 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j9nj2\" (UniqueName: \"kubernetes.io/projected/66b9aef5-3c42-4449-987c-9087653553b5-kube-api-access-j9nj2\") pod \"dnsmasq-dns-6486446b9f-2w9gc\" (UID: \"66b9aef5-3c42-4449-987c-9087653553b5\") " pod="openstack/dnsmasq-dns-6486446b9f-2w9gc" Nov 22 08:14:13 crc kubenswrapper[4789]: I1122 08:14:13.935007 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6486446b9f-2w9gc" Nov 22 08:14:14 crc kubenswrapper[4789]: I1122 08:14:14.039341 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7c6d9948dc-2dpcz"] Nov 22 08:14:14 crc kubenswrapper[4789]: I1122 08:14:14.401706 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6486446b9f-2w9gc"] Nov 22 08:14:14 crc kubenswrapper[4789]: I1122 08:14:14.490615 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c6d9948dc-2dpcz" event={"ID":"e32fe7bc-c0c8-443b-af3e-d9cc58db4f24","Type":"ContainerStarted","Data":"ac1c4ffae0e5f9a9e0a9e26d05d3f8afe5744c521f78f2e52121108a10794560"} Nov 22 08:14:14 crc kubenswrapper[4789]: I1122 08:14:14.494286 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6486446b9f-2w9gc" event={"ID":"66b9aef5-3c42-4449-987c-9087653553b5","Type":"ContainerStarted","Data":"bc250ff924e10b9073d12fc2a8f12d578726b089a226bc9781f93b8fdb6bbf00"} Nov 22 08:14:14 crc kubenswrapper[4789]: I1122 08:14:14.595482 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Nov 22 08:14:14 crc kubenswrapper[4789]: I1122 08:14:14.597968 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 22 08:14:14 crc kubenswrapper[4789]: I1122 08:14:14.600908 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Nov 22 08:14:14 crc kubenswrapper[4789]: I1122 08:14:14.601044 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Nov 22 08:14:14 crc kubenswrapper[4789]: I1122 08:14:14.600950 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Nov 22 08:14:14 crc kubenswrapper[4789]: I1122 08:14:14.601200 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-ggnlw" Nov 22 08:14:14 crc kubenswrapper[4789]: I1122 08:14:14.602372 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Nov 22 08:14:14 crc kubenswrapper[4789]: I1122 08:14:14.602400 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Nov 22 08:14:14 crc kubenswrapper[4789]: I1122 08:14:14.603637 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Nov 22 08:14:14 crc kubenswrapper[4789]: I1122 08:14:14.614177 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 22 08:14:14 crc kubenswrapper[4789]: I1122 08:14:14.663480 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 22 08:14:14 crc kubenswrapper[4789]: I1122 08:14:14.665437 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:14:14 crc kubenswrapper[4789]: I1122 08:14:14.669150 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Nov 22 08:14:14 crc kubenswrapper[4789]: I1122 08:14:14.669208 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Nov 22 08:14:14 crc kubenswrapper[4789]: I1122 08:14:14.669212 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Nov 22 08:14:14 crc kubenswrapper[4789]: I1122 08:14:14.669491 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Nov 22 08:14:14 crc kubenswrapper[4789]: I1122 08:14:14.670060 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Nov 22 08:14:14 crc kubenswrapper[4789]: I1122 08:14:14.670189 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-4hvjr" Nov 22 08:14:14 crc kubenswrapper[4789]: I1122 08:14:14.670234 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Nov 22 08:14:14 crc kubenswrapper[4789]: I1122 08:14:14.683479 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 22 08:14:14 crc kubenswrapper[4789]: I1122 08:14:14.787314 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fv6wj\" (UniqueName: \"kubernetes.io/projected/df1ccc5e-c20c-40ce-bb76-17da02c4bc8c-kube-api-access-fv6wj\") pod \"rabbitmq-cell1-server-0\" (UID: \"df1ccc5e-c20c-40ce-bb76-17da02c4bc8c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:14:14 crc kubenswrapper[4789]: I1122 08:14:14.787353 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/df1ccc5e-c20c-40ce-bb76-17da02c4bc8c-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"df1ccc5e-c20c-40ce-bb76-17da02c4bc8c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:14:14 crc kubenswrapper[4789]: I1122 08:14:14.787386 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/af27dae6-c617-4d16-bb53-19057df7689b-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"af27dae6-c617-4d16-bb53-19057df7689b\") " pod="openstack/rabbitmq-server-0" Nov 22 08:14:14 crc kubenswrapper[4789]: I1122 08:14:14.787404 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/df1ccc5e-c20c-40ce-bb76-17da02c4bc8c-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"df1ccc5e-c20c-40ce-bb76-17da02c4bc8c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:14:14 crc kubenswrapper[4789]: I1122 08:14:14.787434 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/df1ccc5e-c20c-40ce-bb76-17da02c4bc8c-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"df1ccc5e-c20c-40ce-bb76-17da02c4bc8c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:14:14 crc kubenswrapper[4789]: I1122 08:14:14.787479 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/df1ccc5e-c20c-40ce-bb76-17da02c4bc8c-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"df1ccc5e-c20c-40ce-bb76-17da02c4bc8c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:14:14 crc kubenswrapper[4789]: I1122 08:14:14.787530 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/af27dae6-c617-4d16-bb53-19057df7689b-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"af27dae6-c617-4d16-bb53-19057df7689b\") " pod="openstack/rabbitmq-server-0" Nov 22 08:14:14 crc kubenswrapper[4789]: I1122 08:14:14.787555 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/af27dae6-c617-4d16-bb53-19057df7689b-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"af27dae6-c617-4d16-bb53-19057df7689b\") " pod="openstack/rabbitmq-server-0" Nov 22 08:14:14 crc kubenswrapper[4789]: I1122 08:14:14.787575 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/df1ccc5e-c20c-40ce-bb76-17da02c4bc8c-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"df1ccc5e-c20c-40ce-bb76-17da02c4bc8c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:14:14 crc kubenswrapper[4789]: I1122 08:14:14.787594 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"df1ccc5e-c20c-40ce-bb76-17da02c4bc8c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:14:14 crc kubenswrapper[4789]: I1122 08:14:14.787611 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/df1ccc5e-c20c-40ce-bb76-17da02c4bc8c-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"df1ccc5e-c20c-40ce-bb76-17da02c4bc8c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:14:14 crc kubenswrapper[4789]: I1122 08:14:14.787628 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/df1ccc5e-c20c-40ce-bb76-17da02c4bc8c-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"df1ccc5e-c20c-40ce-bb76-17da02c4bc8c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:14:14 crc kubenswrapper[4789]: I1122 08:14:14.787644 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-server-0\" (UID: \"af27dae6-c617-4d16-bb53-19057df7689b\") " pod="openstack/rabbitmq-server-0" Nov 22 08:14:14 crc kubenswrapper[4789]: I1122 08:14:14.787658 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/df1ccc5e-c20c-40ce-bb76-17da02c4bc8c-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"df1ccc5e-c20c-40ce-bb76-17da02c4bc8c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:14:14 crc kubenswrapper[4789]: I1122 08:14:14.787678 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/df1ccc5e-c20c-40ce-bb76-17da02c4bc8c-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"df1ccc5e-c20c-40ce-bb76-17da02c4bc8c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:14:14 crc kubenswrapper[4789]: I1122 08:14:14.787721 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/af27dae6-c617-4d16-bb53-19057df7689b-config-data\") pod \"rabbitmq-server-0\" (UID: \"af27dae6-c617-4d16-bb53-19057df7689b\") " pod="openstack/rabbitmq-server-0" Nov 22 08:14:14 crc kubenswrapper[4789]: I1122 08:14:14.787761 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j7wtx\" (UniqueName: \"kubernetes.io/projected/af27dae6-c617-4d16-bb53-19057df7689b-kube-api-access-j7wtx\") pod \"rabbitmq-server-0\" (UID: \"af27dae6-c617-4d16-bb53-19057df7689b\") " pod="openstack/rabbitmq-server-0" Nov 22 08:14:14 crc kubenswrapper[4789]: I1122 08:14:14.787794 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/af27dae6-c617-4d16-bb53-19057df7689b-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"af27dae6-c617-4d16-bb53-19057df7689b\") " pod="openstack/rabbitmq-server-0" Nov 22 08:14:14 crc kubenswrapper[4789]: I1122 08:14:14.787823 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/af27dae6-c617-4d16-bb53-19057df7689b-server-conf\") pod \"rabbitmq-server-0\" (UID: \"af27dae6-c617-4d16-bb53-19057df7689b\") " pod="openstack/rabbitmq-server-0" Nov 22 08:14:14 crc kubenswrapper[4789]: I1122 08:14:14.787841 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/af27dae6-c617-4d16-bb53-19057df7689b-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"af27dae6-c617-4d16-bb53-19057df7689b\") " pod="openstack/rabbitmq-server-0" Nov 22 08:14:14 crc kubenswrapper[4789]: I1122 08:14:14.787856 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/af27dae6-c617-4d16-bb53-19057df7689b-pod-info\") pod \"rabbitmq-server-0\" (UID: \"af27dae6-c617-4d16-bb53-19057df7689b\") " pod="openstack/rabbitmq-server-0" Nov 22 08:14:14 crc kubenswrapper[4789]: I1122 08:14:14.787874 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/af27dae6-c617-4d16-bb53-19057df7689b-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"af27dae6-c617-4d16-bb53-19057df7689b\") " pod="openstack/rabbitmq-server-0" Nov 22 08:14:14 crc kubenswrapper[4789]: I1122 08:14:14.890133 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-server-0\" (UID: \"af27dae6-c617-4d16-bb53-19057df7689b\") " pod="openstack/rabbitmq-server-0" Nov 22 08:14:14 crc kubenswrapper[4789]: I1122 08:14:14.890175 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/df1ccc5e-c20c-40ce-bb76-17da02c4bc8c-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"df1ccc5e-c20c-40ce-bb76-17da02c4bc8c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:14:14 crc kubenswrapper[4789]: I1122 08:14:14.890199 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/df1ccc5e-c20c-40ce-bb76-17da02c4bc8c-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"df1ccc5e-c20c-40ce-bb76-17da02c4bc8c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:14:14 crc kubenswrapper[4789]: I1122 08:14:14.890231 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/af27dae6-c617-4d16-bb53-19057df7689b-config-data\") pod \"rabbitmq-server-0\" (UID: \"af27dae6-c617-4d16-bb53-19057df7689b\") " pod="openstack/rabbitmq-server-0" Nov 22 08:14:14 crc kubenswrapper[4789]: I1122 08:14:14.890255 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j7wtx\" (UniqueName: \"kubernetes.io/projected/af27dae6-c617-4d16-bb53-19057df7689b-kube-api-access-j7wtx\") pod \"rabbitmq-server-0\" (UID: \"af27dae6-c617-4d16-bb53-19057df7689b\") " pod="openstack/rabbitmq-server-0" Nov 22 08:14:14 crc kubenswrapper[4789]: I1122 08:14:14.890272 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/af27dae6-c617-4d16-bb53-19057df7689b-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"af27dae6-c617-4d16-bb53-19057df7689b\") " pod="openstack/rabbitmq-server-0" Nov 22 08:14:14 crc kubenswrapper[4789]: I1122 08:14:14.890292 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/af27dae6-c617-4d16-bb53-19057df7689b-server-conf\") pod \"rabbitmq-server-0\" (UID: \"af27dae6-c617-4d16-bb53-19057df7689b\") " pod="openstack/rabbitmq-server-0" Nov 22 08:14:14 crc kubenswrapper[4789]: I1122 08:14:14.890306 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/af27dae6-c617-4d16-bb53-19057df7689b-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"af27dae6-c617-4d16-bb53-19057df7689b\") " pod="openstack/rabbitmq-server-0" Nov 22 08:14:14 crc kubenswrapper[4789]: I1122 08:14:14.890323 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/af27dae6-c617-4d16-bb53-19057df7689b-pod-info\") pod \"rabbitmq-server-0\" (UID: \"af27dae6-c617-4d16-bb53-19057df7689b\") " pod="openstack/rabbitmq-server-0" Nov 22 08:14:14 crc kubenswrapper[4789]: I1122 08:14:14.890340 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/af27dae6-c617-4d16-bb53-19057df7689b-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"af27dae6-c617-4d16-bb53-19057df7689b\") " pod="openstack/rabbitmq-server-0" Nov 22 08:14:14 crc kubenswrapper[4789]: I1122 08:14:14.890359 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fv6wj\" (UniqueName: \"kubernetes.io/projected/df1ccc5e-c20c-40ce-bb76-17da02c4bc8c-kube-api-access-fv6wj\") pod \"rabbitmq-cell1-server-0\" (UID: \"df1ccc5e-c20c-40ce-bb76-17da02c4bc8c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:14:14 crc kubenswrapper[4789]: I1122 08:14:14.890374 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/df1ccc5e-c20c-40ce-bb76-17da02c4bc8c-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"df1ccc5e-c20c-40ce-bb76-17da02c4bc8c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:14:14 crc kubenswrapper[4789]: I1122 08:14:14.890389 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/af27dae6-c617-4d16-bb53-19057df7689b-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"af27dae6-c617-4d16-bb53-19057df7689b\") " pod="openstack/rabbitmq-server-0" Nov 22 08:14:14 crc kubenswrapper[4789]: I1122 08:14:14.890404 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/df1ccc5e-c20c-40ce-bb76-17da02c4bc8c-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"df1ccc5e-c20c-40ce-bb76-17da02c4bc8c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:14:14 crc kubenswrapper[4789]: I1122 08:14:14.890430 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/df1ccc5e-c20c-40ce-bb76-17da02c4bc8c-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"df1ccc5e-c20c-40ce-bb76-17da02c4bc8c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:14:14 crc kubenswrapper[4789]: I1122 08:14:14.890446 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/df1ccc5e-c20c-40ce-bb76-17da02c4bc8c-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"df1ccc5e-c20c-40ce-bb76-17da02c4bc8c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:14:14 crc kubenswrapper[4789]: I1122 08:14:14.890465 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/af27dae6-c617-4d16-bb53-19057df7689b-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"af27dae6-c617-4d16-bb53-19057df7689b\") " pod="openstack/rabbitmq-server-0" Nov 22 08:14:14 crc kubenswrapper[4789]: I1122 08:14:14.890488 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/af27dae6-c617-4d16-bb53-19057df7689b-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"af27dae6-c617-4d16-bb53-19057df7689b\") " pod="openstack/rabbitmq-server-0" Nov 22 08:14:14 crc kubenswrapper[4789]: I1122 08:14:14.890507 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/df1ccc5e-c20c-40ce-bb76-17da02c4bc8c-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"df1ccc5e-c20c-40ce-bb76-17da02c4bc8c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:14:14 crc kubenswrapper[4789]: I1122 08:14:14.890526 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"df1ccc5e-c20c-40ce-bb76-17da02c4bc8c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:14:14 crc kubenswrapper[4789]: I1122 08:14:14.890541 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/df1ccc5e-c20c-40ce-bb76-17da02c4bc8c-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"df1ccc5e-c20c-40ce-bb76-17da02c4bc8c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:14:14 crc kubenswrapper[4789]: I1122 08:14:14.890557 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/df1ccc5e-c20c-40ce-bb76-17da02c4bc8c-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"df1ccc5e-c20c-40ce-bb76-17da02c4bc8c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:14:14 crc kubenswrapper[4789]: I1122 08:14:14.891816 4789 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-server-0\" (UID: \"af27dae6-c617-4d16-bb53-19057df7689b\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/rabbitmq-server-0" Nov 22 08:14:14 crc kubenswrapper[4789]: I1122 08:14:14.892582 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/df1ccc5e-c20c-40ce-bb76-17da02c4bc8c-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"df1ccc5e-c20c-40ce-bb76-17da02c4bc8c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:14:14 crc kubenswrapper[4789]: I1122 08:14:14.898324 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/af27dae6-c617-4d16-bb53-19057df7689b-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"af27dae6-c617-4d16-bb53-19057df7689b\") " pod="openstack/rabbitmq-server-0" Nov 22 08:14:14 crc kubenswrapper[4789]: I1122 08:14:14.898818 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/af27dae6-c617-4d16-bb53-19057df7689b-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"af27dae6-c617-4d16-bb53-19057df7689b\") " pod="openstack/rabbitmq-server-0" Nov 22 08:14:14 crc kubenswrapper[4789]: I1122 08:14:14.899299 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/af27dae6-c617-4d16-bb53-19057df7689b-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"af27dae6-c617-4d16-bb53-19057df7689b\") " pod="openstack/rabbitmq-server-0" Nov 22 08:14:14 crc kubenswrapper[4789]: I1122 08:14:14.899307 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/df1ccc5e-c20c-40ce-bb76-17da02c4bc8c-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"df1ccc5e-c20c-40ce-bb76-17da02c4bc8c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:14:14 crc kubenswrapper[4789]: I1122 08:14:14.899888 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/af27dae6-c617-4d16-bb53-19057df7689b-server-conf\") pod \"rabbitmq-server-0\" (UID: \"af27dae6-c617-4d16-bb53-19057df7689b\") " pod="openstack/rabbitmq-server-0" Nov 22 08:14:14 crc kubenswrapper[4789]: I1122 08:14:14.899980 4789 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"df1ccc5e-c20c-40ce-bb76-17da02c4bc8c\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:14:14 crc kubenswrapper[4789]: I1122 08:14:14.900648 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/df1ccc5e-c20c-40ce-bb76-17da02c4bc8c-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"df1ccc5e-c20c-40ce-bb76-17da02c4bc8c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:14:14 crc kubenswrapper[4789]: I1122 08:14:14.901272 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/df1ccc5e-c20c-40ce-bb76-17da02c4bc8c-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"df1ccc5e-c20c-40ce-bb76-17da02c4bc8c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:14:14 crc kubenswrapper[4789]: I1122 08:14:14.901591 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/df1ccc5e-c20c-40ce-bb76-17da02c4bc8c-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"df1ccc5e-c20c-40ce-bb76-17da02c4bc8c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:14:14 crc kubenswrapper[4789]: I1122 08:14:14.901923 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/af27dae6-c617-4d16-bb53-19057df7689b-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"af27dae6-c617-4d16-bb53-19057df7689b\") " pod="openstack/rabbitmq-server-0" Nov 22 08:14:14 crc kubenswrapper[4789]: I1122 08:14:14.902021 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/df1ccc5e-c20c-40ce-bb76-17da02c4bc8c-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"df1ccc5e-c20c-40ce-bb76-17da02c4bc8c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:14:14 crc kubenswrapper[4789]: I1122 08:14:14.902798 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/df1ccc5e-c20c-40ce-bb76-17da02c4bc8c-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"df1ccc5e-c20c-40ce-bb76-17da02c4bc8c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:14:14 crc kubenswrapper[4789]: I1122 08:14:14.914028 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/df1ccc5e-c20c-40ce-bb76-17da02c4bc8c-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"df1ccc5e-c20c-40ce-bb76-17da02c4bc8c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:14:14 crc kubenswrapper[4789]: I1122 08:14:14.914338 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/af27dae6-c617-4d16-bb53-19057df7689b-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"af27dae6-c617-4d16-bb53-19057df7689b\") " pod="openstack/rabbitmq-server-0" Nov 22 08:14:14 crc kubenswrapper[4789]: I1122 08:14:14.919184 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fv6wj\" (UniqueName: \"kubernetes.io/projected/df1ccc5e-c20c-40ce-bb76-17da02c4bc8c-kube-api-access-fv6wj\") pod \"rabbitmq-cell1-server-0\" (UID: \"df1ccc5e-c20c-40ce-bb76-17da02c4bc8c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:14:14 crc kubenswrapper[4789]: I1122 08:14:14.919664 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j7wtx\" (UniqueName: \"kubernetes.io/projected/af27dae6-c617-4d16-bb53-19057df7689b-kube-api-access-j7wtx\") pod \"rabbitmq-server-0\" (UID: \"af27dae6-c617-4d16-bb53-19057df7689b\") " pod="openstack/rabbitmq-server-0" Nov 22 08:14:14 crc kubenswrapper[4789]: I1122 08:14:14.922743 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/af27dae6-c617-4d16-bb53-19057df7689b-config-data\") pod \"rabbitmq-server-0\" (UID: \"af27dae6-c617-4d16-bb53-19057df7689b\") " pod="openstack/rabbitmq-server-0" Nov 22 08:14:14 crc kubenswrapper[4789]: I1122 08:14:14.927849 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/df1ccc5e-c20c-40ce-bb76-17da02c4bc8c-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"df1ccc5e-c20c-40ce-bb76-17da02c4bc8c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:14:14 crc kubenswrapper[4789]: I1122 08:14:14.936966 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"df1ccc5e-c20c-40ce-bb76-17da02c4bc8c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:14:14 crc kubenswrapper[4789]: I1122 08:14:14.955263 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/af27dae6-c617-4d16-bb53-19057df7689b-pod-info\") pod \"rabbitmq-server-0\" (UID: \"af27dae6-c617-4d16-bb53-19057df7689b\") " pod="openstack/rabbitmq-server-0" Nov 22 08:14:14 crc kubenswrapper[4789]: I1122 08:14:14.955815 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/af27dae6-c617-4d16-bb53-19057df7689b-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"af27dae6-c617-4d16-bb53-19057df7689b\") " pod="openstack/rabbitmq-server-0" Nov 22 08:14:14 crc kubenswrapper[4789]: I1122 08:14:14.969638 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-server-0\" (UID: \"af27dae6-c617-4d16-bb53-19057df7689b\") " pod="openstack/rabbitmq-server-0" Nov 22 08:14:14 crc kubenswrapper[4789]: I1122 08:14:14.986539 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:14:15 crc kubenswrapper[4789]: I1122 08:14:15.225695 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 22 08:14:15 crc kubenswrapper[4789]: I1122 08:14:15.514801 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 22 08:14:15 crc kubenswrapper[4789]: W1122 08:14:15.537969 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddf1ccc5e_c20c_40ce_bb76_17da02c4bc8c.slice/crio-9c40add8b63df1edbbdcaf738202f8031af6061b2dda83b3afea044739a87050 WatchSource:0}: Error finding container 9c40add8b63df1edbbdcaf738202f8031af6061b2dda83b3afea044739a87050: Status 404 returned error can't find the container with id 9c40add8b63df1edbbdcaf738202f8031af6061b2dda83b3afea044739a87050 Nov 22 08:14:15 crc kubenswrapper[4789]: I1122 08:14:15.811068 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 22 08:14:15 crc kubenswrapper[4789]: I1122 08:14:15.962715 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Nov 22 08:14:15 crc kubenswrapper[4789]: I1122 08:14:15.965275 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Nov 22 08:14:15 crc kubenswrapper[4789]: I1122 08:14:15.968963 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Nov 22 08:14:15 crc kubenswrapper[4789]: I1122 08:14:15.969799 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Nov 22 08:14:15 crc kubenswrapper[4789]: I1122 08:14:15.973382 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Nov 22 08:14:15 crc kubenswrapper[4789]: I1122 08:14:15.977386 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Nov 22 08:14:15 crc kubenswrapper[4789]: I1122 08:14:15.981612 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-82kvt" Nov 22 08:14:16 crc kubenswrapper[4789]: I1122 08:14:16.003170 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Nov 22 08:14:16 crc kubenswrapper[4789]: I1122 08:14:16.112243 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/51737c30-e744-4266-a235-1ddd1c115c4e-operator-scripts\") pod \"openstack-galera-0\" (UID: \"51737c30-e744-4266-a235-1ddd1c115c4e\") " pod="openstack/openstack-galera-0" Nov 22 08:14:16 crc kubenswrapper[4789]: I1122 08:14:16.112323 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q7b9h\" (UniqueName: \"kubernetes.io/projected/51737c30-e744-4266-a235-1ddd1c115c4e-kube-api-access-q7b9h\") pod \"openstack-galera-0\" (UID: \"51737c30-e744-4266-a235-1ddd1c115c4e\") " pod="openstack/openstack-galera-0" Nov 22 08:14:16 crc kubenswrapper[4789]: I1122 08:14:16.112383 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/51737c30-e744-4266-a235-1ddd1c115c4e-kolla-config\") pod \"openstack-galera-0\" (UID: \"51737c30-e744-4266-a235-1ddd1c115c4e\") " pod="openstack/openstack-galera-0" Nov 22 08:14:16 crc kubenswrapper[4789]: I1122 08:14:16.112425 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51737c30-e744-4266-a235-1ddd1c115c4e-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"51737c30-e744-4266-a235-1ddd1c115c4e\") " pod="openstack/openstack-galera-0" Nov 22 08:14:16 crc kubenswrapper[4789]: I1122 08:14:16.112454 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-galera-0\" (UID: \"51737c30-e744-4266-a235-1ddd1c115c4e\") " pod="openstack/openstack-galera-0" Nov 22 08:14:16 crc kubenswrapper[4789]: I1122 08:14:16.112479 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/51737c30-e744-4266-a235-1ddd1c115c4e-config-data-generated\") pod \"openstack-galera-0\" (UID: \"51737c30-e744-4266-a235-1ddd1c115c4e\") " pod="openstack/openstack-galera-0" Nov 22 08:14:16 crc kubenswrapper[4789]: I1122 08:14:16.112537 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/51737c30-e744-4266-a235-1ddd1c115c4e-config-data-default\") pod \"openstack-galera-0\" (UID: \"51737c30-e744-4266-a235-1ddd1c115c4e\") " pod="openstack/openstack-galera-0" Nov 22 08:14:16 crc kubenswrapper[4789]: I1122 08:14:16.112559 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/51737c30-e744-4266-a235-1ddd1c115c4e-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"51737c30-e744-4266-a235-1ddd1c115c4e\") " pod="openstack/openstack-galera-0" Nov 22 08:14:16 crc kubenswrapper[4789]: I1122 08:14:16.213961 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/51737c30-e744-4266-a235-1ddd1c115c4e-config-data-default\") pod \"openstack-galera-0\" (UID: \"51737c30-e744-4266-a235-1ddd1c115c4e\") " pod="openstack/openstack-galera-0" Nov 22 08:14:16 crc kubenswrapper[4789]: I1122 08:14:16.214025 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/51737c30-e744-4266-a235-1ddd1c115c4e-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"51737c30-e744-4266-a235-1ddd1c115c4e\") " pod="openstack/openstack-galera-0" Nov 22 08:14:16 crc kubenswrapper[4789]: I1122 08:14:16.214060 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/51737c30-e744-4266-a235-1ddd1c115c4e-operator-scripts\") pod \"openstack-galera-0\" (UID: \"51737c30-e744-4266-a235-1ddd1c115c4e\") " pod="openstack/openstack-galera-0" Nov 22 08:14:16 crc kubenswrapper[4789]: I1122 08:14:16.214095 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q7b9h\" (UniqueName: \"kubernetes.io/projected/51737c30-e744-4266-a235-1ddd1c115c4e-kube-api-access-q7b9h\") pod \"openstack-galera-0\" (UID: \"51737c30-e744-4266-a235-1ddd1c115c4e\") " pod="openstack/openstack-galera-0" Nov 22 08:14:16 crc kubenswrapper[4789]: I1122 08:14:16.214152 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/51737c30-e744-4266-a235-1ddd1c115c4e-kolla-config\") pod \"openstack-galera-0\" (UID: \"51737c30-e744-4266-a235-1ddd1c115c4e\") " pod="openstack/openstack-galera-0" Nov 22 08:14:16 crc kubenswrapper[4789]: I1122 08:14:16.214196 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51737c30-e744-4266-a235-1ddd1c115c4e-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"51737c30-e744-4266-a235-1ddd1c115c4e\") " pod="openstack/openstack-galera-0" Nov 22 08:14:16 crc kubenswrapper[4789]: I1122 08:14:16.214228 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-galera-0\" (UID: \"51737c30-e744-4266-a235-1ddd1c115c4e\") " pod="openstack/openstack-galera-0" Nov 22 08:14:16 crc kubenswrapper[4789]: I1122 08:14:16.214259 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/51737c30-e744-4266-a235-1ddd1c115c4e-config-data-generated\") pod \"openstack-galera-0\" (UID: \"51737c30-e744-4266-a235-1ddd1c115c4e\") " pod="openstack/openstack-galera-0" Nov 22 08:14:16 crc kubenswrapper[4789]: I1122 08:14:16.214704 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/51737c30-e744-4266-a235-1ddd1c115c4e-config-data-generated\") pod \"openstack-galera-0\" (UID: \"51737c30-e744-4266-a235-1ddd1c115c4e\") " pod="openstack/openstack-galera-0" Nov 22 08:14:16 crc kubenswrapper[4789]: I1122 08:14:16.215548 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/51737c30-e744-4266-a235-1ddd1c115c4e-config-data-default\") pod \"openstack-galera-0\" (UID: \"51737c30-e744-4266-a235-1ddd1c115c4e\") " pod="openstack/openstack-galera-0" Nov 22 08:14:16 crc kubenswrapper[4789]: I1122 08:14:16.216952 4789 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-galera-0\" (UID: \"51737c30-e744-4266-a235-1ddd1c115c4e\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/openstack-galera-0" Nov 22 08:14:16 crc kubenswrapper[4789]: I1122 08:14:16.217255 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/51737c30-e744-4266-a235-1ddd1c115c4e-kolla-config\") pod \"openstack-galera-0\" (UID: \"51737c30-e744-4266-a235-1ddd1c115c4e\") " pod="openstack/openstack-galera-0" Nov 22 08:14:16 crc kubenswrapper[4789]: I1122 08:14:16.219435 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/51737c30-e744-4266-a235-1ddd1c115c4e-operator-scripts\") pod \"openstack-galera-0\" (UID: \"51737c30-e744-4266-a235-1ddd1c115c4e\") " pod="openstack/openstack-galera-0" Nov 22 08:14:16 crc kubenswrapper[4789]: I1122 08:14:16.222165 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51737c30-e744-4266-a235-1ddd1c115c4e-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"51737c30-e744-4266-a235-1ddd1c115c4e\") " pod="openstack/openstack-galera-0" Nov 22 08:14:16 crc kubenswrapper[4789]: I1122 08:14:16.222700 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/51737c30-e744-4266-a235-1ddd1c115c4e-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"51737c30-e744-4266-a235-1ddd1c115c4e\") " pod="openstack/openstack-galera-0" Nov 22 08:14:16 crc kubenswrapper[4789]: I1122 08:14:16.240585 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q7b9h\" (UniqueName: \"kubernetes.io/projected/51737c30-e744-4266-a235-1ddd1c115c4e-kube-api-access-q7b9h\") pod \"openstack-galera-0\" (UID: \"51737c30-e744-4266-a235-1ddd1c115c4e\") " pod="openstack/openstack-galera-0" Nov 22 08:14:16 crc kubenswrapper[4789]: I1122 08:14:16.264323 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-galera-0\" (UID: \"51737c30-e744-4266-a235-1ddd1c115c4e\") " pod="openstack/openstack-galera-0" Nov 22 08:14:16 crc kubenswrapper[4789]: I1122 08:14:16.307251 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Nov 22 08:14:16 crc kubenswrapper[4789]: I1122 08:14:16.544202 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"df1ccc5e-c20c-40ce-bb76-17da02c4bc8c","Type":"ContainerStarted","Data":"9c40add8b63df1edbbdcaf738202f8031af6061b2dda83b3afea044739a87050"} Nov 22 08:14:16 crc kubenswrapper[4789]: I1122 08:14:16.545079 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"af27dae6-c617-4d16-bb53-19057df7689b","Type":"ContainerStarted","Data":"cc56f8ffb597c855574d458ec8daf6759f5b52074439eea42be55aa3d3b2f5e1"} Nov 22 08:14:16 crc kubenswrapper[4789]: I1122 08:14:16.786832 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Nov 22 08:14:17 crc kubenswrapper[4789]: I1122 08:14:17.360474 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 22 08:14:17 crc kubenswrapper[4789]: I1122 08:14:17.366863 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Nov 22 08:14:17 crc kubenswrapper[4789]: I1122 08:14:17.370607 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Nov 22 08:14:17 crc kubenswrapper[4789]: I1122 08:14:17.371087 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Nov 22 08:14:17 crc kubenswrapper[4789]: I1122 08:14:17.371295 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Nov 22 08:14:17 crc kubenswrapper[4789]: I1122 08:14:17.371494 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-kh5hp" Nov 22 08:14:17 crc kubenswrapper[4789]: I1122 08:14:17.383440 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 22 08:14:17 crc kubenswrapper[4789]: I1122 08:14:17.549422 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7778e90a-45cf-4592-ab3d-68c75258e77f-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"7778e90a-45cf-4592-ab3d-68c75258e77f\") " pod="openstack/openstack-cell1-galera-0" Nov 22 08:14:17 crc kubenswrapper[4789]: I1122 08:14:17.549495 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/7778e90a-45cf-4592-ab3d-68c75258e77f-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"7778e90a-45cf-4592-ab3d-68c75258e77f\") " pod="openstack/openstack-cell1-galera-0" Nov 22 08:14:17 crc kubenswrapper[4789]: I1122 08:14:17.549540 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-cell1-galera-0\" (UID: \"7778e90a-45cf-4592-ab3d-68c75258e77f\") " pod="openstack/openstack-cell1-galera-0" Nov 22 08:14:17 crc kubenswrapper[4789]: I1122 08:14:17.549667 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dctmx\" (UniqueName: \"kubernetes.io/projected/7778e90a-45cf-4592-ab3d-68c75258e77f-kube-api-access-dctmx\") pod \"openstack-cell1-galera-0\" (UID: \"7778e90a-45cf-4592-ab3d-68c75258e77f\") " pod="openstack/openstack-cell1-galera-0" Nov 22 08:14:17 crc kubenswrapper[4789]: I1122 08:14:17.549692 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/7778e90a-45cf-4592-ab3d-68c75258e77f-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"7778e90a-45cf-4592-ab3d-68c75258e77f\") " pod="openstack/openstack-cell1-galera-0" Nov 22 08:14:17 crc kubenswrapper[4789]: I1122 08:14:17.549726 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/7778e90a-45cf-4592-ab3d-68c75258e77f-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"7778e90a-45cf-4592-ab3d-68c75258e77f\") " pod="openstack/openstack-cell1-galera-0" Nov 22 08:14:17 crc kubenswrapper[4789]: I1122 08:14:17.549800 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7778e90a-45cf-4592-ab3d-68c75258e77f-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"7778e90a-45cf-4592-ab3d-68c75258e77f\") " pod="openstack/openstack-cell1-galera-0" Nov 22 08:14:17 crc kubenswrapper[4789]: I1122 08:14:17.549861 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/7778e90a-45cf-4592-ab3d-68c75258e77f-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"7778e90a-45cf-4592-ab3d-68c75258e77f\") " pod="openstack/openstack-cell1-galera-0" Nov 22 08:14:17 crc kubenswrapper[4789]: I1122 08:14:17.570582 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"51737c30-e744-4266-a235-1ddd1c115c4e","Type":"ContainerStarted","Data":"84b015efdd52211da73e99c8df8f0a6af6fb98bd3161fb6bfb2bc5741934fd51"} Nov 22 08:14:17 crc kubenswrapper[4789]: I1122 08:14:17.651731 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dctmx\" (UniqueName: \"kubernetes.io/projected/7778e90a-45cf-4592-ab3d-68c75258e77f-kube-api-access-dctmx\") pod \"openstack-cell1-galera-0\" (UID: \"7778e90a-45cf-4592-ab3d-68c75258e77f\") " pod="openstack/openstack-cell1-galera-0" Nov 22 08:14:17 crc kubenswrapper[4789]: I1122 08:14:17.651807 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/7778e90a-45cf-4592-ab3d-68c75258e77f-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"7778e90a-45cf-4592-ab3d-68c75258e77f\") " pod="openstack/openstack-cell1-galera-0" Nov 22 08:14:17 crc kubenswrapper[4789]: I1122 08:14:17.651843 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/7778e90a-45cf-4592-ab3d-68c75258e77f-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"7778e90a-45cf-4592-ab3d-68c75258e77f\") " pod="openstack/openstack-cell1-galera-0" Nov 22 08:14:17 crc kubenswrapper[4789]: I1122 08:14:17.651874 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7778e90a-45cf-4592-ab3d-68c75258e77f-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"7778e90a-45cf-4592-ab3d-68c75258e77f\") " pod="openstack/openstack-cell1-galera-0" Nov 22 08:14:17 crc kubenswrapper[4789]: I1122 08:14:17.651909 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/7778e90a-45cf-4592-ab3d-68c75258e77f-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"7778e90a-45cf-4592-ab3d-68c75258e77f\") " pod="openstack/openstack-cell1-galera-0" Nov 22 08:14:17 crc kubenswrapper[4789]: I1122 08:14:17.651926 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7778e90a-45cf-4592-ab3d-68c75258e77f-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"7778e90a-45cf-4592-ab3d-68c75258e77f\") " pod="openstack/openstack-cell1-galera-0" Nov 22 08:14:17 crc kubenswrapper[4789]: I1122 08:14:17.651953 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/7778e90a-45cf-4592-ab3d-68c75258e77f-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"7778e90a-45cf-4592-ab3d-68c75258e77f\") " pod="openstack/openstack-cell1-galera-0" Nov 22 08:14:17 crc kubenswrapper[4789]: I1122 08:14:17.651973 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-cell1-galera-0\" (UID: \"7778e90a-45cf-4592-ab3d-68c75258e77f\") " pod="openstack/openstack-cell1-galera-0" Nov 22 08:14:17 crc kubenswrapper[4789]: I1122 08:14:17.652354 4789 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-cell1-galera-0\" (UID: \"7778e90a-45cf-4592-ab3d-68c75258e77f\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/openstack-cell1-galera-0" Nov 22 08:14:17 crc kubenswrapper[4789]: I1122 08:14:17.652489 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/7778e90a-45cf-4592-ab3d-68c75258e77f-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"7778e90a-45cf-4592-ab3d-68c75258e77f\") " pod="openstack/openstack-cell1-galera-0" Nov 22 08:14:17 crc kubenswrapper[4789]: I1122 08:14:17.652631 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/7778e90a-45cf-4592-ab3d-68c75258e77f-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"7778e90a-45cf-4592-ab3d-68c75258e77f\") " pod="openstack/openstack-cell1-galera-0" Nov 22 08:14:17 crc kubenswrapper[4789]: I1122 08:14:17.653724 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/7778e90a-45cf-4592-ab3d-68c75258e77f-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"7778e90a-45cf-4592-ab3d-68c75258e77f\") " pod="openstack/openstack-cell1-galera-0" Nov 22 08:14:17 crc kubenswrapper[4789]: I1122 08:14:17.655166 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7778e90a-45cf-4592-ab3d-68c75258e77f-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"7778e90a-45cf-4592-ab3d-68c75258e77f\") " pod="openstack/openstack-cell1-galera-0" Nov 22 08:14:17 crc kubenswrapper[4789]: I1122 08:14:17.660432 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/7778e90a-45cf-4592-ab3d-68c75258e77f-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"7778e90a-45cf-4592-ab3d-68c75258e77f\") " pod="openstack/openstack-cell1-galera-0" Nov 22 08:14:17 crc kubenswrapper[4789]: I1122 08:14:17.667599 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7778e90a-45cf-4592-ab3d-68c75258e77f-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"7778e90a-45cf-4592-ab3d-68c75258e77f\") " pod="openstack/openstack-cell1-galera-0" Nov 22 08:14:17 crc kubenswrapper[4789]: I1122 08:14:17.672784 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dctmx\" (UniqueName: \"kubernetes.io/projected/7778e90a-45cf-4592-ab3d-68c75258e77f-kube-api-access-dctmx\") pod \"openstack-cell1-galera-0\" (UID: \"7778e90a-45cf-4592-ab3d-68c75258e77f\") " pod="openstack/openstack-cell1-galera-0" Nov 22 08:14:17 crc kubenswrapper[4789]: I1122 08:14:17.687005 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-cell1-galera-0\" (UID: \"7778e90a-45cf-4592-ab3d-68c75258e77f\") " pod="openstack/openstack-cell1-galera-0" Nov 22 08:14:17 crc kubenswrapper[4789]: I1122 08:14:17.696521 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Nov 22 08:14:17 crc kubenswrapper[4789]: I1122 08:14:17.722073 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Nov 22 08:14:17 crc kubenswrapper[4789]: I1122 08:14:17.724075 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Nov 22 08:14:17 crc kubenswrapper[4789]: I1122 08:14:17.726594 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Nov 22 08:14:17 crc kubenswrapper[4789]: I1122 08:14:17.727138 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Nov 22 08:14:17 crc kubenswrapper[4789]: I1122 08:14:17.727424 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-n76ts" Nov 22 08:14:17 crc kubenswrapper[4789]: I1122 08:14:17.733491 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Nov 22 08:14:17 crc kubenswrapper[4789]: I1122 08:14:17.855550 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd7daba2-62fe-48dc-8e83-a3764c57f9ba-memcached-tls-certs\") pod \"memcached-0\" (UID: \"bd7daba2-62fe-48dc-8e83-a3764c57f9ba\") " pod="openstack/memcached-0" Nov 22 08:14:17 crc kubenswrapper[4789]: I1122 08:14:17.855595 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bd7daba2-62fe-48dc-8e83-a3764c57f9ba-config-data\") pod \"memcached-0\" (UID: \"bd7daba2-62fe-48dc-8e83-a3764c57f9ba\") " pod="openstack/memcached-0" Nov 22 08:14:17 crc kubenswrapper[4789]: I1122 08:14:17.855789 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd7daba2-62fe-48dc-8e83-a3764c57f9ba-combined-ca-bundle\") pod \"memcached-0\" (UID: \"bd7daba2-62fe-48dc-8e83-a3764c57f9ba\") " pod="openstack/memcached-0" Nov 22 08:14:17 crc kubenswrapper[4789]: I1122 08:14:17.855959 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/bd7daba2-62fe-48dc-8e83-a3764c57f9ba-kolla-config\") pod \"memcached-0\" (UID: \"bd7daba2-62fe-48dc-8e83-a3764c57f9ba\") " pod="openstack/memcached-0" Nov 22 08:14:17 crc kubenswrapper[4789]: I1122 08:14:17.856062 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-md8pm\" (UniqueName: \"kubernetes.io/projected/bd7daba2-62fe-48dc-8e83-a3764c57f9ba-kube-api-access-md8pm\") pod \"memcached-0\" (UID: \"bd7daba2-62fe-48dc-8e83-a3764c57f9ba\") " pod="openstack/memcached-0" Nov 22 08:14:17 crc kubenswrapper[4789]: I1122 08:14:17.957650 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd7daba2-62fe-48dc-8e83-a3764c57f9ba-combined-ca-bundle\") pod \"memcached-0\" (UID: \"bd7daba2-62fe-48dc-8e83-a3764c57f9ba\") " pod="openstack/memcached-0" Nov 22 08:14:17 crc kubenswrapper[4789]: I1122 08:14:17.957797 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/bd7daba2-62fe-48dc-8e83-a3764c57f9ba-kolla-config\") pod \"memcached-0\" (UID: \"bd7daba2-62fe-48dc-8e83-a3764c57f9ba\") " pod="openstack/memcached-0" Nov 22 08:14:17 crc kubenswrapper[4789]: I1122 08:14:17.957844 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-md8pm\" (UniqueName: \"kubernetes.io/projected/bd7daba2-62fe-48dc-8e83-a3764c57f9ba-kube-api-access-md8pm\") pod \"memcached-0\" (UID: \"bd7daba2-62fe-48dc-8e83-a3764c57f9ba\") " pod="openstack/memcached-0" Nov 22 08:14:17 crc kubenswrapper[4789]: I1122 08:14:17.957892 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd7daba2-62fe-48dc-8e83-a3764c57f9ba-memcached-tls-certs\") pod \"memcached-0\" (UID: \"bd7daba2-62fe-48dc-8e83-a3764c57f9ba\") " pod="openstack/memcached-0" Nov 22 08:14:17 crc kubenswrapper[4789]: I1122 08:14:17.957916 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bd7daba2-62fe-48dc-8e83-a3764c57f9ba-config-data\") pod \"memcached-0\" (UID: \"bd7daba2-62fe-48dc-8e83-a3764c57f9ba\") " pod="openstack/memcached-0" Nov 22 08:14:17 crc kubenswrapper[4789]: I1122 08:14:17.958816 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/bd7daba2-62fe-48dc-8e83-a3764c57f9ba-kolla-config\") pod \"memcached-0\" (UID: \"bd7daba2-62fe-48dc-8e83-a3764c57f9ba\") " pod="openstack/memcached-0" Nov 22 08:14:17 crc kubenswrapper[4789]: I1122 08:14:17.958921 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bd7daba2-62fe-48dc-8e83-a3764c57f9ba-config-data\") pod \"memcached-0\" (UID: \"bd7daba2-62fe-48dc-8e83-a3764c57f9ba\") " pod="openstack/memcached-0" Nov 22 08:14:17 crc kubenswrapper[4789]: I1122 08:14:17.963311 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd7daba2-62fe-48dc-8e83-a3764c57f9ba-memcached-tls-certs\") pod \"memcached-0\" (UID: \"bd7daba2-62fe-48dc-8e83-a3764c57f9ba\") " pod="openstack/memcached-0" Nov 22 08:14:17 crc kubenswrapper[4789]: I1122 08:14:17.965901 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd7daba2-62fe-48dc-8e83-a3764c57f9ba-combined-ca-bundle\") pod \"memcached-0\" (UID: \"bd7daba2-62fe-48dc-8e83-a3764c57f9ba\") " pod="openstack/memcached-0" Nov 22 08:14:17 crc kubenswrapper[4789]: I1122 08:14:17.986832 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-md8pm\" (UniqueName: \"kubernetes.io/projected/bd7daba2-62fe-48dc-8e83-a3764c57f9ba-kube-api-access-md8pm\") pod \"memcached-0\" (UID: \"bd7daba2-62fe-48dc-8e83-a3764c57f9ba\") " pod="openstack/memcached-0" Nov 22 08:14:18 crc kubenswrapper[4789]: I1122 08:14:18.122154 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Nov 22 08:14:18 crc kubenswrapper[4789]: I1122 08:14:18.352140 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 22 08:14:19 crc kubenswrapper[4789]: I1122 08:14:19.561550 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Nov 22 08:14:19 crc kubenswrapper[4789]: I1122 08:14:19.563912 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 22 08:14:19 crc kubenswrapper[4789]: I1122 08:14:19.565899 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-r54gw" Nov 22 08:14:19 crc kubenswrapper[4789]: I1122 08:14:19.581418 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 22 08:14:19 crc kubenswrapper[4789]: I1122 08:14:19.694684 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6t4wv\" (UniqueName: \"kubernetes.io/projected/20a309c0-4fac-480e-b19d-e2ca480a6b6c-kube-api-access-6t4wv\") pod \"kube-state-metrics-0\" (UID: \"20a309c0-4fac-480e-b19d-e2ca480a6b6c\") " pod="openstack/kube-state-metrics-0" Nov 22 08:14:19 crc kubenswrapper[4789]: I1122 08:14:19.796736 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6t4wv\" (UniqueName: \"kubernetes.io/projected/20a309c0-4fac-480e-b19d-e2ca480a6b6c-kube-api-access-6t4wv\") pod \"kube-state-metrics-0\" (UID: \"20a309c0-4fac-480e-b19d-e2ca480a6b6c\") " pod="openstack/kube-state-metrics-0" Nov 22 08:14:19 crc kubenswrapper[4789]: I1122 08:14:19.817854 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6t4wv\" (UniqueName: \"kubernetes.io/projected/20a309c0-4fac-480e-b19d-e2ca480a6b6c-kube-api-access-6t4wv\") pod \"kube-state-metrics-0\" (UID: \"20a309c0-4fac-480e-b19d-e2ca480a6b6c\") " pod="openstack/kube-state-metrics-0" Nov 22 08:14:19 crc kubenswrapper[4789]: I1122 08:14:19.915073 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 22 08:14:22 crc kubenswrapper[4789]: W1122 08:14:22.833238 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7778e90a_45cf_4592_ab3d_68c75258e77f.slice/crio-d686cc7417c8c31972270cec7e65438d0fa24afd68e3c5a354bb3f268af78789 WatchSource:0}: Error finding container d686cc7417c8c31972270cec7e65438d0fa24afd68e3c5a354bb3f268af78789: Status 404 returned error can't find the container with id d686cc7417c8c31972270cec7e65438d0fa24afd68e3c5a354bb3f268af78789 Nov 22 08:14:23 crc kubenswrapper[4789]: I1122 08:14:23.193909 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 22 08:14:23 crc kubenswrapper[4789]: I1122 08:14:23.199002 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Nov 22 08:14:23 crc kubenswrapper[4789]: I1122 08:14:23.201493 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Nov 22 08:14:23 crc kubenswrapper[4789]: I1122 08:14:23.201618 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-6v8rn" Nov 22 08:14:23 crc kubenswrapper[4789]: I1122 08:14:23.205270 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Nov 22 08:14:23 crc kubenswrapper[4789]: I1122 08:14:23.206563 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Nov 22 08:14:23 crc kubenswrapper[4789]: I1122 08:14:23.207318 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Nov 22 08:14:23 crc kubenswrapper[4789]: I1122 08:14:23.207949 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 22 08:14:23 crc kubenswrapper[4789]: I1122 08:14:23.229549 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Nov 22 08:14:23 crc kubenswrapper[4789]: I1122 08:14:23.362210 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b479a8a1-0fba-4a72-b830-2a68e2bd9be4-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"b479a8a1-0fba-4a72-b830-2a68e2bd9be4\") " pod="openstack/ovsdbserver-nb-0" Nov 22 08:14:23 crc kubenswrapper[4789]: I1122 08:14:23.362268 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5jrh5\" (UniqueName: \"kubernetes.io/projected/b479a8a1-0fba-4a72-b830-2a68e2bd9be4-kube-api-access-5jrh5\") pod \"ovsdbserver-nb-0\" (UID: \"b479a8a1-0fba-4a72-b830-2a68e2bd9be4\") " pod="openstack/ovsdbserver-nb-0" Nov 22 08:14:23 crc kubenswrapper[4789]: I1122 08:14:23.362364 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/b479a8a1-0fba-4a72-b830-2a68e2bd9be4-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"b479a8a1-0fba-4a72-b830-2a68e2bd9be4\") " pod="openstack/ovsdbserver-nb-0" Nov 22 08:14:23 crc kubenswrapper[4789]: I1122 08:14:23.362392 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-nb-0\" (UID: \"b479a8a1-0fba-4a72-b830-2a68e2bd9be4\") " pod="openstack/ovsdbserver-nb-0" Nov 22 08:14:23 crc kubenswrapper[4789]: I1122 08:14:23.362418 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/b479a8a1-0fba-4a72-b830-2a68e2bd9be4-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"b479a8a1-0fba-4a72-b830-2a68e2bd9be4\") " pod="openstack/ovsdbserver-nb-0" Nov 22 08:14:23 crc kubenswrapper[4789]: I1122 08:14:23.362440 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b479a8a1-0fba-4a72-b830-2a68e2bd9be4-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"b479a8a1-0fba-4a72-b830-2a68e2bd9be4\") " pod="openstack/ovsdbserver-nb-0" Nov 22 08:14:23 crc kubenswrapper[4789]: I1122 08:14:23.362457 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b479a8a1-0fba-4a72-b830-2a68e2bd9be4-config\") pod \"ovsdbserver-nb-0\" (UID: \"b479a8a1-0fba-4a72-b830-2a68e2bd9be4\") " pod="openstack/ovsdbserver-nb-0" Nov 22 08:14:23 crc kubenswrapper[4789]: I1122 08:14:23.362473 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b479a8a1-0fba-4a72-b830-2a68e2bd9be4-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"b479a8a1-0fba-4a72-b830-2a68e2bd9be4\") " pod="openstack/ovsdbserver-nb-0" Nov 22 08:14:23 crc kubenswrapper[4789]: I1122 08:14:23.464082 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/b479a8a1-0fba-4a72-b830-2a68e2bd9be4-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"b479a8a1-0fba-4a72-b830-2a68e2bd9be4\") " pod="openstack/ovsdbserver-nb-0" Nov 22 08:14:23 crc kubenswrapper[4789]: I1122 08:14:23.464147 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-nb-0\" (UID: \"b479a8a1-0fba-4a72-b830-2a68e2bd9be4\") " pod="openstack/ovsdbserver-nb-0" Nov 22 08:14:23 crc kubenswrapper[4789]: I1122 08:14:23.464178 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/b479a8a1-0fba-4a72-b830-2a68e2bd9be4-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"b479a8a1-0fba-4a72-b830-2a68e2bd9be4\") " pod="openstack/ovsdbserver-nb-0" Nov 22 08:14:23 crc kubenswrapper[4789]: I1122 08:14:23.464203 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b479a8a1-0fba-4a72-b830-2a68e2bd9be4-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"b479a8a1-0fba-4a72-b830-2a68e2bd9be4\") " pod="openstack/ovsdbserver-nb-0" Nov 22 08:14:23 crc kubenswrapper[4789]: I1122 08:14:23.464219 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b479a8a1-0fba-4a72-b830-2a68e2bd9be4-config\") pod \"ovsdbserver-nb-0\" (UID: \"b479a8a1-0fba-4a72-b830-2a68e2bd9be4\") " pod="openstack/ovsdbserver-nb-0" Nov 22 08:14:23 crc kubenswrapper[4789]: I1122 08:14:23.464235 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b479a8a1-0fba-4a72-b830-2a68e2bd9be4-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"b479a8a1-0fba-4a72-b830-2a68e2bd9be4\") " pod="openstack/ovsdbserver-nb-0" Nov 22 08:14:23 crc kubenswrapper[4789]: I1122 08:14:23.464292 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b479a8a1-0fba-4a72-b830-2a68e2bd9be4-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"b479a8a1-0fba-4a72-b830-2a68e2bd9be4\") " pod="openstack/ovsdbserver-nb-0" Nov 22 08:14:23 crc kubenswrapper[4789]: I1122 08:14:23.464323 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5jrh5\" (UniqueName: \"kubernetes.io/projected/b479a8a1-0fba-4a72-b830-2a68e2bd9be4-kube-api-access-5jrh5\") pod \"ovsdbserver-nb-0\" (UID: \"b479a8a1-0fba-4a72-b830-2a68e2bd9be4\") " pod="openstack/ovsdbserver-nb-0" Nov 22 08:14:23 crc kubenswrapper[4789]: I1122 08:14:23.464598 4789 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-nb-0\" (UID: \"b479a8a1-0fba-4a72-b830-2a68e2bd9be4\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/ovsdbserver-nb-0" Nov 22 08:14:23 crc kubenswrapper[4789]: I1122 08:14:23.465591 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b479a8a1-0fba-4a72-b830-2a68e2bd9be4-config\") pod \"ovsdbserver-nb-0\" (UID: \"b479a8a1-0fba-4a72-b830-2a68e2bd9be4\") " pod="openstack/ovsdbserver-nb-0" Nov 22 08:14:23 crc kubenswrapper[4789]: I1122 08:14:23.466675 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b479a8a1-0fba-4a72-b830-2a68e2bd9be4-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"b479a8a1-0fba-4a72-b830-2a68e2bd9be4\") " pod="openstack/ovsdbserver-nb-0" Nov 22 08:14:23 crc kubenswrapper[4789]: I1122 08:14:23.466974 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/b479a8a1-0fba-4a72-b830-2a68e2bd9be4-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"b479a8a1-0fba-4a72-b830-2a68e2bd9be4\") " pod="openstack/ovsdbserver-nb-0" Nov 22 08:14:23 crc kubenswrapper[4789]: I1122 08:14:23.470542 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b479a8a1-0fba-4a72-b830-2a68e2bd9be4-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"b479a8a1-0fba-4a72-b830-2a68e2bd9be4\") " pod="openstack/ovsdbserver-nb-0" Nov 22 08:14:23 crc kubenswrapper[4789]: I1122 08:14:23.471107 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/b479a8a1-0fba-4a72-b830-2a68e2bd9be4-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"b479a8a1-0fba-4a72-b830-2a68e2bd9be4\") " pod="openstack/ovsdbserver-nb-0" Nov 22 08:14:23 crc kubenswrapper[4789]: I1122 08:14:23.485552 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5jrh5\" (UniqueName: \"kubernetes.io/projected/b479a8a1-0fba-4a72-b830-2a68e2bd9be4-kube-api-access-5jrh5\") pod \"ovsdbserver-nb-0\" (UID: \"b479a8a1-0fba-4a72-b830-2a68e2bd9be4\") " pod="openstack/ovsdbserver-nb-0" Nov 22 08:14:23 crc kubenswrapper[4789]: I1122 08:14:23.490305 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b479a8a1-0fba-4a72-b830-2a68e2bd9be4-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"b479a8a1-0fba-4a72-b830-2a68e2bd9be4\") " pod="openstack/ovsdbserver-nb-0" Nov 22 08:14:23 crc kubenswrapper[4789]: I1122 08:14:23.521076 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-nb-0\" (UID: \"b479a8a1-0fba-4a72-b830-2a68e2bd9be4\") " pod="openstack/ovsdbserver-nb-0" Nov 22 08:14:23 crc kubenswrapper[4789]: I1122 08:14:23.529201 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Nov 22 08:14:23 crc kubenswrapper[4789]: I1122 08:14:23.658308 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"7778e90a-45cf-4592-ab3d-68c75258e77f","Type":"ContainerStarted","Data":"d686cc7417c8c31972270cec7e65438d0fa24afd68e3c5a354bb3f268af78789"} Nov 22 08:14:23 crc kubenswrapper[4789]: I1122 08:14:23.815645 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-cl8m4"] Nov 22 08:14:23 crc kubenswrapper[4789]: I1122 08:14:23.816694 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-cl8m4" Nov 22 08:14:23 crc kubenswrapper[4789]: I1122 08:14:23.818970 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Nov 22 08:14:23 crc kubenswrapper[4789]: I1122 08:14:23.819403 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-dlsv7" Nov 22 08:14:23 crc kubenswrapper[4789]: I1122 08:14:23.820875 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Nov 22 08:14:23 crc kubenswrapper[4789]: I1122 08:14:23.832233 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-gjjwv"] Nov 22 08:14:23 crc kubenswrapper[4789]: I1122 08:14:23.835047 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-gjjwv" Nov 22 08:14:23 crc kubenswrapper[4789]: I1122 08:14:23.841409 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-cl8m4"] Nov 22 08:14:23 crc kubenswrapper[4789]: I1122 08:14:23.849804 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-gjjwv"] Nov 22 08:14:23 crc kubenswrapper[4789]: I1122 08:14:23.974355 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6f04e32-3dca-4d6d-8c8d-19ea15f46875-combined-ca-bundle\") pod \"ovn-controller-cl8m4\" (UID: \"e6f04e32-3dca-4d6d-8c8d-19ea15f46875\") " pod="openstack/ovn-controller-cl8m4" Nov 22 08:14:23 crc kubenswrapper[4789]: I1122 08:14:23.974418 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/ea7bbfd1-bdb6-41b2-87f5-2fa2bbe0c804-etc-ovs\") pod \"ovn-controller-ovs-gjjwv\" (UID: \"ea7bbfd1-bdb6-41b2-87f5-2fa2bbe0c804\") " pod="openstack/ovn-controller-ovs-gjjwv" Nov 22 08:14:23 crc kubenswrapper[4789]: I1122 08:14:23.974461 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e6f04e32-3dca-4d6d-8c8d-19ea15f46875-scripts\") pod \"ovn-controller-cl8m4\" (UID: \"e6f04e32-3dca-4d6d-8c8d-19ea15f46875\") " pod="openstack/ovn-controller-cl8m4" Nov 22 08:14:23 crc kubenswrapper[4789]: I1122 08:14:23.974499 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e6f04e32-3dca-4d6d-8c8d-19ea15f46875-var-run\") pod \"ovn-controller-cl8m4\" (UID: \"e6f04e32-3dca-4d6d-8c8d-19ea15f46875\") " pod="openstack/ovn-controller-cl8m4" Nov 22 08:14:23 crc kubenswrapper[4789]: I1122 08:14:23.974524 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/ea7bbfd1-bdb6-41b2-87f5-2fa2bbe0c804-var-run\") pod \"ovn-controller-ovs-gjjwv\" (UID: \"ea7bbfd1-bdb6-41b2-87f5-2fa2bbe0c804\") " pod="openstack/ovn-controller-ovs-gjjwv" Nov 22 08:14:23 crc kubenswrapper[4789]: I1122 08:14:23.974552 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sgqr9\" (UniqueName: \"kubernetes.io/projected/ea7bbfd1-bdb6-41b2-87f5-2fa2bbe0c804-kube-api-access-sgqr9\") pod \"ovn-controller-ovs-gjjwv\" (UID: \"ea7bbfd1-bdb6-41b2-87f5-2fa2bbe0c804\") " pod="openstack/ovn-controller-ovs-gjjwv" Nov 22 08:14:23 crc kubenswrapper[4789]: I1122 08:14:23.974588 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/ea7bbfd1-bdb6-41b2-87f5-2fa2bbe0c804-var-log\") pod \"ovn-controller-ovs-gjjwv\" (UID: \"ea7bbfd1-bdb6-41b2-87f5-2fa2bbe0c804\") " pod="openstack/ovn-controller-ovs-gjjwv" Nov 22 08:14:23 crc kubenswrapper[4789]: I1122 08:14:23.974607 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e6f04e32-3dca-4d6d-8c8d-19ea15f46875-var-run-ovn\") pod \"ovn-controller-cl8m4\" (UID: \"e6f04e32-3dca-4d6d-8c8d-19ea15f46875\") " pod="openstack/ovn-controller-cl8m4" Nov 22 08:14:23 crc kubenswrapper[4789]: I1122 08:14:23.974635 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e6f04e32-3dca-4d6d-8c8d-19ea15f46875-var-log-ovn\") pod \"ovn-controller-cl8m4\" (UID: \"e6f04e32-3dca-4d6d-8c8d-19ea15f46875\") " pod="openstack/ovn-controller-cl8m4" Nov 22 08:14:23 crc kubenswrapper[4789]: I1122 08:14:23.974656 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/ea7bbfd1-bdb6-41b2-87f5-2fa2bbe0c804-var-lib\") pod \"ovn-controller-ovs-gjjwv\" (UID: \"ea7bbfd1-bdb6-41b2-87f5-2fa2bbe0c804\") " pod="openstack/ovn-controller-ovs-gjjwv" Nov 22 08:14:23 crc kubenswrapper[4789]: I1122 08:14:23.974674 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ea7bbfd1-bdb6-41b2-87f5-2fa2bbe0c804-scripts\") pod \"ovn-controller-ovs-gjjwv\" (UID: \"ea7bbfd1-bdb6-41b2-87f5-2fa2bbe0c804\") " pod="openstack/ovn-controller-ovs-gjjwv" Nov 22 08:14:23 crc kubenswrapper[4789]: I1122 08:14:23.974704 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/e6f04e32-3dca-4d6d-8c8d-19ea15f46875-ovn-controller-tls-certs\") pod \"ovn-controller-cl8m4\" (UID: \"e6f04e32-3dca-4d6d-8c8d-19ea15f46875\") " pod="openstack/ovn-controller-cl8m4" Nov 22 08:14:23 crc kubenswrapper[4789]: I1122 08:14:23.974771 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6q9zk\" (UniqueName: \"kubernetes.io/projected/e6f04e32-3dca-4d6d-8c8d-19ea15f46875-kube-api-access-6q9zk\") pod \"ovn-controller-cl8m4\" (UID: \"e6f04e32-3dca-4d6d-8c8d-19ea15f46875\") " pod="openstack/ovn-controller-cl8m4" Nov 22 08:14:24 crc kubenswrapper[4789]: I1122 08:14:24.076094 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sgqr9\" (UniqueName: \"kubernetes.io/projected/ea7bbfd1-bdb6-41b2-87f5-2fa2bbe0c804-kube-api-access-sgqr9\") pod \"ovn-controller-ovs-gjjwv\" (UID: \"ea7bbfd1-bdb6-41b2-87f5-2fa2bbe0c804\") " pod="openstack/ovn-controller-ovs-gjjwv" Nov 22 08:14:24 crc kubenswrapper[4789]: I1122 08:14:24.076153 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/ea7bbfd1-bdb6-41b2-87f5-2fa2bbe0c804-var-log\") pod \"ovn-controller-ovs-gjjwv\" (UID: \"ea7bbfd1-bdb6-41b2-87f5-2fa2bbe0c804\") " pod="openstack/ovn-controller-ovs-gjjwv" Nov 22 08:14:24 crc kubenswrapper[4789]: I1122 08:14:24.076171 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e6f04e32-3dca-4d6d-8c8d-19ea15f46875-var-run-ovn\") pod \"ovn-controller-cl8m4\" (UID: \"e6f04e32-3dca-4d6d-8c8d-19ea15f46875\") " pod="openstack/ovn-controller-cl8m4" Nov 22 08:14:24 crc kubenswrapper[4789]: I1122 08:14:24.076202 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e6f04e32-3dca-4d6d-8c8d-19ea15f46875-var-log-ovn\") pod \"ovn-controller-cl8m4\" (UID: \"e6f04e32-3dca-4d6d-8c8d-19ea15f46875\") " pod="openstack/ovn-controller-cl8m4" Nov 22 08:14:24 crc kubenswrapper[4789]: I1122 08:14:24.076239 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ea7bbfd1-bdb6-41b2-87f5-2fa2bbe0c804-scripts\") pod \"ovn-controller-ovs-gjjwv\" (UID: \"ea7bbfd1-bdb6-41b2-87f5-2fa2bbe0c804\") " pod="openstack/ovn-controller-ovs-gjjwv" Nov 22 08:14:24 crc kubenswrapper[4789]: I1122 08:14:24.076263 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/ea7bbfd1-bdb6-41b2-87f5-2fa2bbe0c804-var-lib\") pod \"ovn-controller-ovs-gjjwv\" (UID: \"ea7bbfd1-bdb6-41b2-87f5-2fa2bbe0c804\") " pod="openstack/ovn-controller-ovs-gjjwv" Nov 22 08:14:24 crc kubenswrapper[4789]: I1122 08:14:24.076304 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/e6f04e32-3dca-4d6d-8c8d-19ea15f46875-ovn-controller-tls-certs\") pod \"ovn-controller-cl8m4\" (UID: \"e6f04e32-3dca-4d6d-8c8d-19ea15f46875\") " pod="openstack/ovn-controller-cl8m4" Nov 22 08:14:24 crc kubenswrapper[4789]: I1122 08:14:24.076352 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6q9zk\" (UniqueName: \"kubernetes.io/projected/e6f04e32-3dca-4d6d-8c8d-19ea15f46875-kube-api-access-6q9zk\") pod \"ovn-controller-cl8m4\" (UID: \"e6f04e32-3dca-4d6d-8c8d-19ea15f46875\") " pod="openstack/ovn-controller-cl8m4" Nov 22 08:14:24 crc kubenswrapper[4789]: I1122 08:14:24.076374 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6f04e32-3dca-4d6d-8c8d-19ea15f46875-combined-ca-bundle\") pod \"ovn-controller-cl8m4\" (UID: \"e6f04e32-3dca-4d6d-8c8d-19ea15f46875\") " pod="openstack/ovn-controller-cl8m4" Nov 22 08:14:24 crc kubenswrapper[4789]: I1122 08:14:24.076398 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/ea7bbfd1-bdb6-41b2-87f5-2fa2bbe0c804-etc-ovs\") pod \"ovn-controller-ovs-gjjwv\" (UID: \"ea7bbfd1-bdb6-41b2-87f5-2fa2bbe0c804\") " pod="openstack/ovn-controller-ovs-gjjwv" Nov 22 08:14:24 crc kubenswrapper[4789]: I1122 08:14:24.076426 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e6f04e32-3dca-4d6d-8c8d-19ea15f46875-scripts\") pod \"ovn-controller-cl8m4\" (UID: \"e6f04e32-3dca-4d6d-8c8d-19ea15f46875\") " pod="openstack/ovn-controller-cl8m4" Nov 22 08:14:24 crc kubenswrapper[4789]: I1122 08:14:24.076452 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e6f04e32-3dca-4d6d-8c8d-19ea15f46875-var-run\") pod \"ovn-controller-cl8m4\" (UID: \"e6f04e32-3dca-4d6d-8c8d-19ea15f46875\") " pod="openstack/ovn-controller-cl8m4" Nov 22 08:14:24 crc kubenswrapper[4789]: I1122 08:14:24.076473 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/ea7bbfd1-bdb6-41b2-87f5-2fa2bbe0c804-var-run\") pod \"ovn-controller-ovs-gjjwv\" (UID: \"ea7bbfd1-bdb6-41b2-87f5-2fa2bbe0c804\") " pod="openstack/ovn-controller-ovs-gjjwv" Nov 22 08:14:24 crc kubenswrapper[4789]: I1122 08:14:24.076861 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e6f04e32-3dca-4d6d-8c8d-19ea15f46875-var-run-ovn\") pod \"ovn-controller-cl8m4\" (UID: \"e6f04e32-3dca-4d6d-8c8d-19ea15f46875\") " pod="openstack/ovn-controller-cl8m4" Nov 22 08:14:24 crc kubenswrapper[4789]: I1122 08:14:24.076908 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/ea7bbfd1-bdb6-41b2-87f5-2fa2bbe0c804-var-log\") pod \"ovn-controller-ovs-gjjwv\" (UID: \"ea7bbfd1-bdb6-41b2-87f5-2fa2bbe0c804\") " pod="openstack/ovn-controller-ovs-gjjwv" Nov 22 08:14:24 crc kubenswrapper[4789]: I1122 08:14:24.077372 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/ea7bbfd1-bdb6-41b2-87f5-2fa2bbe0c804-etc-ovs\") pod \"ovn-controller-ovs-gjjwv\" (UID: \"ea7bbfd1-bdb6-41b2-87f5-2fa2bbe0c804\") " pod="openstack/ovn-controller-ovs-gjjwv" Nov 22 08:14:24 crc kubenswrapper[4789]: I1122 08:14:24.077666 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e6f04e32-3dca-4d6d-8c8d-19ea15f46875-var-run\") pod \"ovn-controller-cl8m4\" (UID: \"e6f04e32-3dca-4d6d-8c8d-19ea15f46875\") " pod="openstack/ovn-controller-cl8m4" Nov 22 08:14:24 crc kubenswrapper[4789]: I1122 08:14:24.079371 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ea7bbfd1-bdb6-41b2-87f5-2fa2bbe0c804-scripts\") pod \"ovn-controller-ovs-gjjwv\" (UID: \"ea7bbfd1-bdb6-41b2-87f5-2fa2bbe0c804\") " pod="openstack/ovn-controller-ovs-gjjwv" Nov 22 08:14:24 crc kubenswrapper[4789]: I1122 08:14:24.079495 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e6f04e32-3dca-4d6d-8c8d-19ea15f46875-scripts\") pod \"ovn-controller-cl8m4\" (UID: \"e6f04e32-3dca-4d6d-8c8d-19ea15f46875\") " pod="openstack/ovn-controller-cl8m4" Nov 22 08:14:24 crc kubenswrapper[4789]: I1122 08:14:24.079593 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e6f04e32-3dca-4d6d-8c8d-19ea15f46875-var-log-ovn\") pod \"ovn-controller-cl8m4\" (UID: \"e6f04e32-3dca-4d6d-8c8d-19ea15f46875\") " pod="openstack/ovn-controller-cl8m4" Nov 22 08:14:24 crc kubenswrapper[4789]: I1122 08:14:24.079992 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/ea7bbfd1-bdb6-41b2-87f5-2fa2bbe0c804-var-run\") pod \"ovn-controller-ovs-gjjwv\" (UID: \"ea7bbfd1-bdb6-41b2-87f5-2fa2bbe0c804\") " pod="openstack/ovn-controller-ovs-gjjwv" Nov 22 08:14:24 crc kubenswrapper[4789]: I1122 08:14:24.080102 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/ea7bbfd1-bdb6-41b2-87f5-2fa2bbe0c804-var-lib\") pod \"ovn-controller-ovs-gjjwv\" (UID: \"ea7bbfd1-bdb6-41b2-87f5-2fa2bbe0c804\") " pod="openstack/ovn-controller-ovs-gjjwv" Nov 22 08:14:24 crc kubenswrapper[4789]: I1122 08:14:24.081121 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6f04e32-3dca-4d6d-8c8d-19ea15f46875-combined-ca-bundle\") pod \"ovn-controller-cl8m4\" (UID: \"e6f04e32-3dca-4d6d-8c8d-19ea15f46875\") " pod="openstack/ovn-controller-cl8m4" Nov 22 08:14:24 crc kubenswrapper[4789]: I1122 08:14:24.085190 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/e6f04e32-3dca-4d6d-8c8d-19ea15f46875-ovn-controller-tls-certs\") pod \"ovn-controller-cl8m4\" (UID: \"e6f04e32-3dca-4d6d-8c8d-19ea15f46875\") " pod="openstack/ovn-controller-cl8m4" Nov 22 08:14:24 crc kubenswrapper[4789]: I1122 08:14:24.091701 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sgqr9\" (UniqueName: \"kubernetes.io/projected/ea7bbfd1-bdb6-41b2-87f5-2fa2bbe0c804-kube-api-access-sgqr9\") pod \"ovn-controller-ovs-gjjwv\" (UID: \"ea7bbfd1-bdb6-41b2-87f5-2fa2bbe0c804\") " pod="openstack/ovn-controller-ovs-gjjwv" Nov 22 08:14:24 crc kubenswrapper[4789]: I1122 08:14:24.093423 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6q9zk\" (UniqueName: \"kubernetes.io/projected/e6f04e32-3dca-4d6d-8c8d-19ea15f46875-kube-api-access-6q9zk\") pod \"ovn-controller-cl8m4\" (UID: \"e6f04e32-3dca-4d6d-8c8d-19ea15f46875\") " pod="openstack/ovn-controller-cl8m4" Nov 22 08:14:24 crc kubenswrapper[4789]: I1122 08:14:24.140730 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-cl8m4" Nov 22 08:14:24 crc kubenswrapper[4789]: I1122 08:14:24.152789 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-gjjwv" Nov 22 08:14:26 crc kubenswrapper[4789]: I1122 08:14:26.991800 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 22 08:14:26 crc kubenswrapper[4789]: I1122 08:14:26.995927 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Nov 22 08:14:27 crc kubenswrapper[4789]: I1122 08:14:27.004029 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Nov 22 08:14:27 crc kubenswrapper[4789]: I1122 08:14:27.004037 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Nov 22 08:14:27 crc kubenswrapper[4789]: I1122 08:14:27.004391 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 22 08:14:27 crc kubenswrapper[4789]: I1122 08:14:27.004735 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-rqx4d" Nov 22 08:14:27 crc kubenswrapper[4789]: I1122 08:14:27.005034 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Nov 22 08:14:27 crc kubenswrapper[4789]: I1122 08:14:27.125473 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c2d44daa-85ba-473a-a29f-2f2587bfd2fa-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"c2d44daa-85ba-473a-a29f-2f2587bfd2fa\") " pod="openstack/ovsdbserver-sb-0" Nov 22 08:14:27 crc kubenswrapper[4789]: I1122 08:14:27.125541 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c2d44daa-85ba-473a-a29f-2f2587bfd2fa-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"c2d44daa-85ba-473a-a29f-2f2587bfd2fa\") " pod="openstack/ovsdbserver-sb-0" Nov 22 08:14:27 crc kubenswrapper[4789]: I1122 08:14:27.125607 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2d44daa-85ba-473a-a29f-2f2587bfd2fa-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"c2d44daa-85ba-473a-a29f-2f2587bfd2fa\") " pod="openstack/ovsdbserver-sb-0" Nov 22 08:14:27 crc kubenswrapper[4789]: I1122 08:14:27.125655 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c2d44daa-85ba-473a-a29f-2f2587bfd2fa-config\") pod \"ovsdbserver-sb-0\" (UID: \"c2d44daa-85ba-473a-a29f-2f2587bfd2fa\") " pod="openstack/ovsdbserver-sb-0" Nov 22 08:14:27 crc kubenswrapper[4789]: I1122 08:14:27.125679 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pcbmp\" (UniqueName: \"kubernetes.io/projected/c2d44daa-85ba-473a-a29f-2f2587bfd2fa-kube-api-access-pcbmp\") pod \"ovsdbserver-sb-0\" (UID: \"c2d44daa-85ba-473a-a29f-2f2587bfd2fa\") " pod="openstack/ovsdbserver-sb-0" Nov 22 08:14:27 crc kubenswrapper[4789]: I1122 08:14:27.125703 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-sb-0\" (UID: \"c2d44daa-85ba-473a-a29f-2f2587bfd2fa\") " pod="openstack/ovsdbserver-sb-0" Nov 22 08:14:27 crc kubenswrapper[4789]: I1122 08:14:27.125839 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2d44daa-85ba-473a-a29f-2f2587bfd2fa-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"c2d44daa-85ba-473a-a29f-2f2587bfd2fa\") " pod="openstack/ovsdbserver-sb-0" Nov 22 08:14:27 crc kubenswrapper[4789]: I1122 08:14:27.125874 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2d44daa-85ba-473a-a29f-2f2587bfd2fa-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"c2d44daa-85ba-473a-a29f-2f2587bfd2fa\") " pod="openstack/ovsdbserver-sb-0" Nov 22 08:14:27 crc kubenswrapper[4789]: I1122 08:14:27.227407 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pcbmp\" (UniqueName: \"kubernetes.io/projected/c2d44daa-85ba-473a-a29f-2f2587bfd2fa-kube-api-access-pcbmp\") pod \"ovsdbserver-sb-0\" (UID: \"c2d44daa-85ba-473a-a29f-2f2587bfd2fa\") " pod="openstack/ovsdbserver-sb-0" Nov 22 08:14:27 crc kubenswrapper[4789]: I1122 08:14:27.227456 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-sb-0\" (UID: \"c2d44daa-85ba-473a-a29f-2f2587bfd2fa\") " pod="openstack/ovsdbserver-sb-0" Nov 22 08:14:27 crc kubenswrapper[4789]: I1122 08:14:27.227500 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2d44daa-85ba-473a-a29f-2f2587bfd2fa-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"c2d44daa-85ba-473a-a29f-2f2587bfd2fa\") " pod="openstack/ovsdbserver-sb-0" Nov 22 08:14:27 crc kubenswrapper[4789]: I1122 08:14:27.227525 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2d44daa-85ba-473a-a29f-2f2587bfd2fa-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"c2d44daa-85ba-473a-a29f-2f2587bfd2fa\") " pod="openstack/ovsdbserver-sb-0" Nov 22 08:14:27 crc kubenswrapper[4789]: I1122 08:14:27.227767 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c2d44daa-85ba-473a-a29f-2f2587bfd2fa-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"c2d44daa-85ba-473a-a29f-2f2587bfd2fa\") " pod="openstack/ovsdbserver-sb-0" Nov 22 08:14:27 crc kubenswrapper[4789]: I1122 08:14:27.227795 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c2d44daa-85ba-473a-a29f-2f2587bfd2fa-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"c2d44daa-85ba-473a-a29f-2f2587bfd2fa\") " pod="openstack/ovsdbserver-sb-0" Nov 22 08:14:27 crc kubenswrapper[4789]: I1122 08:14:27.227826 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2d44daa-85ba-473a-a29f-2f2587bfd2fa-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"c2d44daa-85ba-473a-a29f-2f2587bfd2fa\") " pod="openstack/ovsdbserver-sb-0" Nov 22 08:14:27 crc kubenswrapper[4789]: I1122 08:14:27.227855 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c2d44daa-85ba-473a-a29f-2f2587bfd2fa-config\") pod \"ovsdbserver-sb-0\" (UID: \"c2d44daa-85ba-473a-a29f-2f2587bfd2fa\") " pod="openstack/ovsdbserver-sb-0" Nov 22 08:14:27 crc kubenswrapper[4789]: I1122 08:14:27.228079 4789 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-sb-0\" (UID: \"c2d44daa-85ba-473a-a29f-2f2587bfd2fa\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/ovsdbserver-sb-0" Nov 22 08:14:27 crc kubenswrapper[4789]: I1122 08:14:27.230547 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c2d44daa-85ba-473a-a29f-2f2587bfd2fa-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"c2d44daa-85ba-473a-a29f-2f2587bfd2fa\") " pod="openstack/ovsdbserver-sb-0" Nov 22 08:14:27 crc kubenswrapper[4789]: I1122 08:14:27.230823 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c2d44daa-85ba-473a-a29f-2f2587bfd2fa-config\") pod \"ovsdbserver-sb-0\" (UID: \"c2d44daa-85ba-473a-a29f-2f2587bfd2fa\") " pod="openstack/ovsdbserver-sb-0" Nov 22 08:14:27 crc kubenswrapper[4789]: I1122 08:14:27.231436 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c2d44daa-85ba-473a-a29f-2f2587bfd2fa-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"c2d44daa-85ba-473a-a29f-2f2587bfd2fa\") " pod="openstack/ovsdbserver-sb-0" Nov 22 08:14:27 crc kubenswrapper[4789]: I1122 08:14:27.234768 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2d44daa-85ba-473a-a29f-2f2587bfd2fa-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"c2d44daa-85ba-473a-a29f-2f2587bfd2fa\") " pod="openstack/ovsdbserver-sb-0" Nov 22 08:14:27 crc kubenswrapper[4789]: I1122 08:14:27.234822 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2d44daa-85ba-473a-a29f-2f2587bfd2fa-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"c2d44daa-85ba-473a-a29f-2f2587bfd2fa\") " pod="openstack/ovsdbserver-sb-0" Nov 22 08:14:27 crc kubenswrapper[4789]: I1122 08:14:27.234884 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2d44daa-85ba-473a-a29f-2f2587bfd2fa-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"c2d44daa-85ba-473a-a29f-2f2587bfd2fa\") " pod="openstack/ovsdbserver-sb-0" Nov 22 08:14:27 crc kubenswrapper[4789]: I1122 08:14:27.253796 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-sb-0\" (UID: \"c2d44daa-85ba-473a-a29f-2f2587bfd2fa\") " pod="openstack/ovsdbserver-sb-0" Nov 22 08:14:27 crc kubenswrapper[4789]: I1122 08:14:27.256074 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pcbmp\" (UniqueName: \"kubernetes.io/projected/c2d44daa-85ba-473a-a29f-2f2587bfd2fa-kube-api-access-pcbmp\") pod \"ovsdbserver-sb-0\" (UID: \"c2d44daa-85ba-473a-a29f-2f2587bfd2fa\") " pod="openstack/ovsdbserver-sb-0" Nov 22 08:14:27 crc kubenswrapper[4789]: I1122 08:14:27.324620 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Nov 22 08:14:27 crc kubenswrapper[4789]: W1122 08:14:27.919590 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbd7daba2_62fe_48dc_8e83_a3764c57f9ba.slice/crio-392bd52d69af77337dc6a9c51ad010b1fa6ba70be6b8a408321615e2aaf1b7e5 WatchSource:0}: Error finding container 392bd52d69af77337dc6a9c51ad010b1fa6ba70be6b8a408321615e2aaf1b7e5: Status 404 returned error can't find the container with id 392bd52d69af77337dc6a9c51ad010b1fa6ba70be6b8a408321615e2aaf1b7e5 Nov 22 08:14:28 crc kubenswrapper[4789]: I1122 08:14:28.707343 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"bd7daba2-62fe-48dc-8e83-a3764c57f9ba","Type":"ContainerStarted","Data":"392bd52d69af77337dc6a9c51ad010b1fa6ba70be6b8a408321615e2aaf1b7e5"} Nov 22 08:14:34 crc kubenswrapper[4789]: E1122 08:14:34.545476 4789 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:18f8463fe46fe6081d5682009e92bbcb3df33282b83b0a2857abaece795cf1ba" Nov 22 08:14:34 crc kubenswrapper[4789]: E1122 08:14:34.546232 4789 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:18f8463fe46fe6081d5682009e92bbcb3df33282b83b0a2857abaece795cf1ba,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-llqn6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-7c6d9948dc-2dpcz_openstack(e32fe7bc-c0c8-443b-af3e-d9cc58db4f24): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 22 08:14:34 crc kubenswrapper[4789]: E1122 08:14:34.547793 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-7c6d9948dc-2dpcz" podUID="e32fe7bc-c0c8-443b-af3e-d9cc58db4f24" Nov 22 08:14:34 crc kubenswrapper[4789]: E1122 08:14:34.571826 4789 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:18f8463fe46fe6081d5682009e92bbcb3df33282b83b0a2857abaece795cf1ba" Nov 22 08:14:34 crc kubenswrapper[4789]: E1122 08:14:34.572039 4789 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:18f8463fe46fe6081d5682009e92bbcb3df33282b83b0a2857abaece795cf1ba,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-j9nj2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-6486446b9f-2w9gc_openstack(66b9aef5-3c42-4449-987c-9087653553b5): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 22 08:14:34 crc kubenswrapper[4789]: E1122 08:14:34.573550 4789 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:18f8463fe46fe6081d5682009e92bbcb3df33282b83b0a2857abaece795cf1ba" Nov 22 08:14:34 crc kubenswrapper[4789]: E1122 08:14:34.574477 4789 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:18f8463fe46fe6081d5682009e92bbcb3df33282b83b0a2857abaece795cf1ba,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-c6zvz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-6584b49599-wzdnt_openstack(79936918-3fbc-41b6-9c9c-31757e4a6fc6): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 22 08:14:34 crc kubenswrapper[4789]: E1122 08:14:34.576109 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-6584b49599-wzdnt" podUID="79936918-3fbc-41b6-9c9c-31757e4a6fc6" Nov 22 08:14:34 crc kubenswrapper[4789]: E1122 08:14:34.576184 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-6486446b9f-2w9gc" podUID="66b9aef5-3c42-4449-987c-9087653553b5" Nov 22 08:14:34 crc kubenswrapper[4789]: E1122 08:14:34.614830 4789 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:18f8463fe46fe6081d5682009e92bbcb3df33282b83b0a2857abaece795cf1ba" Nov 22 08:14:34 crc kubenswrapper[4789]: E1122 08:14:34.614984 4789 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:18f8463fe46fe6081d5682009e92bbcb3df33282b83b0a2857abaece795cf1ba,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-b4x4q,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-7bdd77c89-nq45j_openstack(f235f4f5-490c-411c-9a1d-452bae11b2eb): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 22 08:14:34 crc kubenswrapper[4789]: E1122 08:14:34.616586 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-7bdd77c89-nq45j" podUID="f235f4f5-490c-411c-9a1d-452bae11b2eb" Nov 22 08:14:34 crc kubenswrapper[4789]: E1122 08:14:34.753936 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:18f8463fe46fe6081d5682009e92bbcb3df33282b83b0a2857abaece795cf1ba\\\"\"" pod="openstack/dnsmasq-dns-7c6d9948dc-2dpcz" podUID="e32fe7bc-c0c8-443b-af3e-d9cc58db4f24" Nov 22 08:14:34 crc kubenswrapper[4789]: E1122 08:14:34.754780 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:18f8463fe46fe6081d5682009e92bbcb3df33282b83b0a2857abaece795cf1ba\\\"\"" pod="openstack/dnsmasq-dns-6486446b9f-2w9gc" podUID="66b9aef5-3c42-4449-987c-9087653553b5" Nov 22 08:14:34 crc kubenswrapper[4789]: I1122 08:14:34.992947 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-cl8m4"] Nov 22 08:14:35 crc kubenswrapper[4789]: I1122 08:14:35.182260 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 22 08:14:35 crc kubenswrapper[4789]: I1122 08:14:35.273884 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 22 08:14:41 crc kubenswrapper[4789]: W1122 08:14:41.892306 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb479a8a1_0fba_4a72_b830_2a68e2bd9be4.slice/crio-a2247e5c6a79915b5e1814324e8070fcb5461a00e8cac59780b83e145617878f WatchSource:0}: Error finding container a2247e5c6a79915b5e1814324e8070fcb5461a00e8cac59780b83e145617878f: Status 404 returned error can't find the container with id a2247e5c6a79915b5e1814324e8070fcb5461a00e8cac59780b83e145617878f Nov 22 08:14:42 crc kubenswrapper[4789]: I1122 08:14:42.145390 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6584b49599-wzdnt" Nov 22 08:14:42 crc kubenswrapper[4789]: I1122 08:14:42.303642 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c6zvz\" (UniqueName: \"kubernetes.io/projected/79936918-3fbc-41b6-9c9c-31757e4a6fc6-kube-api-access-c6zvz\") pod \"79936918-3fbc-41b6-9c9c-31757e4a6fc6\" (UID: \"79936918-3fbc-41b6-9c9c-31757e4a6fc6\") " Nov 22 08:14:42 crc kubenswrapper[4789]: I1122 08:14:42.303741 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/79936918-3fbc-41b6-9c9c-31757e4a6fc6-config\") pod \"79936918-3fbc-41b6-9c9c-31757e4a6fc6\" (UID: \"79936918-3fbc-41b6-9c9c-31757e4a6fc6\") " Nov 22 08:14:42 crc kubenswrapper[4789]: I1122 08:14:42.303857 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/79936918-3fbc-41b6-9c9c-31757e4a6fc6-dns-svc\") pod \"79936918-3fbc-41b6-9c9c-31757e4a6fc6\" (UID: \"79936918-3fbc-41b6-9c9c-31757e4a6fc6\") " Nov 22 08:14:42 crc kubenswrapper[4789]: I1122 08:14:42.304678 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/79936918-3fbc-41b6-9c9c-31757e4a6fc6-config" (OuterVolumeSpecName: "config") pod "79936918-3fbc-41b6-9c9c-31757e4a6fc6" (UID: "79936918-3fbc-41b6-9c9c-31757e4a6fc6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:14:42 crc kubenswrapper[4789]: I1122 08:14:42.304797 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/79936918-3fbc-41b6-9c9c-31757e4a6fc6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "79936918-3fbc-41b6-9c9c-31757e4a6fc6" (UID: "79936918-3fbc-41b6-9c9c-31757e4a6fc6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:14:42 crc kubenswrapper[4789]: I1122 08:14:42.309143 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79936918-3fbc-41b6-9c9c-31757e4a6fc6-kube-api-access-c6zvz" (OuterVolumeSpecName: "kube-api-access-c6zvz") pod "79936918-3fbc-41b6-9c9c-31757e4a6fc6" (UID: "79936918-3fbc-41b6-9c9c-31757e4a6fc6"). InnerVolumeSpecName "kube-api-access-c6zvz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:14:42 crc kubenswrapper[4789]: I1122 08:14:42.379728 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 22 08:14:42 crc kubenswrapper[4789]: I1122 08:14:42.384330 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bdd77c89-nq45j" Nov 22 08:14:42 crc kubenswrapper[4789]: I1122 08:14:42.406451 4789 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/79936918-3fbc-41b6-9c9c-31757e4a6fc6-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 22 08:14:42 crc kubenswrapper[4789]: I1122 08:14:42.406526 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c6zvz\" (UniqueName: \"kubernetes.io/projected/79936918-3fbc-41b6-9c9c-31757e4a6fc6-kube-api-access-c6zvz\") on node \"crc\" DevicePath \"\"" Nov 22 08:14:42 crc kubenswrapper[4789]: I1122 08:14:42.406593 4789 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/79936918-3fbc-41b6-9c9c-31757e4a6fc6-config\") on node \"crc\" DevicePath \"\"" Nov 22 08:14:42 crc kubenswrapper[4789]: I1122 08:14:42.508870 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f235f4f5-490c-411c-9a1d-452bae11b2eb-config\") pod \"f235f4f5-490c-411c-9a1d-452bae11b2eb\" (UID: \"f235f4f5-490c-411c-9a1d-452bae11b2eb\") " Nov 22 08:14:42 crc kubenswrapper[4789]: I1122 08:14:42.509258 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b4x4q\" (UniqueName: \"kubernetes.io/projected/f235f4f5-490c-411c-9a1d-452bae11b2eb-kube-api-access-b4x4q\") pod \"f235f4f5-490c-411c-9a1d-452bae11b2eb\" (UID: \"f235f4f5-490c-411c-9a1d-452bae11b2eb\") " Nov 22 08:14:42 crc kubenswrapper[4789]: I1122 08:14:42.510664 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f235f4f5-490c-411c-9a1d-452bae11b2eb-config" (OuterVolumeSpecName: "config") pod "f235f4f5-490c-411c-9a1d-452bae11b2eb" (UID: "f235f4f5-490c-411c-9a1d-452bae11b2eb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:14:42 crc kubenswrapper[4789]: I1122 08:14:42.516132 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-gjjwv"] Nov 22 08:14:42 crc kubenswrapper[4789]: I1122 08:14:42.597783 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f235f4f5-490c-411c-9a1d-452bae11b2eb-kube-api-access-b4x4q" (OuterVolumeSpecName: "kube-api-access-b4x4q") pod "f235f4f5-490c-411c-9a1d-452bae11b2eb" (UID: "f235f4f5-490c-411c-9a1d-452bae11b2eb"). InnerVolumeSpecName "kube-api-access-b4x4q". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:14:42 crc kubenswrapper[4789]: I1122 08:14:42.612983 4789 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f235f4f5-490c-411c-9a1d-452bae11b2eb-config\") on node \"crc\" DevicePath \"\"" Nov 22 08:14:42 crc kubenswrapper[4789]: I1122 08:14:42.613015 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b4x4q\" (UniqueName: \"kubernetes.io/projected/f235f4f5-490c-411c-9a1d-452bae11b2eb-kube-api-access-b4x4q\") on node \"crc\" DevicePath \"\"" Nov 22 08:14:42 crc kubenswrapper[4789]: I1122 08:14:42.817312 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6584b49599-wzdnt" event={"ID":"79936918-3fbc-41b6-9c9c-31757e4a6fc6","Type":"ContainerDied","Data":"bb243ce6a40860b4c430b1230bf4b32fa8e8c774a4f42356768a69b1b12a6d7a"} Nov 22 08:14:42 crc kubenswrapper[4789]: I1122 08:14:42.817409 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6584b49599-wzdnt" Nov 22 08:14:42 crc kubenswrapper[4789]: I1122 08:14:42.824690 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"20a309c0-4fac-480e-b19d-e2ca480a6b6c","Type":"ContainerStarted","Data":"35ebde5382595cf8e01a5138ab3b0185954a6bcd09892ad0565fa7abbcf20083"} Nov 22 08:14:42 crc kubenswrapper[4789]: I1122 08:14:42.826580 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-cl8m4" event={"ID":"e6f04e32-3dca-4d6d-8c8d-19ea15f46875","Type":"ContainerStarted","Data":"fc8064f7f067e6d0e94c00d7f53341762b5f36a673c2815d162023b49dc6f860"} Nov 22 08:14:42 crc kubenswrapper[4789]: I1122 08:14:42.830129 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"7778e90a-45cf-4592-ab3d-68c75258e77f","Type":"ContainerStarted","Data":"d7f3371f37d974bca9431b646921bf07a58c48cac9257957026636d71074f44a"} Nov 22 08:14:42 crc kubenswrapper[4789]: I1122 08:14:42.832146 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-gjjwv" event={"ID":"ea7bbfd1-bdb6-41b2-87f5-2fa2bbe0c804","Type":"ContainerStarted","Data":"54196c5860ec1652f19b3f03aeab4525193f85c282c2d460c4352d671b9a3c89"} Nov 22 08:14:42 crc kubenswrapper[4789]: I1122 08:14:42.834449 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"c2d44daa-85ba-473a-a29f-2f2587bfd2fa","Type":"ContainerStarted","Data":"5d15ab07cc697aa469f00e4a7b8a7654fd8f31ee3b23783cc09bc42215d760b4"} Nov 22 08:14:42 crc kubenswrapper[4789]: I1122 08:14:42.841423 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bdd77c89-nq45j" event={"ID":"f235f4f5-490c-411c-9a1d-452bae11b2eb","Type":"ContainerDied","Data":"e53dd5d61e2512f60bcff635d2a631f5e0eddf6a64deca3b00f3780d16962654"} Nov 22 08:14:42 crc kubenswrapper[4789]: I1122 08:14:42.841513 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bdd77c89-nq45j" Nov 22 08:14:42 crc kubenswrapper[4789]: I1122 08:14:42.843367 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"51737c30-e744-4266-a235-1ddd1c115c4e","Type":"ContainerStarted","Data":"33d695a2ed2acc937455627b156a150d63f1b1ea5504a7291a8492dfa0057ff4"} Nov 22 08:14:42 crc kubenswrapper[4789]: I1122 08:14:42.846504 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"bd7daba2-62fe-48dc-8e83-a3764c57f9ba","Type":"ContainerStarted","Data":"5b957f4fae8ca653bafc627989ba49dfb91fbb03c46b4dab0b8bb04bf744fde1"} Nov 22 08:14:42 crc kubenswrapper[4789]: I1122 08:14:42.846658 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Nov 22 08:14:42 crc kubenswrapper[4789]: I1122 08:14:42.848258 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"b479a8a1-0fba-4a72-b830-2a68e2bd9be4","Type":"ContainerStarted","Data":"a2247e5c6a79915b5e1814324e8070fcb5461a00e8cac59780b83e145617878f"} Nov 22 08:14:42 crc kubenswrapper[4789]: I1122 08:14:42.930662 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=11.875907145 podStartE2EDuration="25.930642219s" podCreationTimestamp="2025-11-22 08:14:17 +0000 UTC" firstStartedPulling="2025-11-22 08:14:27.922643938 +0000 UTC m=+1082.157044211" lastFinishedPulling="2025-11-22 08:14:41.977379012 +0000 UTC m=+1096.211779285" observedRunningTime="2025-11-22 08:14:42.900553101 +0000 UTC m=+1097.134953394" watchObservedRunningTime="2025-11-22 08:14:42.930642219 +0000 UTC m=+1097.165042492" Nov 22 08:14:42 crc kubenswrapper[4789]: I1122 08:14:42.945508 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6584b49599-wzdnt"] Nov 22 08:14:42 crc kubenswrapper[4789]: I1122 08:14:42.950265 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6584b49599-wzdnt"] Nov 22 08:14:42 crc kubenswrapper[4789]: I1122 08:14:42.977030 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7bdd77c89-nq45j"] Nov 22 08:14:42 crc kubenswrapper[4789]: I1122 08:14:42.981728 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7bdd77c89-nq45j"] Nov 22 08:14:43 crc kubenswrapper[4789]: I1122 08:14:43.880553 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"df1ccc5e-c20c-40ce-bb76-17da02c4bc8c","Type":"ContainerStarted","Data":"b52c35865cafb51202b9dc7fd7ca36b2d1afa82daff2e035973c3e4405cd63b2"} Nov 22 08:14:43 crc kubenswrapper[4789]: I1122 08:14:43.885956 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"af27dae6-c617-4d16-bb53-19057df7689b","Type":"ContainerStarted","Data":"e7ed5333a86c250ae63dba829dea4c481af2d77e17c98c355f0fc1115c16d37c"} Nov 22 08:14:43 crc kubenswrapper[4789]: I1122 08:14:43.976891 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="79936918-3fbc-41b6-9c9c-31757e4a6fc6" path="/var/lib/kubelet/pods/79936918-3fbc-41b6-9c9c-31757e4a6fc6/volumes" Nov 22 08:14:43 crc kubenswrapper[4789]: I1122 08:14:43.977517 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f235f4f5-490c-411c-9a1d-452bae11b2eb" path="/var/lib/kubelet/pods/f235f4f5-490c-411c-9a1d-452bae11b2eb/volumes" Nov 22 08:14:46 crc kubenswrapper[4789]: I1122 08:14:46.906206 4789 generic.go:334] "Generic (PLEG): container finished" podID="51737c30-e744-4266-a235-1ddd1c115c4e" containerID="33d695a2ed2acc937455627b156a150d63f1b1ea5504a7291a8492dfa0057ff4" exitCode=0 Nov 22 08:14:46 crc kubenswrapper[4789]: I1122 08:14:46.906292 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"51737c30-e744-4266-a235-1ddd1c115c4e","Type":"ContainerDied","Data":"33d695a2ed2acc937455627b156a150d63f1b1ea5504a7291a8492dfa0057ff4"} Nov 22 08:14:46 crc kubenswrapper[4789]: I1122 08:14:46.909407 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"c2d44daa-85ba-473a-a29f-2f2587bfd2fa","Type":"ContainerStarted","Data":"5934c831e5ac971068391d8eae11c99de866f9730f20b4589ef524cc941e8925"} Nov 22 08:14:46 crc kubenswrapper[4789]: I1122 08:14:46.910901 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"b479a8a1-0fba-4a72-b830-2a68e2bd9be4","Type":"ContainerStarted","Data":"e5a7462789382347a7131dd361c3d7d327e655110338393d4c8b6bf390a10581"} Nov 22 08:14:46 crc kubenswrapper[4789]: I1122 08:14:46.912278 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"20a309c0-4fac-480e-b19d-e2ca480a6b6c","Type":"ContainerStarted","Data":"39f28765bd714106b4fd5010156d2c44ca1777ed061f27f9555f21c3d40d7898"} Nov 22 08:14:46 crc kubenswrapper[4789]: I1122 08:14:46.912442 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Nov 22 08:14:46 crc kubenswrapper[4789]: I1122 08:14:46.913557 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-cl8m4" event={"ID":"e6f04e32-3dca-4d6d-8c8d-19ea15f46875","Type":"ContainerStarted","Data":"69b36fb458f52b4942a142b1e419c9070aee17243c01c2f38d55a3927fb2471b"} Nov 22 08:14:46 crc kubenswrapper[4789]: I1122 08:14:46.913608 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-cl8m4" Nov 22 08:14:46 crc kubenswrapper[4789]: I1122 08:14:46.915161 4789 generic.go:334] "Generic (PLEG): container finished" podID="7778e90a-45cf-4592-ab3d-68c75258e77f" containerID="d7f3371f37d974bca9431b646921bf07a58c48cac9257957026636d71074f44a" exitCode=0 Nov 22 08:14:46 crc kubenswrapper[4789]: I1122 08:14:46.915186 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"7778e90a-45cf-4592-ab3d-68c75258e77f","Type":"ContainerDied","Data":"d7f3371f37d974bca9431b646921bf07a58c48cac9257957026636d71074f44a"} Nov 22 08:14:46 crc kubenswrapper[4789]: I1122 08:14:46.916905 4789 generic.go:334] "Generic (PLEG): container finished" podID="ea7bbfd1-bdb6-41b2-87f5-2fa2bbe0c804" containerID="2f7a4e3253eb8433e52127426b63ae3e8beaf3eb88f54315b104723521300732" exitCode=0 Nov 22 08:14:46 crc kubenswrapper[4789]: I1122 08:14:46.916974 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-gjjwv" event={"ID":"ea7bbfd1-bdb6-41b2-87f5-2fa2bbe0c804","Type":"ContainerDied","Data":"2f7a4e3253eb8433e52127426b63ae3e8beaf3eb88f54315b104723521300732"} Nov 22 08:14:46 crc kubenswrapper[4789]: I1122 08:14:46.955710 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-cl8m4" podStartSLOduration=19.492999173 podStartE2EDuration="23.955693442s" podCreationTimestamp="2025-11-22 08:14:23 +0000 UTC" firstStartedPulling="2025-11-22 08:14:41.888017131 +0000 UTC m=+1096.122417404" lastFinishedPulling="2025-11-22 08:14:46.3507114 +0000 UTC m=+1100.585111673" observedRunningTime="2025-11-22 08:14:46.95119336 +0000 UTC m=+1101.185593633" watchObservedRunningTime="2025-11-22 08:14:46.955693442 +0000 UTC m=+1101.190093715" Nov 22 08:14:46 crc kubenswrapper[4789]: I1122 08:14:46.981178 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=23.45836381 podStartE2EDuration="27.981150756s" podCreationTimestamp="2025-11-22 08:14:19 +0000 UTC" firstStartedPulling="2025-11-22 08:14:41.890426007 +0000 UTC m=+1096.124826290" lastFinishedPulling="2025-11-22 08:14:46.413212963 +0000 UTC m=+1100.647613236" observedRunningTime="2025-11-22 08:14:46.967926056 +0000 UTC m=+1101.202326329" watchObservedRunningTime="2025-11-22 08:14:46.981150756 +0000 UTC m=+1101.215551039" Nov 22 08:14:47 crc kubenswrapper[4789]: I1122 08:14:47.928896 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"7778e90a-45cf-4592-ab3d-68c75258e77f","Type":"ContainerStarted","Data":"457c7b6970858f76eae1ebef0e72149d82cc972a01c6db89dadcfb3d19b98a68"} Nov 22 08:14:47 crc kubenswrapper[4789]: I1122 08:14:47.932730 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-gjjwv" event={"ID":"ea7bbfd1-bdb6-41b2-87f5-2fa2bbe0c804","Type":"ContainerStarted","Data":"67b136bca1246a0951fc02e1c7e01a72fa22f5419383c7ccc0f46339b1a3c71e"} Nov 22 08:14:47 crc kubenswrapper[4789]: I1122 08:14:47.932790 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-gjjwv" event={"ID":"ea7bbfd1-bdb6-41b2-87f5-2fa2bbe0c804","Type":"ContainerStarted","Data":"1b989084e5dab59b4a483185d946bb052626ad8a2ee7a41fc1f5fbe5477abbf1"} Nov 22 08:14:47 crc kubenswrapper[4789]: I1122 08:14:47.932835 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-gjjwv" Nov 22 08:14:47 crc kubenswrapper[4789]: I1122 08:14:47.932943 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-gjjwv" Nov 22 08:14:47 crc kubenswrapper[4789]: I1122 08:14:47.935032 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"51737c30-e744-4266-a235-1ddd1c115c4e","Type":"ContainerStarted","Data":"914867369cdd970b1bef0052800036a8e7447ce3af496680a98b6d1587e56c8a"} Nov 22 08:14:47 crc kubenswrapper[4789]: I1122 08:14:47.953525 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=12.905173395 podStartE2EDuration="31.953507786s" podCreationTimestamp="2025-11-22 08:14:16 +0000 UTC" firstStartedPulling="2025-11-22 08:14:22.836210636 +0000 UTC m=+1077.070610919" lastFinishedPulling="2025-11-22 08:14:41.884545037 +0000 UTC m=+1096.118945310" observedRunningTime="2025-11-22 08:14:47.952635473 +0000 UTC m=+1102.187035756" watchObservedRunningTime="2025-11-22 08:14:47.953507786 +0000 UTC m=+1102.187908059" Nov 22 08:14:47 crc kubenswrapper[4789]: I1122 08:14:47.977806 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-gjjwv" podStartSLOduration=21.567654142 podStartE2EDuration="24.977788078s" podCreationTimestamp="2025-11-22 08:14:23 +0000 UTC" firstStartedPulling="2025-11-22 08:14:42.605703381 +0000 UTC m=+1096.840103654" lastFinishedPulling="2025-11-22 08:14:46.015837317 +0000 UTC m=+1100.250237590" observedRunningTime="2025-11-22 08:14:47.971322442 +0000 UTC m=+1102.205722735" watchObservedRunningTime="2025-11-22 08:14:47.977788078 +0000 UTC m=+1102.212188351" Nov 22 08:14:47 crc kubenswrapper[4789]: I1122 08:14:47.993234 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=16.238461243 podStartE2EDuration="33.993218569s" podCreationTimestamp="2025-11-22 08:14:14 +0000 UTC" firstStartedPulling="2025-11-22 08:14:16.839539649 +0000 UTC m=+1071.073939922" lastFinishedPulling="2025-11-22 08:14:34.594296975 +0000 UTC m=+1088.828697248" observedRunningTime="2025-11-22 08:14:47.990818013 +0000 UTC m=+1102.225218286" watchObservedRunningTime="2025-11-22 08:14:47.993218569 +0000 UTC m=+1102.227618842" Nov 22 08:14:48 crc kubenswrapper[4789]: I1122 08:14:48.124514 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Nov 22 08:14:50 crc kubenswrapper[4789]: E1122 08:14:50.318520 4789 upgradeaware.go:441] Error proxying data from backend to client: writeto tcp 38.129.56.56:54352->38.129.56.56:37377: read tcp 38.129.56.56:54352->38.129.56.56:37377: read: connection reset by peer Nov 22 08:14:50 crc kubenswrapper[4789]: E1122 08:14:50.318543 4789 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.129.56.56:54352->38.129.56.56:37377: write tcp 38.129.56.56:54352->38.129.56.56:37377: write: broken pipe Nov 22 08:14:50 crc kubenswrapper[4789]: I1122 08:14:50.968617 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"c2d44daa-85ba-473a-a29f-2f2587bfd2fa","Type":"ContainerStarted","Data":"24b184b0f96da505559a622488c439ffd5a687549cb9ddaad2920334de7d21c6"} Nov 22 08:14:50 crc kubenswrapper[4789]: I1122 08:14:50.970066 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"b479a8a1-0fba-4a72-b830-2a68e2bd9be4","Type":"ContainerStarted","Data":"20136ef231dded82429a8166fd1d78a25c786d1535791bb1db9789596b9686e2"} Nov 22 08:14:50 crc kubenswrapper[4789]: I1122 08:14:50.971657 4789 generic.go:334] "Generic (PLEG): container finished" podID="e32fe7bc-c0c8-443b-af3e-d9cc58db4f24" containerID="aa072d21743fd61b576d25e52ac024348caaf4f7d8a62f95f279868643113ed8" exitCode=0 Nov 22 08:14:50 crc kubenswrapper[4789]: I1122 08:14:50.971717 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c6d9948dc-2dpcz" event={"ID":"e32fe7bc-c0c8-443b-af3e-d9cc58db4f24","Type":"ContainerDied","Data":"aa072d21743fd61b576d25e52ac024348caaf4f7d8a62f95f279868643113ed8"} Nov 22 08:14:50 crc kubenswrapper[4789]: I1122 08:14:50.973435 4789 generic.go:334] "Generic (PLEG): container finished" podID="66b9aef5-3c42-4449-987c-9087653553b5" containerID="670465358e9760832a3d128dd0dd08cd1e21b7cd4bea04525a642eac7ddda14d" exitCode=0 Nov 22 08:14:50 crc kubenswrapper[4789]: I1122 08:14:50.973475 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6486446b9f-2w9gc" event={"ID":"66b9aef5-3c42-4449-987c-9087653553b5","Type":"ContainerDied","Data":"670465358e9760832a3d128dd0dd08cd1e21b7cd4bea04525a642eac7ddda14d"} Nov 22 08:14:50 crc kubenswrapper[4789]: I1122 08:14:50.992884 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=17.772844754 podStartE2EDuration="25.992870241s" podCreationTimestamp="2025-11-22 08:14:25 +0000 UTC" firstStartedPulling="2025-11-22 08:14:42.392443421 +0000 UTC m=+1096.626843694" lastFinishedPulling="2025-11-22 08:14:50.612468908 +0000 UTC m=+1104.846869181" observedRunningTime="2025-11-22 08:14:50.987729872 +0000 UTC m=+1105.222130155" watchObservedRunningTime="2025-11-22 08:14:50.992870241 +0000 UTC m=+1105.227270514" Nov 22 08:14:51 crc kubenswrapper[4789]: I1122 08:14:51.017423 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=20.361593031 podStartE2EDuration="29.01740357s" podCreationTimestamp="2025-11-22 08:14:22 +0000 UTC" firstStartedPulling="2025-11-22 08:14:41.940741025 +0000 UTC m=+1096.175141298" lastFinishedPulling="2025-11-22 08:14:50.596551564 +0000 UTC m=+1104.830951837" observedRunningTime="2025-11-22 08:14:51.01484121 +0000 UTC m=+1105.249241483" watchObservedRunningTime="2025-11-22 08:14:51.01740357 +0000 UTC m=+1105.251803853" Nov 22 08:14:51 crc kubenswrapper[4789]: I1122 08:14:51.325030 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Nov 22 08:14:51 crc kubenswrapper[4789]: I1122 08:14:51.363814 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Nov 22 08:14:51 crc kubenswrapper[4789]: I1122 08:14:51.987320 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6486446b9f-2w9gc" event={"ID":"66b9aef5-3c42-4449-987c-9087653553b5","Type":"ContainerStarted","Data":"6b0f95e060c81173508b35aea811dae2268821ef6427dd3647924e9202f650c1"} Nov 22 08:14:51 crc kubenswrapper[4789]: I1122 08:14:51.987811 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6486446b9f-2w9gc" Nov 22 08:14:51 crc kubenswrapper[4789]: I1122 08:14:51.989293 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c6d9948dc-2dpcz" event={"ID":"e32fe7bc-c0c8-443b-af3e-d9cc58db4f24","Type":"ContainerStarted","Data":"4c86ea65c880f7ffe8e1357a4462d7ace08f91605d53b35fb585f327d1976286"} Nov 22 08:14:51 crc kubenswrapper[4789]: I1122 08:14:51.990109 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Nov 22 08:14:51 crc kubenswrapper[4789]: I1122 08:14:51.990127 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7c6d9948dc-2dpcz" Nov 22 08:14:52 crc kubenswrapper[4789]: I1122 08:14:52.011740 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6486446b9f-2w9gc" podStartSLOduration=2.761252844 podStartE2EDuration="39.011720859s" podCreationTimestamp="2025-11-22 08:14:13 +0000 UTC" firstStartedPulling="2025-11-22 08:14:14.412100677 +0000 UTC m=+1068.646500950" lastFinishedPulling="2025-11-22 08:14:50.662568692 +0000 UTC m=+1104.896968965" observedRunningTime="2025-11-22 08:14:52.002087537 +0000 UTC m=+1106.236487810" watchObservedRunningTime="2025-11-22 08:14:52.011720859 +0000 UTC m=+1106.246121162" Nov 22 08:14:52 crc kubenswrapper[4789]: I1122 08:14:52.025083 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7c6d9948dc-2dpcz" podStartSLOduration=2.4280631010000002 podStartE2EDuration="39.025067792s" podCreationTimestamp="2025-11-22 08:14:13 +0000 UTC" firstStartedPulling="2025-11-22 08:14:14.063739482 +0000 UTC m=+1068.298139755" lastFinishedPulling="2025-11-22 08:14:50.660744173 +0000 UTC m=+1104.895144446" observedRunningTime="2025-11-22 08:14:52.024096157 +0000 UTC m=+1106.258496440" watchObservedRunningTime="2025-11-22 08:14:52.025067792 +0000 UTC m=+1106.259468065" Nov 22 08:14:52 crc kubenswrapper[4789]: I1122 08:14:52.033059 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Nov 22 08:14:52 crc kubenswrapper[4789]: I1122 08:14:52.275517 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6486446b9f-2w9gc"] Nov 22 08:14:52 crc kubenswrapper[4789]: I1122 08:14:52.302597 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-65c9b8d4f7-kfsdh"] Nov 22 08:14:52 crc kubenswrapper[4789]: I1122 08:14:52.303960 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-65c9b8d4f7-kfsdh" Nov 22 08:14:52 crc kubenswrapper[4789]: I1122 08:14:52.313418 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-c6fhr"] Nov 22 08:14:52 crc kubenswrapper[4789]: I1122 08:14:52.314641 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-c6fhr" Nov 22 08:14:52 crc kubenswrapper[4789]: I1122 08:14:52.315388 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Nov 22 08:14:52 crc kubenswrapper[4789]: I1122 08:14:52.324125 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Nov 22 08:14:52 crc kubenswrapper[4789]: I1122 08:14:52.328021 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-65c9b8d4f7-kfsdh"] Nov 22 08:14:52 crc kubenswrapper[4789]: I1122 08:14:52.351733 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-c6fhr"] Nov 22 08:14:52 crc kubenswrapper[4789]: I1122 08:14:52.487994 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2e9862f9-e69a-472f-9db9-638c95ee3912-config\") pod \"dnsmasq-dns-65c9b8d4f7-kfsdh\" (UID: \"2e9862f9-e69a-472f-9db9-638c95ee3912\") " pod="openstack/dnsmasq-dns-65c9b8d4f7-kfsdh" Nov 22 08:14:52 crc kubenswrapper[4789]: I1122 08:14:52.488365 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h5hkh\" (UniqueName: \"kubernetes.io/projected/2e9862f9-e69a-472f-9db9-638c95ee3912-kube-api-access-h5hkh\") pod \"dnsmasq-dns-65c9b8d4f7-kfsdh\" (UID: \"2e9862f9-e69a-472f-9db9-638c95ee3912\") " pod="openstack/dnsmasq-dns-65c9b8d4f7-kfsdh" Nov 22 08:14:52 crc kubenswrapper[4789]: I1122 08:14:52.488425 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2e9862f9-e69a-472f-9db9-638c95ee3912-dns-svc\") pod \"dnsmasq-dns-65c9b8d4f7-kfsdh\" (UID: \"2e9862f9-e69a-472f-9db9-638c95ee3912\") " pod="openstack/dnsmasq-dns-65c9b8d4f7-kfsdh" Nov 22 08:14:52 crc kubenswrapper[4789]: I1122 08:14:52.488470 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0e0f4cec-d122-4b44-8344-1cd6e515ef19-config\") pod \"ovn-controller-metrics-c6fhr\" (UID: \"0e0f4cec-d122-4b44-8344-1cd6e515ef19\") " pod="openstack/ovn-controller-metrics-c6fhr" Nov 22 08:14:52 crc kubenswrapper[4789]: I1122 08:14:52.488530 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/0e0f4cec-d122-4b44-8344-1cd6e515ef19-ovn-rundir\") pod \"ovn-controller-metrics-c6fhr\" (UID: \"0e0f4cec-d122-4b44-8344-1cd6e515ef19\") " pod="openstack/ovn-controller-metrics-c6fhr" Nov 22 08:14:52 crc kubenswrapper[4789]: I1122 08:14:52.488628 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e0f4cec-d122-4b44-8344-1cd6e515ef19-combined-ca-bundle\") pod \"ovn-controller-metrics-c6fhr\" (UID: \"0e0f4cec-d122-4b44-8344-1cd6e515ef19\") " pod="openstack/ovn-controller-metrics-c6fhr" Nov 22 08:14:52 crc kubenswrapper[4789]: I1122 08:14:52.488667 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmgl6\" (UniqueName: \"kubernetes.io/projected/0e0f4cec-d122-4b44-8344-1cd6e515ef19-kube-api-access-vmgl6\") pod \"ovn-controller-metrics-c6fhr\" (UID: \"0e0f4cec-d122-4b44-8344-1cd6e515ef19\") " pod="openstack/ovn-controller-metrics-c6fhr" Nov 22 08:14:52 crc kubenswrapper[4789]: I1122 08:14:52.488778 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2e9862f9-e69a-472f-9db9-638c95ee3912-ovsdbserver-sb\") pod \"dnsmasq-dns-65c9b8d4f7-kfsdh\" (UID: \"2e9862f9-e69a-472f-9db9-638c95ee3912\") " pod="openstack/dnsmasq-dns-65c9b8d4f7-kfsdh" Nov 22 08:14:52 crc kubenswrapper[4789]: I1122 08:14:52.488856 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/0e0f4cec-d122-4b44-8344-1cd6e515ef19-ovs-rundir\") pod \"ovn-controller-metrics-c6fhr\" (UID: \"0e0f4cec-d122-4b44-8344-1cd6e515ef19\") " pod="openstack/ovn-controller-metrics-c6fhr" Nov 22 08:14:52 crc kubenswrapper[4789]: I1122 08:14:52.488885 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/0e0f4cec-d122-4b44-8344-1cd6e515ef19-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-c6fhr\" (UID: \"0e0f4cec-d122-4b44-8344-1cd6e515ef19\") " pod="openstack/ovn-controller-metrics-c6fhr" Nov 22 08:14:52 crc kubenswrapper[4789]: I1122 08:14:52.590449 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e0f4cec-d122-4b44-8344-1cd6e515ef19-combined-ca-bundle\") pod \"ovn-controller-metrics-c6fhr\" (UID: \"0e0f4cec-d122-4b44-8344-1cd6e515ef19\") " pod="openstack/ovn-controller-metrics-c6fhr" Nov 22 08:14:52 crc kubenswrapper[4789]: I1122 08:14:52.590491 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmgl6\" (UniqueName: \"kubernetes.io/projected/0e0f4cec-d122-4b44-8344-1cd6e515ef19-kube-api-access-vmgl6\") pod \"ovn-controller-metrics-c6fhr\" (UID: \"0e0f4cec-d122-4b44-8344-1cd6e515ef19\") " pod="openstack/ovn-controller-metrics-c6fhr" Nov 22 08:14:52 crc kubenswrapper[4789]: I1122 08:14:52.590511 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2e9862f9-e69a-472f-9db9-638c95ee3912-ovsdbserver-sb\") pod \"dnsmasq-dns-65c9b8d4f7-kfsdh\" (UID: \"2e9862f9-e69a-472f-9db9-638c95ee3912\") " pod="openstack/dnsmasq-dns-65c9b8d4f7-kfsdh" Nov 22 08:14:52 crc kubenswrapper[4789]: I1122 08:14:52.590544 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/0e0f4cec-d122-4b44-8344-1cd6e515ef19-ovs-rundir\") pod \"ovn-controller-metrics-c6fhr\" (UID: \"0e0f4cec-d122-4b44-8344-1cd6e515ef19\") " pod="openstack/ovn-controller-metrics-c6fhr" Nov 22 08:14:52 crc kubenswrapper[4789]: I1122 08:14:52.590560 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/0e0f4cec-d122-4b44-8344-1cd6e515ef19-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-c6fhr\" (UID: \"0e0f4cec-d122-4b44-8344-1cd6e515ef19\") " pod="openstack/ovn-controller-metrics-c6fhr" Nov 22 08:14:52 crc kubenswrapper[4789]: I1122 08:14:52.590583 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2e9862f9-e69a-472f-9db9-638c95ee3912-config\") pod \"dnsmasq-dns-65c9b8d4f7-kfsdh\" (UID: \"2e9862f9-e69a-472f-9db9-638c95ee3912\") " pod="openstack/dnsmasq-dns-65c9b8d4f7-kfsdh" Nov 22 08:14:52 crc kubenswrapper[4789]: I1122 08:14:52.590598 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h5hkh\" (UniqueName: \"kubernetes.io/projected/2e9862f9-e69a-472f-9db9-638c95ee3912-kube-api-access-h5hkh\") pod \"dnsmasq-dns-65c9b8d4f7-kfsdh\" (UID: \"2e9862f9-e69a-472f-9db9-638c95ee3912\") " pod="openstack/dnsmasq-dns-65c9b8d4f7-kfsdh" Nov 22 08:14:52 crc kubenswrapper[4789]: I1122 08:14:52.590631 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2e9862f9-e69a-472f-9db9-638c95ee3912-dns-svc\") pod \"dnsmasq-dns-65c9b8d4f7-kfsdh\" (UID: \"2e9862f9-e69a-472f-9db9-638c95ee3912\") " pod="openstack/dnsmasq-dns-65c9b8d4f7-kfsdh" Nov 22 08:14:52 crc kubenswrapper[4789]: I1122 08:14:52.590652 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0e0f4cec-d122-4b44-8344-1cd6e515ef19-config\") pod \"ovn-controller-metrics-c6fhr\" (UID: \"0e0f4cec-d122-4b44-8344-1cd6e515ef19\") " pod="openstack/ovn-controller-metrics-c6fhr" Nov 22 08:14:52 crc kubenswrapper[4789]: I1122 08:14:52.590666 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/0e0f4cec-d122-4b44-8344-1cd6e515ef19-ovn-rundir\") pod \"ovn-controller-metrics-c6fhr\" (UID: \"0e0f4cec-d122-4b44-8344-1cd6e515ef19\") " pod="openstack/ovn-controller-metrics-c6fhr" Nov 22 08:14:52 crc kubenswrapper[4789]: I1122 08:14:52.596075 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2e9862f9-e69a-472f-9db9-638c95ee3912-dns-svc\") pod \"dnsmasq-dns-65c9b8d4f7-kfsdh\" (UID: \"2e9862f9-e69a-472f-9db9-638c95ee3912\") " pod="openstack/dnsmasq-dns-65c9b8d4f7-kfsdh" Nov 22 08:14:52 crc kubenswrapper[4789]: I1122 08:14:52.596101 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/0e0f4cec-d122-4b44-8344-1cd6e515ef19-ovs-rundir\") pod \"ovn-controller-metrics-c6fhr\" (UID: \"0e0f4cec-d122-4b44-8344-1cd6e515ef19\") " pod="openstack/ovn-controller-metrics-c6fhr" Nov 22 08:14:52 crc kubenswrapper[4789]: I1122 08:14:52.596367 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2e9862f9-e69a-472f-9db9-638c95ee3912-config\") pod \"dnsmasq-dns-65c9b8d4f7-kfsdh\" (UID: \"2e9862f9-e69a-472f-9db9-638c95ee3912\") " pod="openstack/dnsmasq-dns-65c9b8d4f7-kfsdh" Nov 22 08:14:52 crc kubenswrapper[4789]: I1122 08:14:52.596369 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2e9862f9-e69a-472f-9db9-638c95ee3912-ovsdbserver-sb\") pod \"dnsmasq-dns-65c9b8d4f7-kfsdh\" (UID: \"2e9862f9-e69a-472f-9db9-638c95ee3912\") " pod="openstack/dnsmasq-dns-65c9b8d4f7-kfsdh" Nov 22 08:14:52 crc kubenswrapper[4789]: I1122 08:14:52.596960 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0e0f4cec-d122-4b44-8344-1cd6e515ef19-config\") pod \"ovn-controller-metrics-c6fhr\" (UID: \"0e0f4cec-d122-4b44-8344-1cd6e515ef19\") " pod="openstack/ovn-controller-metrics-c6fhr" Nov 22 08:14:52 crc kubenswrapper[4789]: I1122 08:14:52.597688 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/0e0f4cec-d122-4b44-8344-1cd6e515ef19-ovn-rundir\") pod \"ovn-controller-metrics-c6fhr\" (UID: \"0e0f4cec-d122-4b44-8344-1cd6e515ef19\") " pod="openstack/ovn-controller-metrics-c6fhr" Nov 22 08:14:52 crc kubenswrapper[4789]: I1122 08:14:52.601624 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e0f4cec-d122-4b44-8344-1cd6e515ef19-combined-ca-bundle\") pod \"ovn-controller-metrics-c6fhr\" (UID: \"0e0f4cec-d122-4b44-8344-1cd6e515ef19\") " pod="openstack/ovn-controller-metrics-c6fhr" Nov 22 08:14:52 crc kubenswrapper[4789]: I1122 08:14:52.609353 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/0e0f4cec-d122-4b44-8344-1cd6e515ef19-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-c6fhr\" (UID: \"0e0f4cec-d122-4b44-8344-1cd6e515ef19\") " pod="openstack/ovn-controller-metrics-c6fhr" Nov 22 08:14:52 crc kubenswrapper[4789]: I1122 08:14:52.615292 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmgl6\" (UniqueName: \"kubernetes.io/projected/0e0f4cec-d122-4b44-8344-1cd6e515ef19-kube-api-access-vmgl6\") pod \"ovn-controller-metrics-c6fhr\" (UID: \"0e0f4cec-d122-4b44-8344-1cd6e515ef19\") " pod="openstack/ovn-controller-metrics-c6fhr" Nov 22 08:14:52 crc kubenswrapper[4789]: I1122 08:14:52.616059 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h5hkh\" (UniqueName: \"kubernetes.io/projected/2e9862f9-e69a-472f-9db9-638c95ee3912-kube-api-access-h5hkh\") pod \"dnsmasq-dns-65c9b8d4f7-kfsdh\" (UID: \"2e9862f9-e69a-472f-9db9-638c95ee3912\") " pod="openstack/dnsmasq-dns-65c9b8d4f7-kfsdh" Nov 22 08:14:52 crc kubenswrapper[4789]: I1122 08:14:52.621398 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-65c9b8d4f7-kfsdh" Nov 22 08:14:52 crc kubenswrapper[4789]: I1122 08:14:52.638547 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-c6fhr" Nov 22 08:14:52 crc kubenswrapper[4789]: I1122 08:14:52.697539 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7c6d9948dc-2dpcz"] Nov 22 08:14:52 crc kubenswrapper[4789]: I1122 08:14:52.714469 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c476d78c5-djb2g"] Nov 22 08:14:52 crc kubenswrapper[4789]: I1122 08:14:52.715624 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c476d78c5-djb2g" Nov 22 08:14:52 crc kubenswrapper[4789]: I1122 08:14:52.722886 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Nov 22 08:14:52 crc kubenswrapper[4789]: I1122 08:14:52.765047 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c476d78c5-djb2g"] Nov 22 08:14:52 crc kubenswrapper[4789]: I1122 08:14:52.896594 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0cf0fb02-e586-4daa-88c1-6556b376c02a-ovsdbserver-sb\") pod \"dnsmasq-dns-5c476d78c5-djb2g\" (UID: \"0cf0fb02-e586-4daa-88c1-6556b376c02a\") " pod="openstack/dnsmasq-dns-5c476d78c5-djb2g" Nov 22 08:14:52 crc kubenswrapper[4789]: I1122 08:14:52.896631 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xhkvg\" (UniqueName: \"kubernetes.io/projected/0cf0fb02-e586-4daa-88c1-6556b376c02a-kube-api-access-xhkvg\") pod \"dnsmasq-dns-5c476d78c5-djb2g\" (UID: \"0cf0fb02-e586-4daa-88c1-6556b376c02a\") " pod="openstack/dnsmasq-dns-5c476d78c5-djb2g" Nov 22 08:14:52 crc kubenswrapper[4789]: I1122 08:14:52.896691 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0cf0fb02-e586-4daa-88c1-6556b376c02a-dns-svc\") pod \"dnsmasq-dns-5c476d78c5-djb2g\" (UID: \"0cf0fb02-e586-4daa-88c1-6556b376c02a\") " pod="openstack/dnsmasq-dns-5c476d78c5-djb2g" Nov 22 08:14:52 crc kubenswrapper[4789]: I1122 08:14:52.896736 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0cf0fb02-e586-4daa-88c1-6556b376c02a-config\") pod \"dnsmasq-dns-5c476d78c5-djb2g\" (UID: \"0cf0fb02-e586-4daa-88c1-6556b376c02a\") " pod="openstack/dnsmasq-dns-5c476d78c5-djb2g" Nov 22 08:14:52 crc kubenswrapper[4789]: I1122 08:14:52.896776 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0cf0fb02-e586-4daa-88c1-6556b376c02a-ovsdbserver-nb\") pod \"dnsmasq-dns-5c476d78c5-djb2g\" (UID: \"0cf0fb02-e586-4daa-88c1-6556b376c02a\") " pod="openstack/dnsmasq-dns-5c476d78c5-djb2g" Nov 22 08:14:52 crc kubenswrapper[4789]: I1122 08:14:52.997424 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0cf0fb02-e586-4daa-88c1-6556b376c02a-config\") pod \"dnsmasq-dns-5c476d78c5-djb2g\" (UID: \"0cf0fb02-e586-4daa-88c1-6556b376c02a\") " pod="openstack/dnsmasq-dns-5c476d78c5-djb2g" Nov 22 08:14:52 crc kubenswrapper[4789]: I1122 08:14:52.997472 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0cf0fb02-e586-4daa-88c1-6556b376c02a-ovsdbserver-nb\") pod \"dnsmasq-dns-5c476d78c5-djb2g\" (UID: \"0cf0fb02-e586-4daa-88c1-6556b376c02a\") " pod="openstack/dnsmasq-dns-5c476d78c5-djb2g" Nov 22 08:14:52 crc kubenswrapper[4789]: I1122 08:14:52.997529 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0cf0fb02-e586-4daa-88c1-6556b376c02a-ovsdbserver-sb\") pod \"dnsmasq-dns-5c476d78c5-djb2g\" (UID: \"0cf0fb02-e586-4daa-88c1-6556b376c02a\") " pod="openstack/dnsmasq-dns-5c476d78c5-djb2g" Nov 22 08:14:52 crc kubenswrapper[4789]: I1122 08:14:52.997558 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xhkvg\" (UniqueName: \"kubernetes.io/projected/0cf0fb02-e586-4daa-88c1-6556b376c02a-kube-api-access-xhkvg\") pod \"dnsmasq-dns-5c476d78c5-djb2g\" (UID: \"0cf0fb02-e586-4daa-88c1-6556b376c02a\") " pod="openstack/dnsmasq-dns-5c476d78c5-djb2g" Nov 22 08:14:52 crc kubenswrapper[4789]: I1122 08:14:52.997629 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0cf0fb02-e586-4daa-88c1-6556b376c02a-dns-svc\") pod \"dnsmasq-dns-5c476d78c5-djb2g\" (UID: \"0cf0fb02-e586-4daa-88c1-6556b376c02a\") " pod="openstack/dnsmasq-dns-5c476d78c5-djb2g" Nov 22 08:14:52 crc kubenswrapper[4789]: I1122 08:14:52.998849 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0cf0fb02-e586-4daa-88c1-6556b376c02a-config\") pod \"dnsmasq-dns-5c476d78c5-djb2g\" (UID: \"0cf0fb02-e586-4daa-88c1-6556b376c02a\") " pod="openstack/dnsmasq-dns-5c476d78c5-djb2g" Nov 22 08:14:52 crc kubenswrapper[4789]: I1122 08:14:52.998864 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0cf0fb02-e586-4daa-88c1-6556b376c02a-ovsdbserver-nb\") pod \"dnsmasq-dns-5c476d78c5-djb2g\" (UID: \"0cf0fb02-e586-4daa-88c1-6556b376c02a\") " pod="openstack/dnsmasq-dns-5c476d78c5-djb2g" Nov 22 08:14:52 crc kubenswrapper[4789]: I1122 08:14:52.998914 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0cf0fb02-e586-4daa-88c1-6556b376c02a-ovsdbserver-sb\") pod \"dnsmasq-dns-5c476d78c5-djb2g\" (UID: \"0cf0fb02-e586-4daa-88c1-6556b376c02a\") " pod="openstack/dnsmasq-dns-5c476d78c5-djb2g" Nov 22 08:14:52 crc kubenswrapper[4789]: I1122 08:14:52.998956 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0cf0fb02-e586-4daa-88c1-6556b376c02a-dns-svc\") pod \"dnsmasq-dns-5c476d78c5-djb2g\" (UID: \"0cf0fb02-e586-4daa-88c1-6556b376c02a\") " pod="openstack/dnsmasq-dns-5c476d78c5-djb2g" Nov 22 08:14:53 crc kubenswrapper[4789]: I1122 08:14:53.019865 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xhkvg\" (UniqueName: \"kubernetes.io/projected/0cf0fb02-e586-4daa-88c1-6556b376c02a-kube-api-access-xhkvg\") pod \"dnsmasq-dns-5c476d78c5-djb2g\" (UID: \"0cf0fb02-e586-4daa-88c1-6556b376c02a\") " pod="openstack/dnsmasq-dns-5c476d78c5-djb2g" Nov 22 08:14:53 crc kubenswrapper[4789]: I1122 08:14:53.062048 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c476d78c5-djb2g" Nov 22 08:14:53 crc kubenswrapper[4789]: I1122 08:14:53.123972 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-65c9b8d4f7-kfsdh"] Nov 22 08:14:53 crc kubenswrapper[4789]: W1122 08:14:53.131930 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2e9862f9_e69a_472f_9db9_638c95ee3912.slice/crio-5c6eecd336e919c7bef4462ee52c51e3602fe0cc8c361f4b47e2e0b7fe6c6f79 WatchSource:0}: Error finding container 5c6eecd336e919c7bef4462ee52c51e3602fe0cc8c361f4b47e2e0b7fe6c6f79: Status 404 returned error can't find the container with id 5c6eecd336e919c7bef4462ee52c51e3602fe0cc8c361f4b47e2e0b7fe6c6f79 Nov 22 08:14:53 crc kubenswrapper[4789]: I1122 08:14:53.180426 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-c6fhr"] Nov 22 08:14:53 crc kubenswrapper[4789]: I1122 08:14:53.483983 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c476d78c5-djb2g"] Nov 22 08:14:53 crc kubenswrapper[4789]: I1122 08:14:53.531059 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Nov 22 08:14:53 crc kubenswrapper[4789]: I1122 08:14:53.531105 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Nov 22 08:14:53 crc kubenswrapper[4789]: I1122 08:14:53.570170 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Nov 22 08:14:54 crc kubenswrapper[4789]: I1122 08:14:54.004640 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-c6fhr" event={"ID":"0e0f4cec-d122-4b44-8344-1cd6e515ef19","Type":"ContainerStarted","Data":"be23df9cafd27d4f3a27d999d43a646f1290997459e6935a566ba4cdb1a42ced"} Nov 22 08:14:54 crc kubenswrapper[4789]: I1122 08:14:54.005010 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-c6fhr" event={"ID":"0e0f4cec-d122-4b44-8344-1cd6e515ef19","Type":"ContainerStarted","Data":"2fd5eb40834e47aaefa236c3ccef9e6085af065823d68feff7f4218854a71a69"} Nov 22 08:14:54 crc kubenswrapper[4789]: I1122 08:14:54.006276 4789 generic.go:334] "Generic (PLEG): container finished" podID="2e9862f9-e69a-472f-9db9-638c95ee3912" containerID="e9e5f2aec8b3d285a6e9e963b44ca3fa3a8b87702afb2dd44b4d4baf6c4a3391" exitCode=0 Nov 22 08:14:54 crc kubenswrapper[4789]: I1122 08:14:54.006349 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65c9b8d4f7-kfsdh" event={"ID":"2e9862f9-e69a-472f-9db9-638c95ee3912","Type":"ContainerDied","Data":"e9e5f2aec8b3d285a6e9e963b44ca3fa3a8b87702afb2dd44b4d4baf6c4a3391"} Nov 22 08:14:54 crc kubenswrapper[4789]: I1122 08:14:54.006414 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65c9b8d4f7-kfsdh" event={"ID":"2e9862f9-e69a-472f-9db9-638c95ee3912","Type":"ContainerStarted","Data":"5c6eecd336e919c7bef4462ee52c51e3602fe0cc8c361f4b47e2e0b7fe6c6f79"} Nov 22 08:14:54 crc kubenswrapper[4789]: I1122 08:14:54.008276 4789 generic.go:334] "Generic (PLEG): container finished" podID="0cf0fb02-e586-4daa-88c1-6556b376c02a" containerID="98e61e9ab6d1be90a45fd2a32392caf89eebaff16982d2f7347fb57dd8948bdb" exitCode=0 Nov 22 08:14:54 crc kubenswrapper[4789]: I1122 08:14:54.008516 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6486446b9f-2w9gc" podUID="66b9aef5-3c42-4449-987c-9087653553b5" containerName="dnsmasq-dns" containerID="cri-o://6b0f95e060c81173508b35aea811dae2268821ef6427dd3647924e9202f650c1" gracePeriod=10 Nov 22 08:14:54 crc kubenswrapper[4789]: I1122 08:14:54.008905 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7c6d9948dc-2dpcz" podUID="e32fe7bc-c0c8-443b-af3e-d9cc58db4f24" containerName="dnsmasq-dns" containerID="cri-o://4c86ea65c880f7ffe8e1357a4462d7ace08f91605d53b35fb585f327d1976286" gracePeriod=10 Nov 22 08:14:54 crc kubenswrapper[4789]: I1122 08:14:54.009010 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c476d78c5-djb2g" event={"ID":"0cf0fb02-e586-4daa-88c1-6556b376c02a","Type":"ContainerDied","Data":"98e61e9ab6d1be90a45fd2a32392caf89eebaff16982d2f7347fb57dd8948bdb"} Nov 22 08:14:54 crc kubenswrapper[4789]: I1122 08:14:54.009037 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c476d78c5-djb2g" event={"ID":"0cf0fb02-e586-4daa-88c1-6556b376c02a","Type":"ContainerStarted","Data":"7721d00be02c91565b524c85dff133a1bb86958e34fd9092e9d73d985f6ce81b"} Nov 22 08:14:54 crc kubenswrapper[4789]: I1122 08:14:54.035095 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-c6fhr" podStartSLOduration=2.035076164 podStartE2EDuration="2.035076164s" podCreationTimestamp="2025-11-22 08:14:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:14:54.027996251 +0000 UTC m=+1108.262396544" watchObservedRunningTime="2025-11-22 08:14:54.035076164 +0000 UTC m=+1108.269476427" Nov 22 08:14:54 crc kubenswrapper[4789]: I1122 08:14:54.071546 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Nov 22 08:14:54 crc kubenswrapper[4789]: I1122 08:14:54.351287 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Nov 22 08:14:54 crc kubenswrapper[4789]: I1122 08:14:54.353368 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Nov 22 08:14:54 crc kubenswrapper[4789]: I1122 08:14:54.356430 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-clrdr" Nov 22 08:14:54 crc kubenswrapper[4789]: I1122 08:14:54.358824 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Nov 22 08:14:54 crc kubenswrapper[4789]: I1122 08:14:54.358868 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Nov 22 08:14:54 crc kubenswrapper[4789]: I1122 08:14:54.359782 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Nov 22 08:14:54 crc kubenswrapper[4789]: I1122 08:14:54.373848 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Nov 22 08:14:54 crc kubenswrapper[4789]: I1122 08:14:54.421067 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/ec49bfda-737c-49c9-8bc3-cfceb08e5f88-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"ec49bfda-737c-49c9-8bc3-cfceb08e5f88\") " pod="openstack/ovn-northd-0" Nov 22 08:14:54 crc kubenswrapper[4789]: I1122 08:14:54.421132 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ec49bfda-737c-49c9-8bc3-cfceb08e5f88-scripts\") pod \"ovn-northd-0\" (UID: \"ec49bfda-737c-49c9-8bc3-cfceb08e5f88\") " pod="openstack/ovn-northd-0" Nov 22 08:14:54 crc kubenswrapper[4789]: I1122 08:14:54.421160 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ec49bfda-737c-49c9-8bc3-cfceb08e5f88-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"ec49bfda-737c-49c9-8bc3-cfceb08e5f88\") " pod="openstack/ovn-northd-0" Nov 22 08:14:54 crc kubenswrapper[4789]: I1122 08:14:54.421187 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec49bfda-737c-49c9-8bc3-cfceb08e5f88-config\") pod \"ovn-northd-0\" (UID: \"ec49bfda-737c-49c9-8bc3-cfceb08e5f88\") " pod="openstack/ovn-northd-0" Nov 22 08:14:54 crc kubenswrapper[4789]: I1122 08:14:54.421220 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ec49bfda-737c-49c9-8bc3-cfceb08e5f88-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"ec49bfda-737c-49c9-8bc3-cfceb08e5f88\") " pod="openstack/ovn-northd-0" Nov 22 08:14:54 crc kubenswrapper[4789]: I1122 08:14:54.421234 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mgwcx\" (UniqueName: \"kubernetes.io/projected/ec49bfda-737c-49c9-8bc3-cfceb08e5f88-kube-api-access-mgwcx\") pod \"ovn-northd-0\" (UID: \"ec49bfda-737c-49c9-8bc3-cfceb08e5f88\") " pod="openstack/ovn-northd-0" Nov 22 08:14:54 crc kubenswrapper[4789]: I1122 08:14:54.421269 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec49bfda-737c-49c9-8bc3-cfceb08e5f88-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"ec49bfda-737c-49c9-8bc3-cfceb08e5f88\") " pod="openstack/ovn-northd-0" Nov 22 08:14:54 crc kubenswrapper[4789]: I1122 08:14:54.465652 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c6d9948dc-2dpcz" Nov 22 08:14:54 crc kubenswrapper[4789]: I1122 08:14:54.489172 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6486446b9f-2w9gc" Nov 22 08:14:54 crc kubenswrapper[4789]: I1122 08:14:54.522393 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e32fe7bc-c0c8-443b-af3e-d9cc58db4f24-dns-svc\") pod \"e32fe7bc-c0c8-443b-af3e-d9cc58db4f24\" (UID: \"e32fe7bc-c0c8-443b-af3e-d9cc58db4f24\") " Nov 22 08:14:54 crc kubenswrapper[4789]: I1122 08:14:54.522706 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-llqn6\" (UniqueName: \"kubernetes.io/projected/e32fe7bc-c0c8-443b-af3e-d9cc58db4f24-kube-api-access-llqn6\") pod \"e32fe7bc-c0c8-443b-af3e-d9cc58db4f24\" (UID: \"e32fe7bc-c0c8-443b-af3e-d9cc58db4f24\") " Nov 22 08:14:54 crc kubenswrapper[4789]: I1122 08:14:54.522739 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e32fe7bc-c0c8-443b-af3e-d9cc58db4f24-config\") pod \"e32fe7bc-c0c8-443b-af3e-d9cc58db4f24\" (UID: \"e32fe7bc-c0c8-443b-af3e-d9cc58db4f24\") " Nov 22 08:14:54 crc kubenswrapper[4789]: I1122 08:14:54.523012 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/ec49bfda-737c-49c9-8bc3-cfceb08e5f88-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"ec49bfda-737c-49c9-8bc3-cfceb08e5f88\") " pod="openstack/ovn-northd-0" Nov 22 08:14:54 crc kubenswrapper[4789]: I1122 08:14:54.523071 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ec49bfda-737c-49c9-8bc3-cfceb08e5f88-scripts\") pod \"ovn-northd-0\" (UID: \"ec49bfda-737c-49c9-8bc3-cfceb08e5f88\") " pod="openstack/ovn-northd-0" Nov 22 08:14:54 crc kubenswrapper[4789]: I1122 08:14:54.523108 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ec49bfda-737c-49c9-8bc3-cfceb08e5f88-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"ec49bfda-737c-49c9-8bc3-cfceb08e5f88\") " pod="openstack/ovn-northd-0" Nov 22 08:14:54 crc kubenswrapper[4789]: I1122 08:14:54.523133 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec49bfda-737c-49c9-8bc3-cfceb08e5f88-config\") pod \"ovn-northd-0\" (UID: \"ec49bfda-737c-49c9-8bc3-cfceb08e5f88\") " pod="openstack/ovn-northd-0" Nov 22 08:14:54 crc kubenswrapper[4789]: I1122 08:14:54.523172 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ec49bfda-737c-49c9-8bc3-cfceb08e5f88-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"ec49bfda-737c-49c9-8bc3-cfceb08e5f88\") " pod="openstack/ovn-northd-0" Nov 22 08:14:54 crc kubenswrapper[4789]: I1122 08:14:54.523187 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mgwcx\" (UniqueName: \"kubernetes.io/projected/ec49bfda-737c-49c9-8bc3-cfceb08e5f88-kube-api-access-mgwcx\") pod \"ovn-northd-0\" (UID: \"ec49bfda-737c-49c9-8bc3-cfceb08e5f88\") " pod="openstack/ovn-northd-0" Nov 22 08:14:54 crc kubenswrapper[4789]: I1122 08:14:54.523230 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec49bfda-737c-49c9-8bc3-cfceb08e5f88-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"ec49bfda-737c-49c9-8bc3-cfceb08e5f88\") " pod="openstack/ovn-northd-0" Nov 22 08:14:54 crc kubenswrapper[4789]: I1122 08:14:54.524722 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec49bfda-737c-49c9-8bc3-cfceb08e5f88-config\") pod \"ovn-northd-0\" (UID: \"ec49bfda-737c-49c9-8bc3-cfceb08e5f88\") " pod="openstack/ovn-northd-0" Nov 22 08:14:54 crc kubenswrapper[4789]: I1122 08:14:54.524773 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ec49bfda-737c-49c9-8bc3-cfceb08e5f88-scripts\") pod \"ovn-northd-0\" (UID: \"ec49bfda-737c-49c9-8bc3-cfceb08e5f88\") " pod="openstack/ovn-northd-0" Nov 22 08:14:54 crc kubenswrapper[4789]: I1122 08:14:54.525420 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ec49bfda-737c-49c9-8bc3-cfceb08e5f88-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"ec49bfda-737c-49c9-8bc3-cfceb08e5f88\") " pod="openstack/ovn-northd-0" Nov 22 08:14:54 crc kubenswrapper[4789]: I1122 08:14:54.526904 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec49bfda-737c-49c9-8bc3-cfceb08e5f88-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"ec49bfda-737c-49c9-8bc3-cfceb08e5f88\") " pod="openstack/ovn-northd-0" Nov 22 08:14:54 crc kubenswrapper[4789]: I1122 08:14:54.529597 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/ec49bfda-737c-49c9-8bc3-cfceb08e5f88-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"ec49bfda-737c-49c9-8bc3-cfceb08e5f88\") " pod="openstack/ovn-northd-0" Nov 22 08:14:54 crc kubenswrapper[4789]: I1122 08:14:54.533048 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ec49bfda-737c-49c9-8bc3-cfceb08e5f88-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"ec49bfda-737c-49c9-8bc3-cfceb08e5f88\") " pod="openstack/ovn-northd-0" Nov 22 08:14:54 crc kubenswrapper[4789]: I1122 08:14:54.533169 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e32fe7bc-c0c8-443b-af3e-d9cc58db4f24-kube-api-access-llqn6" (OuterVolumeSpecName: "kube-api-access-llqn6") pod "e32fe7bc-c0c8-443b-af3e-d9cc58db4f24" (UID: "e32fe7bc-c0c8-443b-af3e-d9cc58db4f24"). InnerVolumeSpecName "kube-api-access-llqn6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:14:54 crc kubenswrapper[4789]: I1122 08:14:54.538830 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mgwcx\" (UniqueName: \"kubernetes.io/projected/ec49bfda-737c-49c9-8bc3-cfceb08e5f88-kube-api-access-mgwcx\") pod \"ovn-northd-0\" (UID: \"ec49bfda-737c-49c9-8bc3-cfceb08e5f88\") " pod="openstack/ovn-northd-0" Nov 22 08:14:54 crc kubenswrapper[4789]: I1122 08:14:54.569318 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e32fe7bc-c0c8-443b-af3e-d9cc58db4f24-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e32fe7bc-c0c8-443b-af3e-d9cc58db4f24" (UID: "e32fe7bc-c0c8-443b-af3e-d9cc58db4f24"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:14:54 crc kubenswrapper[4789]: I1122 08:14:54.582172 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e32fe7bc-c0c8-443b-af3e-d9cc58db4f24-config" (OuterVolumeSpecName: "config") pod "e32fe7bc-c0c8-443b-af3e-d9cc58db4f24" (UID: "e32fe7bc-c0c8-443b-af3e-d9cc58db4f24"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:14:54 crc kubenswrapper[4789]: I1122 08:14:54.624807 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j9nj2\" (UniqueName: \"kubernetes.io/projected/66b9aef5-3c42-4449-987c-9087653553b5-kube-api-access-j9nj2\") pod \"66b9aef5-3c42-4449-987c-9087653553b5\" (UID: \"66b9aef5-3c42-4449-987c-9087653553b5\") " Nov 22 08:14:54 crc kubenswrapper[4789]: I1122 08:14:54.624883 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/66b9aef5-3c42-4449-987c-9087653553b5-dns-svc\") pod \"66b9aef5-3c42-4449-987c-9087653553b5\" (UID: \"66b9aef5-3c42-4449-987c-9087653553b5\") " Nov 22 08:14:54 crc kubenswrapper[4789]: I1122 08:14:54.624961 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/66b9aef5-3c42-4449-987c-9087653553b5-config\") pod \"66b9aef5-3c42-4449-987c-9087653553b5\" (UID: \"66b9aef5-3c42-4449-987c-9087653553b5\") " Nov 22 08:14:54 crc kubenswrapper[4789]: I1122 08:14:54.626306 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-llqn6\" (UniqueName: \"kubernetes.io/projected/e32fe7bc-c0c8-443b-af3e-d9cc58db4f24-kube-api-access-llqn6\") on node \"crc\" DevicePath \"\"" Nov 22 08:14:54 crc kubenswrapper[4789]: I1122 08:14:54.626365 4789 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e32fe7bc-c0c8-443b-af3e-d9cc58db4f24-config\") on node \"crc\" DevicePath \"\"" Nov 22 08:14:54 crc kubenswrapper[4789]: I1122 08:14:54.626379 4789 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e32fe7bc-c0c8-443b-af3e-d9cc58db4f24-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 22 08:14:54 crc kubenswrapper[4789]: I1122 08:14:54.629974 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66b9aef5-3c42-4449-987c-9087653553b5-kube-api-access-j9nj2" (OuterVolumeSpecName: "kube-api-access-j9nj2") pod "66b9aef5-3c42-4449-987c-9087653553b5" (UID: "66b9aef5-3c42-4449-987c-9087653553b5"). InnerVolumeSpecName "kube-api-access-j9nj2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:14:54 crc kubenswrapper[4789]: I1122 08:14:54.671930 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Nov 22 08:14:54 crc kubenswrapper[4789]: I1122 08:14:54.676294 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/66b9aef5-3c42-4449-987c-9087653553b5-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "66b9aef5-3c42-4449-987c-9087653553b5" (UID: "66b9aef5-3c42-4449-987c-9087653553b5"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:14:54 crc kubenswrapper[4789]: I1122 08:14:54.694067 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/66b9aef5-3c42-4449-987c-9087653553b5-config" (OuterVolumeSpecName: "config") pod "66b9aef5-3c42-4449-987c-9087653553b5" (UID: "66b9aef5-3c42-4449-987c-9087653553b5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:14:54 crc kubenswrapper[4789]: I1122 08:14:54.728519 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j9nj2\" (UniqueName: \"kubernetes.io/projected/66b9aef5-3c42-4449-987c-9087653553b5-kube-api-access-j9nj2\") on node \"crc\" DevicePath \"\"" Nov 22 08:14:54 crc kubenswrapper[4789]: I1122 08:14:54.728564 4789 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/66b9aef5-3c42-4449-987c-9087653553b5-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 22 08:14:54 crc kubenswrapper[4789]: I1122 08:14:54.728579 4789 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/66b9aef5-3c42-4449-987c-9087653553b5-config\") on node \"crc\" DevicePath \"\"" Nov 22 08:14:55 crc kubenswrapper[4789]: I1122 08:14:55.018228 4789 generic.go:334] "Generic (PLEG): container finished" podID="e32fe7bc-c0c8-443b-af3e-d9cc58db4f24" containerID="4c86ea65c880f7ffe8e1357a4462d7ace08f91605d53b35fb585f327d1976286" exitCode=0 Nov 22 08:14:55 crc kubenswrapper[4789]: I1122 08:14:55.018287 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c6d9948dc-2dpcz" Nov 22 08:14:55 crc kubenswrapper[4789]: I1122 08:14:55.018314 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c6d9948dc-2dpcz" event={"ID":"e32fe7bc-c0c8-443b-af3e-d9cc58db4f24","Type":"ContainerDied","Data":"4c86ea65c880f7ffe8e1357a4462d7ace08f91605d53b35fb585f327d1976286"} Nov 22 08:14:55 crc kubenswrapper[4789]: I1122 08:14:55.018730 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c6d9948dc-2dpcz" event={"ID":"e32fe7bc-c0c8-443b-af3e-d9cc58db4f24","Type":"ContainerDied","Data":"ac1c4ffae0e5f9a9e0a9e26d05d3f8afe5744c521f78f2e52121108a10794560"} Nov 22 08:14:55 crc kubenswrapper[4789]: I1122 08:14:55.018780 4789 scope.go:117] "RemoveContainer" containerID="4c86ea65c880f7ffe8e1357a4462d7ace08f91605d53b35fb585f327d1976286" Nov 22 08:14:55 crc kubenswrapper[4789]: I1122 08:14:55.022098 4789 generic.go:334] "Generic (PLEG): container finished" podID="66b9aef5-3c42-4449-987c-9087653553b5" containerID="6b0f95e060c81173508b35aea811dae2268821ef6427dd3647924e9202f650c1" exitCode=0 Nov 22 08:14:55 crc kubenswrapper[4789]: I1122 08:14:55.022209 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6486446b9f-2w9gc" event={"ID":"66b9aef5-3c42-4449-987c-9087653553b5","Type":"ContainerDied","Data":"6b0f95e060c81173508b35aea811dae2268821ef6427dd3647924e9202f650c1"} Nov 22 08:14:55 crc kubenswrapper[4789]: I1122 08:14:55.022233 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6486446b9f-2w9gc" event={"ID":"66b9aef5-3c42-4449-987c-9087653553b5","Type":"ContainerDied","Data":"bc250ff924e10b9073d12fc2a8f12d578726b089a226bc9781f93b8fdb6bbf00"} Nov 22 08:14:55 crc kubenswrapper[4789]: I1122 08:14:55.022287 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6486446b9f-2w9gc" Nov 22 08:14:55 crc kubenswrapper[4789]: I1122 08:14:55.024413 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c476d78c5-djb2g" event={"ID":"0cf0fb02-e586-4daa-88c1-6556b376c02a","Type":"ContainerStarted","Data":"3bba78e22b91a98d5a6fd8c4d50cedda778719b31f2900449fafb4e224b7a2bb"} Nov 22 08:14:55 crc kubenswrapper[4789]: I1122 08:14:55.024521 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c476d78c5-djb2g" Nov 22 08:14:55 crc kubenswrapper[4789]: I1122 08:14:55.028214 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65c9b8d4f7-kfsdh" event={"ID":"2e9862f9-e69a-472f-9db9-638c95ee3912","Type":"ContainerStarted","Data":"449fe80a53cf4a823fa98663c8779562ef32de7a57fd4ceac3350994ab0ee9d8"} Nov 22 08:14:55 crc kubenswrapper[4789]: I1122 08:14:55.028253 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-65c9b8d4f7-kfsdh" Nov 22 08:14:55 crc kubenswrapper[4789]: I1122 08:14:55.039803 4789 scope.go:117] "RemoveContainer" containerID="aa072d21743fd61b576d25e52ac024348caaf4f7d8a62f95f279868643113ed8" Nov 22 08:14:55 crc kubenswrapper[4789]: I1122 08:14:55.049617 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5c476d78c5-djb2g" podStartSLOduration=3.049599324 podStartE2EDuration="3.049599324s" podCreationTimestamp="2025-11-22 08:14:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:14:55.046301803 +0000 UTC m=+1109.280702086" watchObservedRunningTime="2025-11-22 08:14:55.049599324 +0000 UTC m=+1109.283999597" Nov 22 08:14:55 crc kubenswrapper[4789]: I1122 08:14:55.067443 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-65c9b8d4f7-kfsdh" podStartSLOduration=3.067424129 podStartE2EDuration="3.067424129s" podCreationTimestamp="2025-11-22 08:14:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:14:55.063208824 +0000 UTC m=+1109.297609117" watchObservedRunningTime="2025-11-22 08:14:55.067424129 +0000 UTC m=+1109.301824402" Nov 22 08:14:55 crc kubenswrapper[4789]: I1122 08:14:55.078766 4789 scope.go:117] "RemoveContainer" containerID="4c86ea65c880f7ffe8e1357a4462d7ace08f91605d53b35fb585f327d1976286" Nov 22 08:14:55 crc kubenswrapper[4789]: E1122 08:14:55.079303 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4c86ea65c880f7ffe8e1357a4462d7ace08f91605d53b35fb585f327d1976286\": container with ID starting with 4c86ea65c880f7ffe8e1357a4462d7ace08f91605d53b35fb585f327d1976286 not found: ID does not exist" containerID="4c86ea65c880f7ffe8e1357a4462d7ace08f91605d53b35fb585f327d1976286" Nov 22 08:14:55 crc kubenswrapper[4789]: I1122 08:14:55.079429 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4c86ea65c880f7ffe8e1357a4462d7ace08f91605d53b35fb585f327d1976286"} err="failed to get container status \"4c86ea65c880f7ffe8e1357a4462d7ace08f91605d53b35fb585f327d1976286\": rpc error: code = NotFound desc = could not find container \"4c86ea65c880f7ffe8e1357a4462d7ace08f91605d53b35fb585f327d1976286\": container with ID starting with 4c86ea65c880f7ffe8e1357a4462d7ace08f91605d53b35fb585f327d1976286 not found: ID does not exist" Nov 22 08:14:55 crc kubenswrapper[4789]: I1122 08:14:55.079533 4789 scope.go:117] "RemoveContainer" containerID="aa072d21743fd61b576d25e52ac024348caaf4f7d8a62f95f279868643113ed8" Nov 22 08:14:55 crc kubenswrapper[4789]: E1122 08:14:55.080140 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aa072d21743fd61b576d25e52ac024348caaf4f7d8a62f95f279868643113ed8\": container with ID starting with aa072d21743fd61b576d25e52ac024348caaf4f7d8a62f95f279868643113ed8 not found: ID does not exist" containerID="aa072d21743fd61b576d25e52ac024348caaf4f7d8a62f95f279868643113ed8" Nov 22 08:14:55 crc kubenswrapper[4789]: I1122 08:14:55.080278 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa072d21743fd61b576d25e52ac024348caaf4f7d8a62f95f279868643113ed8"} err="failed to get container status \"aa072d21743fd61b576d25e52ac024348caaf4f7d8a62f95f279868643113ed8\": rpc error: code = NotFound desc = could not find container \"aa072d21743fd61b576d25e52ac024348caaf4f7d8a62f95f279868643113ed8\": container with ID starting with aa072d21743fd61b576d25e52ac024348caaf4f7d8a62f95f279868643113ed8 not found: ID does not exist" Nov 22 08:14:55 crc kubenswrapper[4789]: I1122 08:14:55.080306 4789 scope.go:117] "RemoveContainer" containerID="6b0f95e060c81173508b35aea811dae2268821ef6427dd3647924e9202f650c1" Nov 22 08:14:55 crc kubenswrapper[4789]: I1122 08:14:55.087833 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6486446b9f-2w9gc"] Nov 22 08:14:55 crc kubenswrapper[4789]: I1122 08:14:55.093794 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6486446b9f-2w9gc"] Nov 22 08:14:55 crc kubenswrapper[4789]: I1122 08:14:55.100766 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7c6d9948dc-2dpcz"] Nov 22 08:14:55 crc kubenswrapper[4789]: I1122 08:14:55.113339 4789 scope.go:117] "RemoveContainer" containerID="670465358e9760832a3d128dd0dd08cd1e21b7cd4bea04525a642eac7ddda14d" Nov 22 08:14:55 crc kubenswrapper[4789]: I1122 08:14:55.132540 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7c6d9948dc-2dpcz"] Nov 22 08:14:55 crc kubenswrapper[4789]: I1122 08:14:55.143057 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Nov 22 08:14:55 crc kubenswrapper[4789]: I1122 08:14:55.158804 4789 scope.go:117] "RemoveContainer" containerID="6b0f95e060c81173508b35aea811dae2268821ef6427dd3647924e9202f650c1" Nov 22 08:14:55 crc kubenswrapper[4789]: E1122 08:14:55.159240 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b0f95e060c81173508b35aea811dae2268821ef6427dd3647924e9202f650c1\": container with ID starting with 6b0f95e060c81173508b35aea811dae2268821ef6427dd3647924e9202f650c1 not found: ID does not exist" containerID="6b0f95e060c81173508b35aea811dae2268821ef6427dd3647924e9202f650c1" Nov 22 08:14:55 crc kubenswrapper[4789]: I1122 08:14:55.159297 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b0f95e060c81173508b35aea811dae2268821ef6427dd3647924e9202f650c1"} err="failed to get container status \"6b0f95e060c81173508b35aea811dae2268821ef6427dd3647924e9202f650c1\": rpc error: code = NotFound desc = could not find container \"6b0f95e060c81173508b35aea811dae2268821ef6427dd3647924e9202f650c1\": container with ID starting with 6b0f95e060c81173508b35aea811dae2268821ef6427dd3647924e9202f650c1 not found: ID does not exist" Nov 22 08:14:55 crc kubenswrapper[4789]: I1122 08:14:55.159330 4789 scope.go:117] "RemoveContainer" containerID="670465358e9760832a3d128dd0dd08cd1e21b7cd4bea04525a642eac7ddda14d" Nov 22 08:14:55 crc kubenswrapper[4789]: E1122 08:14:55.159842 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"670465358e9760832a3d128dd0dd08cd1e21b7cd4bea04525a642eac7ddda14d\": container with ID starting with 670465358e9760832a3d128dd0dd08cd1e21b7cd4bea04525a642eac7ddda14d not found: ID does not exist" containerID="670465358e9760832a3d128dd0dd08cd1e21b7cd4bea04525a642eac7ddda14d" Nov 22 08:14:55 crc kubenswrapper[4789]: I1122 08:14:55.159875 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"670465358e9760832a3d128dd0dd08cd1e21b7cd4bea04525a642eac7ddda14d"} err="failed to get container status \"670465358e9760832a3d128dd0dd08cd1e21b7cd4bea04525a642eac7ddda14d\": rpc error: code = NotFound desc = could not find container \"670465358e9760832a3d128dd0dd08cd1e21b7cd4bea04525a642eac7ddda14d\": container with ID starting with 670465358e9760832a3d128dd0dd08cd1e21b7cd4bea04525a642eac7ddda14d not found: ID does not exist" Nov 22 08:14:55 crc kubenswrapper[4789]: I1122 08:14:55.975911 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="66b9aef5-3c42-4449-987c-9087653553b5" path="/var/lib/kubelet/pods/66b9aef5-3c42-4449-987c-9087653553b5/volumes" Nov 22 08:14:55 crc kubenswrapper[4789]: I1122 08:14:55.976483 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e32fe7bc-c0c8-443b-af3e-d9cc58db4f24" path="/var/lib/kubelet/pods/e32fe7bc-c0c8-443b-af3e-d9cc58db4f24/volumes" Nov 22 08:14:56 crc kubenswrapper[4789]: I1122 08:14:56.037309 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"ec49bfda-737c-49c9-8bc3-cfceb08e5f88","Type":"ContainerStarted","Data":"566fc2a6b9270ccb9e2dc5f2894e971326907cf27b61b5574dea406d07640f44"} Nov 22 08:14:56 crc kubenswrapper[4789]: I1122 08:14:56.307419 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Nov 22 08:14:56 crc kubenswrapper[4789]: I1122 08:14:56.307552 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Nov 22 08:14:56 crc kubenswrapper[4789]: I1122 08:14:56.369488 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Nov 22 08:14:57 crc kubenswrapper[4789]: I1122 08:14:57.047722 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"ec49bfda-737c-49c9-8bc3-cfceb08e5f88","Type":"ContainerStarted","Data":"62e620eba821b42c74f8626a0ef3802bd76163463291252312f6fa74fa4fe3b2"} Nov 22 08:14:57 crc kubenswrapper[4789]: I1122 08:14:57.048142 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"ec49bfda-737c-49c9-8bc3-cfceb08e5f88","Type":"ContainerStarted","Data":"50f4343420331094f7487eb78c1961ca353aa2e89fa7ee2ceebc52644bd7b6ad"} Nov 22 08:14:57 crc kubenswrapper[4789]: I1122 08:14:57.065139 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.206203563 podStartE2EDuration="3.065118064s" podCreationTimestamp="2025-11-22 08:14:54 +0000 UTC" firstStartedPulling="2025-11-22 08:14:55.12768174 +0000 UTC m=+1109.362082013" lastFinishedPulling="2025-11-22 08:14:55.986596241 +0000 UTC m=+1110.220996514" observedRunningTime="2025-11-22 08:14:57.063741306 +0000 UTC m=+1111.298141579" watchObservedRunningTime="2025-11-22 08:14:57.065118064 +0000 UTC m=+1111.299518337" Nov 22 08:14:57 crc kubenswrapper[4789]: I1122 08:14:57.109663 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Nov 22 08:14:57 crc kubenswrapper[4789]: I1122 08:14:57.697451 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Nov 22 08:14:57 crc kubenswrapper[4789]: I1122 08:14:57.697517 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Nov 22 08:14:57 crc kubenswrapper[4789]: I1122 08:14:57.700262 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-fdc5-account-create-lqqqs"] Nov 22 08:14:57 crc kubenswrapper[4789]: E1122 08:14:57.700587 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e32fe7bc-c0c8-443b-af3e-d9cc58db4f24" containerName="dnsmasq-dns" Nov 22 08:14:57 crc kubenswrapper[4789]: I1122 08:14:57.700619 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="e32fe7bc-c0c8-443b-af3e-d9cc58db4f24" containerName="dnsmasq-dns" Nov 22 08:14:57 crc kubenswrapper[4789]: E1122 08:14:57.700639 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66b9aef5-3c42-4449-987c-9087653553b5" containerName="init" Nov 22 08:14:57 crc kubenswrapper[4789]: I1122 08:14:57.700645 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="66b9aef5-3c42-4449-987c-9087653553b5" containerName="init" Nov 22 08:14:57 crc kubenswrapper[4789]: E1122 08:14:57.700661 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66b9aef5-3c42-4449-987c-9087653553b5" containerName="dnsmasq-dns" Nov 22 08:14:57 crc kubenswrapper[4789]: I1122 08:14:57.700668 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="66b9aef5-3c42-4449-987c-9087653553b5" containerName="dnsmasq-dns" Nov 22 08:14:57 crc kubenswrapper[4789]: E1122 08:14:57.700679 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e32fe7bc-c0c8-443b-af3e-d9cc58db4f24" containerName="init" Nov 22 08:14:57 crc kubenswrapper[4789]: I1122 08:14:57.700685 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="e32fe7bc-c0c8-443b-af3e-d9cc58db4f24" containerName="init" Nov 22 08:14:57 crc kubenswrapper[4789]: I1122 08:14:57.700857 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="e32fe7bc-c0c8-443b-af3e-d9cc58db4f24" containerName="dnsmasq-dns" Nov 22 08:14:57 crc kubenswrapper[4789]: I1122 08:14:57.700871 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="66b9aef5-3c42-4449-987c-9087653553b5" containerName="dnsmasq-dns" Nov 22 08:14:57 crc kubenswrapper[4789]: I1122 08:14:57.701371 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-fdc5-account-create-lqqqs" Nov 22 08:14:57 crc kubenswrapper[4789]: I1122 08:14:57.703518 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Nov 22 08:14:57 crc kubenswrapper[4789]: I1122 08:14:57.713380 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-fdc5-account-create-lqqqs"] Nov 22 08:14:57 crc kubenswrapper[4789]: I1122 08:14:57.755867 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-2sgdb"] Nov 22 08:14:57 crc kubenswrapper[4789]: I1122 08:14:57.757243 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-2sgdb" Nov 22 08:14:57 crc kubenswrapper[4789]: I1122 08:14:57.763272 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-2sgdb"] Nov 22 08:14:57 crc kubenswrapper[4789]: I1122 08:14:57.786713 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c6mc6\" (UniqueName: \"kubernetes.io/projected/033d4c4c-73bc-4aa4-a124-aff465a011d3-kube-api-access-c6mc6\") pod \"keystone-fdc5-account-create-lqqqs\" (UID: \"033d4c4c-73bc-4aa4-a124-aff465a011d3\") " pod="openstack/keystone-fdc5-account-create-lqqqs" Nov 22 08:14:57 crc kubenswrapper[4789]: I1122 08:14:57.787048 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/033d4c4c-73bc-4aa4-a124-aff465a011d3-operator-scripts\") pod \"keystone-fdc5-account-create-lqqqs\" (UID: \"033d4c4c-73bc-4aa4-a124-aff465a011d3\") " pod="openstack/keystone-fdc5-account-create-lqqqs" Nov 22 08:14:57 crc kubenswrapper[4789]: I1122 08:14:57.791805 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Nov 22 08:14:57 crc kubenswrapper[4789]: I1122 08:14:57.888058 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/033d4c4c-73bc-4aa4-a124-aff465a011d3-operator-scripts\") pod \"keystone-fdc5-account-create-lqqqs\" (UID: \"033d4c4c-73bc-4aa4-a124-aff465a011d3\") " pod="openstack/keystone-fdc5-account-create-lqqqs" Nov 22 08:14:57 crc kubenswrapper[4789]: I1122 08:14:57.888145 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qnhdz\" (UniqueName: \"kubernetes.io/projected/a5b8f813-91e7-4b56-803d-44167089fdf3-kube-api-access-qnhdz\") pod \"keystone-db-create-2sgdb\" (UID: \"a5b8f813-91e7-4b56-803d-44167089fdf3\") " pod="openstack/keystone-db-create-2sgdb" Nov 22 08:14:57 crc kubenswrapper[4789]: I1122 08:14:57.888182 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c6mc6\" (UniqueName: \"kubernetes.io/projected/033d4c4c-73bc-4aa4-a124-aff465a011d3-kube-api-access-c6mc6\") pod \"keystone-fdc5-account-create-lqqqs\" (UID: \"033d4c4c-73bc-4aa4-a124-aff465a011d3\") " pod="openstack/keystone-fdc5-account-create-lqqqs" Nov 22 08:14:57 crc kubenswrapper[4789]: I1122 08:14:57.888277 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a5b8f813-91e7-4b56-803d-44167089fdf3-operator-scripts\") pod \"keystone-db-create-2sgdb\" (UID: \"a5b8f813-91e7-4b56-803d-44167089fdf3\") " pod="openstack/keystone-db-create-2sgdb" Nov 22 08:14:57 crc kubenswrapper[4789]: I1122 08:14:57.889114 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/033d4c4c-73bc-4aa4-a124-aff465a011d3-operator-scripts\") pod \"keystone-fdc5-account-create-lqqqs\" (UID: \"033d4c4c-73bc-4aa4-a124-aff465a011d3\") " pod="openstack/keystone-fdc5-account-create-lqqqs" Nov 22 08:14:57 crc kubenswrapper[4789]: I1122 08:14:57.906545 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c6mc6\" (UniqueName: \"kubernetes.io/projected/033d4c4c-73bc-4aa4-a124-aff465a011d3-kube-api-access-c6mc6\") pod \"keystone-fdc5-account-create-lqqqs\" (UID: \"033d4c4c-73bc-4aa4-a124-aff465a011d3\") " pod="openstack/keystone-fdc5-account-create-lqqqs" Nov 22 08:14:57 crc kubenswrapper[4789]: I1122 08:14:57.953627 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-b867h"] Nov 22 08:14:57 crc kubenswrapper[4789]: I1122 08:14:57.954546 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-b867h" Nov 22 08:14:57 crc kubenswrapper[4789]: I1122 08:14:57.978412 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-b867h"] Nov 22 08:14:57 crc kubenswrapper[4789]: I1122 08:14:57.995664 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a5b8f813-91e7-4b56-803d-44167089fdf3-operator-scripts\") pod \"keystone-db-create-2sgdb\" (UID: \"a5b8f813-91e7-4b56-803d-44167089fdf3\") " pod="openstack/keystone-db-create-2sgdb" Nov 22 08:14:57 crc kubenswrapper[4789]: I1122 08:14:57.996085 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dktxh\" (UniqueName: \"kubernetes.io/projected/95dc84f1-08e8-4e12-8181-4a52830326c1-kube-api-access-dktxh\") pod \"placement-db-create-b867h\" (UID: \"95dc84f1-08e8-4e12-8181-4a52830326c1\") " pod="openstack/placement-db-create-b867h" Nov 22 08:14:57 crc kubenswrapper[4789]: I1122 08:14:57.996205 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qnhdz\" (UniqueName: \"kubernetes.io/projected/a5b8f813-91e7-4b56-803d-44167089fdf3-kube-api-access-qnhdz\") pod \"keystone-db-create-2sgdb\" (UID: \"a5b8f813-91e7-4b56-803d-44167089fdf3\") " pod="openstack/keystone-db-create-2sgdb" Nov 22 08:14:57 crc kubenswrapper[4789]: I1122 08:14:57.996247 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/95dc84f1-08e8-4e12-8181-4a52830326c1-operator-scripts\") pod \"placement-db-create-b867h\" (UID: \"95dc84f1-08e8-4e12-8181-4a52830326c1\") " pod="openstack/placement-db-create-b867h" Nov 22 08:14:57 crc kubenswrapper[4789]: I1122 08:14:57.996957 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a5b8f813-91e7-4b56-803d-44167089fdf3-operator-scripts\") pod \"keystone-db-create-2sgdb\" (UID: \"a5b8f813-91e7-4b56-803d-44167089fdf3\") " pod="openstack/keystone-db-create-2sgdb" Nov 22 08:14:58 crc kubenswrapper[4789]: I1122 08:14:58.021973 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-fdc5-account-create-lqqqs" Nov 22 08:14:58 crc kubenswrapper[4789]: I1122 08:14:58.022527 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qnhdz\" (UniqueName: \"kubernetes.io/projected/a5b8f813-91e7-4b56-803d-44167089fdf3-kube-api-access-qnhdz\") pod \"keystone-db-create-2sgdb\" (UID: \"a5b8f813-91e7-4b56-803d-44167089fdf3\") " pod="openstack/keystone-db-create-2sgdb" Nov 22 08:14:58 crc kubenswrapper[4789]: I1122 08:14:58.034266 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-d5fc-account-create-tbdhr"] Nov 22 08:14:58 crc kubenswrapper[4789]: I1122 08:14:58.035398 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-d5fc-account-create-tbdhr" Nov 22 08:14:58 crc kubenswrapper[4789]: I1122 08:14:58.037764 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Nov 22 08:14:58 crc kubenswrapper[4789]: I1122 08:14:58.054375 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Nov 22 08:14:58 crc kubenswrapper[4789]: I1122 08:14:58.061301 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-d5fc-account-create-tbdhr"] Nov 22 08:14:58 crc kubenswrapper[4789]: I1122 08:14:58.080245 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-2sgdb" Nov 22 08:14:58 crc kubenswrapper[4789]: I1122 08:14:58.099257 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/28b7b6e7-2620-4b4b-b3ab-2da093176b81-operator-scripts\") pod \"placement-d5fc-account-create-tbdhr\" (UID: \"28b7b6e7-2620-4b4b-b3ab-2da093176b81\") " pod="openstack/placement-d5fc-account-create-tbdhr" Nov 22 08:14:58 crc kubenswrapper[4789]: I1122 08:14:58.099302 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dktxh\" (UniqueName: \"kubernetes.io/projected/95dc84f1-08e8-4e12-8181-4a52830326c1-kube-api-access-dktxh\") pod \"placement-db-create-b867h\" (UID: \"95dc84f1-08e8-4e12-8181-4a52830326c1\") " pod="openstack/placement-db-create-b867h" Nov 22 08:14:58 crc kubenswrapper[4789]: I1122 08:14:58.099543 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fjnff\" (UniqueName: \"kubernetes.io/projected/28b7b6e7-2620-4b4b-b3ab-2da093176b81-kube-api-access-fjnff\") pod \"placement-d5fc-account-create-tbdhr\" (UID: \"28b7b6e7-2620-4b4b-b3ab-2da093176b81\") " pod="openstack/placement-d5fc-account-create-tbdhr" Nov 22 08:14:58 crc kubenswrapper[4789]: I1122 08:14:58.099626 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/95dc84f1-08e8-4e12-8181-4a52830326c1-operator-scripts\") pod \"placement-db-create-b867h\" (UID: \"95dc84f1-08e8-4e12-8181-4a52830326c1\") " pod="openstack/placement-db-create-b867h" Nov 22 08:14:58 crc kubenswrapper[4789]: I1122 08:14:58.101258 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/95dc84f1-08e8-4e12-8181-4a52830326c1-operator-scripts\") pod \"placement-db-create-b867h\" (UID: \"95dc84f1-08e8-4e12-8181-4a52830326c1\") " pod="openstack/placement-db-create-b867h" Nov 22 08:14:58 crc kubenswrapper[4789]: I1122 08:14:58.138138 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dktxh\" (UniqueName: \"kubernetes.io/projected/95dc84f1-08e8-4e12-8181-4a52830326c1-kube-api-access-dktxh\") pod \"placement-db-create-b867h\" (UID: \"95dc84f1-08e8-4e12-8181-4a52830326c1\") " pod="openstack/placement-db-create-b867h" Nov 22 08:14:58 crc kubenswrapper[4789]: I1122 08:14:58.161625 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Nov 22 08:14:58 crc kubenswrapper[4789]: I1122 08:14:58.201361 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/28b7b6e7-2620-4b4b-b3ab-2da093176b81-operator-scripts\") pod \"placement-d5fc-account-create-tbdhr\" (UID: \"28b7b6e7-2620-4b4b-b3ab-2da093176b81\") " pod="openstack/placement-d5fc-account-create-tbdhr" Nov 22 08:14:58 crc kubenswrapper[4789]: I1122 08:14:58.201772 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fjnff\" (UniqueName: \"kubernetes.io/projected/28b7b6e7-2620-4b4b-b3ab-2da093176b81-kube-api-access-fjnff\") pod \"placement-d5fc-account-create-tbdhr\" (UID: \"28b7b6e7-2620-4b4b-b3ab-2da093176b81\") " pod="openstack/placement-d5fc-account-create-tbdhr" Nov 22 08:14:58 crc kubenswrapper[4789]: I1122 08:14:58.202879 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/28b7b6e7-2620-4b4b-b3ab-2da093176b81-operator-scripts\") pod \"placement-d5fc-account-create-tbdhr\" (UID: \"28b7b6e7-2620-4b4b-b3ab-2da093176b81\") " pod="openstack/placement-d5fc-account-create-tbdhr" Nov 22 08:14:58 crc kubenswrapper[4789]: I1122 08:14:58.233457 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fjnff\" (UniqueName: \"kubernetes.io/projected/28b7b6e7-2620-4b4b-b3ab-2da093176b81-kube-api-access-fjnff\") pod \"placement-d5fc-account-create-tbdhr\" (UID: \"28b7b6e7-2620-4b4b-b3ab-2da093176b81\") " pod="openstack/placement-d5fc-account-create-tbdhr" Nov 22 08:14:58 crc kubenswrapper[4789]: I1122 08:14:58.293561 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-b867h" Nov 22 08:14:58 crc kubenswrapper[4789]: I1122 08:14:58.359440 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-9bv8q"] Nov 22 08:14:58 crc kubenswrapper[4789]: I1122 08:14:58.360522 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-9bv8q" Nov 22 08:14:58 crc kubenswrapper[4789]: I1122 08:14:58.365830 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-e4c3-account-create-679fs"] Nov 22 08:14:58 crc kubenswrapper[4789]: I1122 08:14:58.368482 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-e4c3-account-create-679fs" Nov 22 08:14:58 crc kubenswrapper[4789]: I1122 08:14:58.370887 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Nov 22 08:14:58 crc kubenswrapper[4789]: I1122 08:14:58.372236 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-9bv8q"] Nov 22 08:14:58 crc kubenswrapper[4789]: I1122 08:14:58.385352 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-e4c3-account-create-679fs"] Nov 22 08:14:58 crc kubenswrapper[4789]: I1122 08:14:58.403837 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r4kjl\" (UniqueName: \"kubernetes.io/projected/dab42b3a-5fd5-43cf-801d-eeb4eef0ac98-kube-api-access-r4kjl\") pod \"glance-db-create-9bv8q\" (UID: \"dab42b3a-5fd5-43cf-801d-eeb4eef0ac98\") " pod="openstack/glance-db-create-9bv8q" Nov 22 08:14:58 crc kubenswrapper[4789]: I1122 08:14:58.403919 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/56293959-d97b-4df2-948e-d8422250945f-operator-scripts\") pod \"glance-e4c3-account-create-679fs\" (UID: \"56293959-d97b-4df2-948e-d8422250945f\") " pod="openstack/glance-e4c3-account-create-679fs" Nov 22 08:14:58 crc kubenswrapper[4789]: I1122 08:14:58.403967 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dab42b3a-5fd5-43cf-801d-eeb4eef0ac98-operator-scripts\") pod \"glance-db-create-9bv8q\" (UID: \"dab42b3a-5fd5-43cf-801d-eeb4eef0ac98\") " pod="openstack/glance-db-create-9bv8q" Nov 22 08:14:58 crc kubenswrapper[4789]: I1122 08:14:58.404000 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xdh8j\" (UniqueName: \"kubernetes.io/projected/56293959-d97b-4df2-948e-d8422250945f-kube-api-access-xdh8j\") pod \"glance-e4c3-account-create-679fs\" (UID: \"56293959-d97b-4df2-948e-d8422250945f\") " pod="openstack/glance-e4c3-account-create-679fs" Nov 22 08:14:58 crc kubenswrapper[4789]: I1122 08:14:58.484230 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-d5fc-account-create-tbdhr" Nov 22 08:14:58 crc kubenswrapper[4789]: I1122 08:14:58.505384 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/56293959-d97b-4df2-948e-d8422250945f-operator-scripts\") pod \"glance-e4c3-account-create-679fs\" (UID: \"56293959-d97b-4df2-948e-d8422250945f\") " pod="openstack/glance-e4c3-account-create-679fs" Nov 22 08:14:58 crc kubenswrapper[4789]: I1122 08:14:58.505452 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dab42b3a-5fd5-43cf-801d-eeb4eef0ac98-operator-scripts\") pod \"glance-db-create-9bv8q\" (UID: \"dab42b3a-5fd5-43cf-801d-eeb4eef0ac98\") " pod="openstack/glance-db-create-9bv8q" Nov 22 08:14:58 crc kubenswrapper[4789]: I1122 08:14:58.505474 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xdh8j\" (UniqueName: \"kubernetes.io/projected/56293959-d97b-4df2-948e-d8422250945f-kube-api-access-xdh8j\") pod \"glance-e4c3-account-create-679fs\" (UID: \"56293959-d97b-4df2-948e-d8422250945f\") " pod="openstack/glance-e4c3-account-create-679fs" Nov 22 08:14:58 crc kubenswrapper[4789]: I1122 08:14:58.505549 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r4kjl\" (UniqueName: \"kubernetes.io/projected/dab42b3a-5fd5-43cf-801d-eeb4eef0ac98-kube-api-access-r4kjl\") pod \"glance-db-create-9bv8q\" (UID: \"dab42b3a-5fd5-43cf-801d-eeb4eef0ac98\") " pod="openstack/glance-db-create-9bv8q" Nov 22 08:14:58 crc kubenswrapper[4789]: I1122 08:14:58.506246 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dab42b3a-5fd5-43cf-801d-eeb4eef0ac98-operator-scripts\") pod \"glance-db-create-9bv8q\" (UID: \"dab42b3a-5fd5-43cf-801d-eeb4eef0ac98\") " pod="openstack/glance-db-create-9bv8q" Nov 22 08:14:58 crc kubenswrapper[4789]: I1122 08:14:58.506855 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/56293959-d97b-4df2-948e-d8422250945f-operator-scripts\") pod \"glance-e4c3-account-create-679fs\" (UID: \"56293959-d97b-4df2-948e-d8422250945f\") " pod="openstack/glance-e4c3-account-create-679fs" Nov 22 08:14:58 crc kubenswrapper[4789]: I1122 08:14:58.509606 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-fdc5-account-create-lqqqs"] Nov 22 08:14:58 crc kubenswrapper[4789]: W1122 08:14:58.509987 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod033d4c4c_73bc_4aa4_a124_aff465a011d3.slice/crio-7f2024cfc098b195b1cc722b50c6cf67491c011de84c07def82cf3450d37ca8c WatchSource:0}: Error finding container 7f2024cfc098b195b1cc722b50c6cf67491c011de84c07def82cf3450d37ca8c: Status 404 returned error can't find the container with id 7f2024cfc098b195b1cc722b50c6cf67491c011de84c07def82cf3450d37ca8c Nov 22 08:14:58 crc kubenswrapper[4789]: I1122 08:14:58.531161 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xdh8j\" (UniqueName: \"kubernetes.io/projected/56293959-d97b-4df2-948e-d8422250945f-kube-api-access-xdh8j\") pod \"glance-e4c3-account-create-679fs\" (UID: \"56293959-d97b-4df2-948e-d8422250945f\") " pod="openstack/glance-e4c3-account-create-679fs" Nov 22 08:14:58 crc kubenswrapper[4789]: I1122 08:14:58.536177 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r4kjl\" (UniqueName: \"kubernetes.io/projected/dab42b3a-5fd5-43cf-801d-eeb4eef0ac98-kube-api-access-r4kjl\") pod \"glance-db-create-9bv8q\" (UID: \"dab42b3a-5fd5-43cf-801d-eeb4eef0ac98\") " pod="openstack/glance-db-create-9bv8q" Nov 22 08:14:58 crc kubenswrapper[4789]: I1122 08:14:58.622241 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-2sgdb"] Nov 22 08:14:58 crc kubenswrapper[4789]: W1122 08:14:58.639027 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda5b8f813_91e7_4b56_803d_44167089fdf3.slice/crio-a430218101b211ab32ab59169fe5c4c1f7e33f518397677eb50e9d51f70b2953 WatchSource:0}: Error finding container a430218101b211ab32ab59169fe5c4c1f7e33f518397677eb50e9d51f70b2953: Status 404 returned error can't find the container with id a430218101b211ab32ab59169fe5c4c1f7e33f518397677eb50e9d51f70b2953 Nov 22 08:14:58 crc kubenswrapper[4789]: I1122 08:14:58.680670 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-9bv8q" Nov 22 08:14:58 crc kubenswrapper[4789]: I1122 08:14:58.689224 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-e4c3-account-create-679fs" Nov 22 08:14:58 crc kubenswrapper[4789]: I1122 08:14:58.784635 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-b867h"] Nov 22 08:14:58 crc kubenswrapper[4789]: I1122 08:14:58.976933 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-d5fc-account-create-tbdhr"] Nov 22 08:14:58 crc kubenswrapper[4789]: W1122 08:14:58.992438 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod28b7b6e7_2620_4b4b_b3ab_2da093176b81.slice/crio-e2cc28fe39ec39a0219f3955b2eed1eb32fa3d2dcc4ce20b3823980d399352f3 WatchSource:0}: Error finding container e2cc28fe39ec39a0219f3955b2eed1eb32fa3d2dcc4ce20b3823980d399352f3: Status 404 returned error can't find the container with id e2cc28fe39ec39a0219f3955b2eed1eb32fa3d2dcc4ce20b3823980d399352f3 Nov 22 08:14:59 crc kubenswrapper[4789]: I1122 08:14:59.061911 4789 generic.go:334] "Generic (PLEG): container finished" podID="a5b8f813-91e7-4b56-803d-44167089fdf3" containerID="f3a3af787ea165ab682819efab7cdd97d21971e55c7d4175c91c2858a3196481" exitCode=0 Nov 22 08:14:59 crc kubenswrapper[4789]: I1122 08:14:59.062036 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-2sgdb" event={"ID":"a5b8f813-91e7-4b56-803d-44167089fdf3","Type":"ContainerDied","Data":"f3a3af787ea165ab682819efab7cdd97d21971e55c7d4175c91c2858a3196481"} Nov 22 08:14:59 crc kubenswrapper[4789]: I1122 08:14:59.062086 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-2sgdb" event={"ID":"a5b8f813-91e7-4b56-803d-44167089fdf3","Type":"ContainerStarted","Data":"a430218101b211ab32ab59169fe5c4c1f7e33f518397677eb50e9d51f70b2953"} Nov 22 08:14:59 crc kubenswrapper[4789]: I1122 08:14:59.064634 4789 generic.go:334] "Generic (PLEG): container finished" podID="033d4c4c-73bc-4aa4-a124-aff465a011d3" containerID="b30099df66f7f5c0ada8e51fea6fd51be78fed2a973b15f47861296095d47eb8" exitCode=0 Nov 22 08:14:59 crc kubenswrapper[4789]: I1122 08:14:59.064678 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-fdc5-account-create-lqqqs" event={"ID":"033d4c4c-73bc-4aa4-a124-aff465a011d3","Type":"ContainerDied","Data":"b30099df66f7f5c0ada8e51fea6fd51be78fed2a973b15f47861296095d47eb8"} Nov 22 08:14:59 crc kubenswrapper[4789]: I1122 08:14:59.064714 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-fdc5-account-create-lqqqs" event={"ID":"033d4c4c-73bc-4aa4-a124-aff465a011d3","Type":"ContainerStarted","Data":"7f2024cfc098b195b1cc722b50c6cf67491c011de84c07def82cf3450d37ca8c"} Nov 22 08:14:59 crc kubenswrapper[4789]: I1122 08:14:59.066615 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-b867h" event={"ID":"95dc84f1-08e8-4e12-8181-4a52830326c1","Type":"ContainerStarted","Data":"182aec76e43aa8c76574cf4f35eb8b948a8c3f29ee5bab1c406a601bb189642a"} Nov 22 08:14:59 crc kubenswrapper[4789]: I1122 08:14:59.066656 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-b867h" event={"ID":"95dc84f1-08e8-4e12-8181-4a52830326c1","Type":"ContainerStarted","Data":"a74bc1a8cbd9a4a35de96e242e51cdc3f0d7f6a353883bb747148f07b20b0455"} Nov 22 08:14:59 crc kubenswrapper[4789]: I1122 08:14:59.067922 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-d5fc-account-create-tbdhr" event={"ID":"28b7b6e7-2620-4b4b-b3ab-2da093176b81","Type":"ContainerStarted","Data":"e2cc28fe39ec39a0219f3955b2eed1eb32fa3d2dcc4ce20b3823980d399352f3"} Nov 22 08:14:59 crc kubenswrapper[4789]: I1122 08:14:59.103771 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-create-b867h" podStartSLOduration=2.103735914 podStartE2EDuration="2.103735914s" podCreationTimestamp="2025-11-22 08:14:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:14:59.098491661 +0000 UTC m=+1113.332891934" watchObservedRunningTime="2025-11-22 08:14:59.103735914 +0000 UTC m=+1113.338136187" Nov 22 08:14:59 crc kubenswrapper[4789]: I1122 08:14:59.154299 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-9bv8q"] Nov 22 08:14:59 crc kubenswrapper[4789]: W1122 08:14:59.157576 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddab42b3a_5fd5_43cf_801d_eeb4eef0ac98.slice/crio-a5b302ca0f39c9ee31e30fd87e2ee195c622df65a57a77fe996b43c2269968ef WatchSource:0}: Error finding container a5b302ca0f39c9ee31e30fd87e2ee195c622df65a57a77fe996b43c2269968ef: Status 404 returned error can't find the container with id a5b302ca0f39c9ee31e30fd87e2ee195c622df65a57a77fe996b43c2269968ef Nov 22 08:14:59 crc kubenswrapper[4789]: I1122 08:14:59.230585 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-e4c3-account-create-679fs"] Nov 22 08:14:59 crc kubenswrapper[4789]: W1122 08:14:59.249139 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod56293959_d97b_4df2_948e_d8422250945f.slice/crio-2a0c199a57b52847a700f00232e5f9948cd62a357d894166d84156e2c90a9cb2 WatchSource:0}: Error finding container 2a0c199a57b52847a700f00232e5f9948cd62a357d894166d84156e2c90a9cb2: Status 404 returned error can't find the container with id 2a0c199a57b52847a700f00232e5f9948cd62a357d894166d84156e2c90a9cb2 Nov 22 08:14:59 crc kubenswrapper[4789]: I1122 08:14:59.943136 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Nov 22 08:15:00 crc kubenswrapper[4789]: I1122 08:15:00.075869 4789 generic.go:334] "Generic (PLEG): container finished" podID="dab42b3a-5fd5-43cf-801d-eeb4eef0ac98" containerID="a7f9032ef73f529862f33e1e701f05ec568e012f2117e4463fe6532cb25f24b7" exitCode=0 Nov 22 08:15:00 crc kubenswrapper[4789]: I1122 08:15:00.075932 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-9bv8q" event={"ID":"dab42b3a-5fd5-43cf-801d-eeb4eef0ac98","Type":"ContainerDied","Data":"a7f9032ef73f529862f33e1e701f05ec568e012f2117e4463fe6532cb25f24b7"} Nov 22 08:15:00 crc kubenswrapper[4789]: I1122 08:15:00.075966 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-9bv8q" event={"ID":"dab42b3a-5fd5-43cf-801d-eeb4eef0ac98","Type":"ContainerStarted","Data":"a5b302ca0f39c9ee31e30fd87e2ee195c622df65a57a77fe996b43c2269968ef"} Nov 22 08:15:00 crc kubenswrapper[4789]: I1122 08:15:00.078894 4789 generic.go:334] "Generic (PLEG): container finished" podID="56293959-d97b-4df2-948e-d8422250945f" containerID="c9d9a47ab6213afb223079802b4594dbfeae14df6a8064de3fdfbc3d3211e5be" exitCode=0 Nov 22 08:15:00 crc kubenswrapper[4789]: I1122 08:15:00.078965 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-e4c3-account-create-679fs" event={"ID":"56293959-d97b-4df2-948e-d8422250945f","Type":"ContainerDied","Data":"c9d9a47ab6213afb223079802b4594dbfeae14df6a8064de3fdfbc3d3211e5be"} Nov 22 08:15:00 crc kubenswrapper[4789]: I1122 08:15:00.078995 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-e4c3-account-create-679fs" event={"ID":"56293959-d97b-4df2-948e-d8422250945f","Type":"ContainerStarted","Data":"2a0c199a57b52847a700f00232e5f9948cd62a357d894166d84156e2c90a9cb2"} Nov 22 08:15:00 crc kubenswrapper[4789]: I1122 08:15:00.080374 4789 generic.go:334] "Generic (PLEG): container finished" podID="28b7b6e7-2620-4b4b-b3ab-2da093176b81" containerID="c2f913b8d0601111137c4e3f14da86191b8ed52f5df71712374e6677e4851fbf" exitCode=0 Nov 22 08:15:00 crc kubenswrapper[4789]: I1122 08:15:00.080432 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-d5fc-account-create-tbdhr" event={"ID":"28b7b6e7-2620-4b4b-b3ab-2da093176b81","Type":"ContainerDied","Data":"c2f913b8d0601111137c4e3f14da86191b8ed52f5df71712374e6677e4851fbf"} Nov 22 08:15:00 crc kubenswrapper[4789]: I1122 08:15:00.081678 4789 generic.go:334] "Generic (PLEG): container finished" podID="95dc84f1-08e8-4e12-8181-4a52830326c1" containerID="182aec76e43aa8c76574cf4f35eb8b948a8c3f29ee5bab1c406a601bb189642a" exitCode=0 Nov 22 08:15:00 crc kubenswrapper[4789]: I1122 08:15:00.081715 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-b867h" event={"ID":"95dc84f1-08e8-4e12-8181-4a52830326c1","Type":"ContainerDied","Data":"182aec76e43aa8c76574cf4f35eb8b948a8c3f29ee5bab1c406a601bb189642a"} Nov 22 08:15:00 crc kubenswrapper[4789]: I1122 08:15:00.149635 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396655-pm4q9"] Nov 22 08:15:00 crc kubenswrapper[4789]: I1122 08:15:00.150735 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396655-pm4q9" Nov 22 08:15:00 crc kubenswrapper[4789]: I1122 08:15:00.152513 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 22 08:15:00 crc kubenswrapper[4789]: I1122 08:15:00.152737 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 22 08:15:00 crc kubenswrapper[4789]: I1122 08:15:00.168880 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396655-pm4q9"] Nov 22 08:15:00 crc kubenswrapper[4789]: I1122 08:15:00.245506 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2a71b7c2-042d-4b11-8b37-867c4dddddf2-config-volume\") pod \"collect-profiles-29396655-pm4q9\" (UID: \"2a71b7c2-042d-4b11-8b37-867c4dddddf2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396655-pm4q9" Nov 22 08:15:00 crc kubenswrapper[4789]: I1122 08:15:00.245696 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l25fh\" (UniqueName: \"kubernetes.io/projected/2a71b7c2-042d-4b11-8b37-867c4dddddf2-kube-api-access-l25fh\") pod \"collect-profiles-29396655-pm4q9\" (UID: \"2a71b7c2-042d-4b11-8b37-867c4dddddf2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396655-pm4q9" Nov 22 08:15:00 crc kubenswrapper[4789]: I1122 08:15:00.245868 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2a71b7c2-042d-4b11-8b37-867c4dddddf2-secret-volume\") pod \"collect-profiles-29396655-pm4q9\" (UID: \"2a71b7c2-042d-4b11-8b37-867c4dddddf2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396655-pm4q9" Nov 22 08:15:00 crc kubenswrapper[4789]: I1122 08:15:00.347603 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2a71b7c2-042d-4b11-8b37-867c4dddddf2-config-volume\") pod \"collect-profiles-29396655-pm4q9\" (UID: \"2a71b7c2-042d-4b11-8b37-867c4dddddf2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396655-pm4q9" Nov 22 08:15:00 crc kubenswrapper[4789]: I1122 08:15:00.347728 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l25fh\" (UniqueName: \"kubernetes.io/projected/2a71b7c2-042d-4b11-8b37-867c4dddddf2-kube-api-access-l25fh\") pod \"collect-profiles-29396655-pm4q9\" (UID: \"2a71b7c2-042d-4b11-8b37-867c4dddddf2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396655-pm4q9" Nov 22 08:15:00 crc kubenswrapper[4789]: I1122 08:15:00.347822 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2a71b7c2-042d-4b11-8b37-867c4dddddf2-secret-volume\") pod \"collect-profiles-29396655-pm4q9\" (UID: \"2a71b7c2-042d-4b11-8b37-867c4dddddf2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396655-pm4q9" Nov 22 08:15:00 crc kubenswrapper[4789]: I1122 08:15:00.348473 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2a71b7c2-042d-4b11-8b37-867c4dddddf2-config-volume\") pod \"collect-profiles-29396655-pm4q9\" (UID: \"2a71b7c2-042d-4b11-8b37-867c4dddddf2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396655-pm4q9" Nov 22 08:15:00 crc kubenswrapper[4789]: I1122 08:15:00.353734 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2a71b7c2-042d-4b11-8b37-867c4dddddf2-secret-volume\") pod \"collect-profiles-29396655-pm4q9\" (UID: \"2a71b7c2-042d-4b11-8b37-867c4dddddf2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396655-pm4q9" Nov 22 08:15:00 crc kubenswrapper[4789]: I1122 08:15:00.373239 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l25fh\" (UniqueName: \"kubernetes.io/projected/2a71b7c2-042d-4b11-8b37-867c4dddddf2-kube-api-access-l25fh\") pod \"collect-profiles-29396655-pm4q9\" (UID: \"2a71b7c2-042d-4b11-8b37-867c4dddddf2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396655-pm4q9" Nov 22 08:15:00 crc kubenswrapper[4789]: I1122 08:15:00.469135 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396655-pm4q9" Nov 22 08:15:00 crc kubenswrapper[4789]: I1122 08:15:00.469300 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-fdc5-account-create-lqqqs" Nov 22 08:15:00 crc kubenswrapper[4789]: I1122 08:15:00.482291 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-2sgdb" Nov 22 08:15:00 crc kubenswrapper[4789]: I1122 08:15:00.550313 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qnhdz\" (UniqueName: \"kubernetes.io/projected/a5b8f813-91e7-4b56-803d-44167089fdf3-kube-api-access-qnhdz\") pod \"a5b8f813-91e7-4b56-803d-44167089fdf3\" (UID: \"a5b8f813-91e7-4b56-803d-44167089fdf3\") " Nov 22 08:15:00 crc kubenswrapper[4789]: I1122 08:15:00.550403 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a5b8f813-91e7-4b56-803d-44167089fdf3-operator-scripts\") pod \"a5b8f813-91e7-4b56-803d-44167089fdf3\" (UID: \"a5b8f813-91e7-4b56-803d-44167089fdf3\") " Nov 22 08:15:00 crc kubenswrapper[4789]: I1122 08:15:00.550459 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/033d4c4c-73bc-4aa4-a124-aff465a011d3-operator-scripts\") pod \"033d4c4c-73bc-4aa4-a124-aff465a011d3\" (UID: \"033d4c4c-73bc-4aa4-a124-aff465a011d3\") " Nov 22 08:15:00 crc kubenswrapper[4789]: I1122 08:15:00.550504 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c6mc6\" (UniqueName: \"kubernetes.io/projected/033d4c4c-73bc-4aa4-a124-aff465a011d3-kube-api-access-c6mc6\") pod \"033d4c4c-73bc-4aa4-a124-aff465a011d3\" (UID: \"033d4c4c-73bc-4aa4-a124-aff465a011d3\") " Nov 22 08:15:00 crc kubenswrapper[4789]: I1122 08:15:00.551338 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a5b8f813-91e7-4b56-803d-44167089fdf3-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a5b8f813-91e7-4b56-803d-44167089fdf3" (UID: "a5b8f813-91e7-4b56-803d-44167089fdf3"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:15:00 crc kubenswrapper[4789]: I1122 08:15:00.551737 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/033d4c4c-73bc-4aa4-a124-aff465a011d3-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "033d4c4c-73bc-4aa4-a124-aff465a011d3" (UID: "033d4c4c-73bc-4aa4-a124-aff465a011d3"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:15:00 crc kubenswrapper[4789]: I1122 08:15:00.554594 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/033d4c4c-73bc-4aa4-a124-aff465a011d3-kube-api-access-c6mc6" (OuterVolumeSpecName: "kube-api-access-c6mc6") pod "033d4c4c-73bc-4aa4-a124-aff465a011d3" (UID: "033d4c4c-73bc-4aa4-a124-aff465a011d3"). InnerVolumeSpecName "kube-api-access-c6mc6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:15:00 crc kubenswrapper[4789]: I1122 08:15:00.554977 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5b8f813-91e7-4b56-803d-44167089fdf3-kube-api-access-qnhdz" (OuterVolumeSpecName: "kube-api-access-qnhdz") pod "a5b8f813-91e7-4b56-803d-44167089fdf3" (UID: "a5b8f813-91e7-4b56-803d-44167089fdf3"). InnerVolumeSpecName "kube-api-access-qnhdz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:15:00 crc kubenswrapper[4789]: I1122 08:15:00.652375 4789 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a5b8f813-91e7-4b56-803d-44167089fdf3-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:15:00 crc kubenswrapper[4789]: I1122 08:15:00.652719 4789 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/033d4c4c-73bc-4aa4-a124-aff465a011d3-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:15:00 crc kubenswrapper[4789]: I1122 08:15:00.652733 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c6mc6\" (UniqueName: \"kubernetes.io/projected/033d4c4c-73bc-4aa4-a124-aff465a011d3-kube-api-access-c6mc6\") on node \"crc\" DevicePath \"\"" Nov 22 08:15:00 crc kubenswrapper[4789]: I1122 08:15:00.652747 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qnhdz\" (UniqueName: \"kubernetes.io/projected/a5b8f813-91e7-4b56-803d-44167089fdf3-kube-api-access-qnhdz\") on node \"crc\" DevicePath \"\"" Nov 22 08:15:00 crc kubenswrapper[4789]: I1122 08:15:00.896097 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396655-pm4q9"] Nov 22 08:15:01 crc kubenswrapper[4789]: I1122 08:15:01.089985 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-2sgdb" Nov 22 08:15:01 crc kubenswrapper[4789]: I1122 08:15:01.090020 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-2sgdb" event={"ID":"a5b8f813-91e7-4b56-803d-44167089fdf3","Type":"ContainerDied","Data":"a430218101b211ab32ab59169fe5c4c1f7e33f518397677eb50e9d51f70b2953"} Nov 22 08:15:01 crc kubenswrapper[4789]: I1122 08:15:01.090058 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a430218101b211ab32ab59169fe5c4c1f7e33f518397677eb50e9d51f70b2953" Nov 22 08:15:01 crc kubenswrapper[4789]: I1122 08:15:01.093018 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-fdc5-account-create-lqqqs" event={"ID":"033d4c4c-73bc-4aa4-a124-aff465a011d3","Type":"ContainerDied","Data":"7f2024cfc098b195b1cc722b50c6cf67491c011de84c07def82cf3450d37ca8c"} Nov 22 08:15:01 crc kubenswrapper[4789]: I1122 08:15:01.093047 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7f2024cfc098b195b1cc722b50c6cf67491c011de84c07def82cf3450d37ca8c" Nov 22 08:15:01 crc kubenswrapper[4789]: I1122 08:15:01.093079 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-fdc5-account-create-lqqqs" Nov 22 08:15:01 crc kubenswrapper[4789]: I1122 08:15:01.097980 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29396655-pm4q9" event={"ID":"2a71b7c2-042d-4b11-8b37-867c4dddddf2","Type":"ContainerStarted","Data":"254a79e2a772d7a8f8af426f899fa26e8d86e738a8471b562508261c46095767"} Nov 22 08:15:01 crc kubenswrapper[4789]: I1122 08:15:01.098015 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29396655-pm4q9" event={"ID":"2a71b7c2-042d-4b11-8b37-867c4dddddf2","Type":"ContainerStarted","Data":"e758d1b9f70d2a137e96e20e65a64765e04cef7d9a279eb27f9958bb68c00536"} Nov 22 08:15:01 crc kubenswrapper[4789]: I1122 08:15:01.114339 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29396655-pm4q9" podStartSLOduration=1.114321061 podStartE2EDuration="1.114321061s" podCreationTimestamp="2025-11-22 08:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:15:01.112509542 +0000 UTC m=+1115.346909825" watchObservedRunningTime="2025-11-22 08:15:01.114321061 +0000 UTC m=+1115.348721334" Nov 22 08:15:01 crc kubenswrapper[4789]: I1122 08:15:01.469113 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-e4c3-account-create-679fs" Nov 22 08:15:01 crc kubenswrapper[4789]: I1122 08:15:01.568462 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/56293959-d97b-4df2-948e-d8422250945f-operator-scripts\") pod \"56293959-d97b-4df2-948e-d8422250945f\" (UID: \"56293959-d97b-4df2-948e-d8422250945f\") " Nov 22 08:15:01 crc kubenswrapper[4789]: I1122 08:15:01.568717 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xdh8j\" (UniqueName: \"kubernetes.io/projected/56293959-d97b-4df2-948e-d8422250945f-kube-api-access-xdh8j\") pod \"56293959-d97b-4df2-948e-d8422250945f\" (UID: \"56293959-d97b-4df2-948e-d8422250945f\") " Nov 22 08:15:01 crc kubenswrapper[4789]: I1122 08:15:01.569667 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/56293959-d97b-4df2-948e-d8422250945f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "56293959-d97b-4df2-948e-d8422250945f" (UID: "56293959-d97b-4df2-948e-d8422250945f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:15:01 crc kubenswrapper[4789]: I1122 08:15:01.574823 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56293959-d97b-4df2-948e-d8422250945f-kube-api-access-xdh8j" (OuterVolumeSpecName: "kube-api-access-xdh8j") pod "56293959-d97b-4df2-948e-d8422250945f" (UID: "56293959-d97b-4df2-948e-d8422250945f"). InnerVolumeSpecName "kube-api-access-xdh8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:15:01 crc kubenswrapper[4789]: I1122 08:15:01.643542 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-d5fc-account-create-tbdhr" Nov 22 08:15:01 crc kubenswrapper[4789]: I1122 08:15:01.653332 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-9bv8q" Nov 22 08:15:01 crc kubenswrapper[4789]: I1122 08:15:01.665778 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-b867h" Nov 22 08:15:01 crc kubenswrapper[4789]: I1122 08:15:01.670265 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fjnff\" (UniqueName: \"kubernetes.io/projected/28b7b6e7-2620-4b4b-b3ab-2da093176b81-kube-api-access-fjnff\") pod \"28b7b6e7-2620-4b4b-b3ab-2da093176b81\" (UID: \"28b7b6e7-2620-4b4b-b3ab-2da093176b81\") " Nov 22 08:15:01 crc kubenswrapper[4789]: I1122 08:15:01.670562 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/28b7b6e7-2620-4b4b-b3ab-2da093176b81-operator-scripts\") pod \"28b7b6e7-2620-4b4b-b3ab-2da093176b81\" (UID: \"28b7b6e7-2620-4b4b-b3ab-2da093176b81\") " Nov 22 08:15:01 crc kubenswrapper[4789]: I1122 08:15:01.670957 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xdh8j\" (UniqueName: \"kubernetes.io/projected/56293959-d97b-4df2-948e-d8422250945f-kube-api-access-xdh8j\") on node \"crc\" DevicePath \"\"" Nov 22 08:15:01 crc kubenswrapper[4789]: I1122 08:15:01.671076 4789 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/56293959-d97b-4df2-948e-d8422250945f-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:15:01 crc kubenswrapper[4789]: I1122 08:15:01.671327 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/28b7b6e7-2620-4b4b-b3ab-2da093176b81-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "28b7b6e7-2620-4b4b-b3ab-2da093176b81" (UID: "28b7b6e7-2620-4b4b-b3ab-2da093176b81"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:15:01 crc kubenswrapper[4789]: I1122 08:15:01.677842 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/28b7b6e7-2620-4b4b-b3ab-2da093176b81-kube-api-access-fjnff" (OuterVolumeSpecName: "kube-api-access-fjnff") pod "28b7b6e7-2620-4b4b-b3ab-2da093176b81" (UID: "28b7b6e7-2620-4b4b-b3ab-2da093176b81"). InnerVolumeSpecName "kube-api-access-fjnff". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:15:01 crc kubenswrapper[4789]: I1122 08:15:01.772254 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r4kjl\" (UniqueName: \"kubernetes.io/projected/dab42b3a-5fd5-43cf-801d-eeb4eef0ac98-kube-api-access-r4kjl\") pod \"dab42b3a-5fd5-43cf-801d-eeb4eef0ac98\" (UID: \"dab42b3a-5fd5-43cf-801d-eeb4eef0ac98\") " Nov 22 08:15:01 crc kubenswrapper[4789]: I1122 08:15:01.772407 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dab42b3a-5fd5-43cf-801d-eeb4eef0ac98-operator-scripts\") pod \"dab42b3a-5fd5-43cf-801d-eeb4eef0ac98\" (UID: \"dab42b3a-5fd5-43cf-801d-eeb4eef0ac98\") " Nov 22 08:15:01 crc kubenswrapper[4789]: I1122 08:15:01.772475 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/95dc84f1-08e8-4e12-8181-4a52830326c1-operator-scripts\") pod \"95dc84f1-08e8-4e12-8181-4a52830326c1\" (UID: \"95dc84f1-08e8-4e12-8181-4a52830326c1\") " Nov 22 08:15:01 crc kubenswrapper[4789]: I1122 08:15:01.772541 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dktxh\" (UniqueName: \"kubernetes.io/projected/95dc84f1-08e8-4e12-8181-4a52830326c1-kube-api-access-dktxh\") pod \"95dc84f1-08e8-4e12-8181-4a52830326c1\" (UID: \"95dc84f1-08e8-4e12-8181-4a52830326c1\") " Nov 22 08:15:01 crc kubenswrapper[4789]: I1122 08:15:01.773028 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/95dc84f1-08e8-4e12-8181-4a52830326c1-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "95dc84f1-08e8-4e12-8181-4a52830326c1" (UID: "95dc84f1-08e8-4e12-8181-4a52830326c1"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:15:01 crc kubenswrapper[4789]: I1122 08:15:01.773055 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dab42b3a-5fd5-43cf-801d-eeb4eef0ac98-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "dab42b3a-5fd5-43cf-801d-eeb4eef0ac98" (UID: "dab42b3a-5fd5-43cf-801d-eeb4eef0ac98"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:15:01 crc kubenswrapper[4789]: I1122 08:15:01.773137 4789 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/28b7b6e7-2620-4b4b-b3ab-2da093176b81-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:15:01 crc kubenswrapper[4789]: I1122 08:15:01.773175 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fjnff\" (UniqueName: \"kubernetes.io/projected/28b7b6e7-2620-4b4b-b3ab-2da093176b81-kube-api-access-fjnff\") on node \"crc\" DevicePath \"\"" Nov 22 08:15:01 crc kubenswrapper[4789]: I1122 08:15:01.777191 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95dc84f1-08e8-4e12-8181-4a52830326c1-kube-api-access-dktxh" (OuterVolumeSpecName: "kube-api-access-dktxh") pod "95dc84f1-08e8-4e12-8181-4a52830326c1" (UID: "95dc84f1-08e8-4e12-8181-4a52830326c1"). InnerVolumeSpecName "kube-api-access-dktxh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:15:01 crc kubenswrapper[4789]: I1122 08:15:01.777301 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dab42b3a-5fd5-43cf-801d-eeb4eef0ac98-kube-api-access-r4kjl" (OuterVolumeSpecName: "kube-api-access-r4kjl") pod "dab42b3a-5fd5-43cf-801d-eeb4eef0ac98" (UID: "dab42b3a-5fd5-43cf-801d-eeb4eef0ac98"). InnerVolumeSpecName "kube-api-access-r4kjl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:15:01 crc kubenswrapper[4789]: I1122 08:15:01.875273 4789 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dab42b3a-5fd5-43cf-801d-eeb4eef0ac98-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:15:01 crc kubenswrapper[4789]: I1122 08:15:01.875312 4789 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/95dc84f1-08e8-4e12-8181-4a52830326c1-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:15:01 crc kubenswrapper[4789]: I1122 08:15:01.875324 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dktxh\" (UniqueName: \"kubernetes.io/projected/95dc84f1-08e8-4e12-8181-4a52830326c1-kube-api-access-dktxh\") on node \"crc\" DevicePath \"\"" Nov 22 08:15:01 crc kubenswrapper[4789]: I1122 08:15:01.875335 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r4kjl\" (UniqueName: \"kubernetes.io/projected/dab42b3a-5fd5-43cf-801d-eeb4eef0ac98-kube-api-access-r4kjl\") on node \"crc\" DevicePath \"\"" Nov 22 08:15:02 crc kubenswrapper[4789]: I1122 08:15:02.106513 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-9bv8q" event={"ID":"dab42b3a-5fd5-43cf-801d-eeb4eef0ac98","Type":"ContainerDied","Data":"a5b302ca0f39c9ee31e30fd87e2ee195c622df65a57a77fe996b43c2269968ef"} Nov 22 08:15:02 crc kubenswrapper[4789]: I1122 08:15:02.106572 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-9bv8q" Nov 22 08:15:02 crc kubenswrapper[4789]: I1122 08:15:02.106621 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a5b302ca0f39c9ee31e30fd87e2ee195c622df65a57a77fe996b43c2269968ef" Nov 22 08:15:02 crc kubenswrapper[4789]: I1122 08:15:02.109009 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-d5fc-account-create-tbdhr" event={"ID":"28b7b6e7-2620-4b4b-b3ab-2da093176b81","Type":"ContainerDied","Data":"e2cc28fe39ec39a0219f3955b2eed1eb32fa3d2dcc4ce20b3823980d399352f3"} Nov 22 08:15:02 crc kubenswrapper[4789]: I1122 08:15:02.109050 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e2cc28fe39ec39a0219f3955b2eed1eb32fa3d2dcc4ce20b3823980d399352f3" Nov 22 08:15:02 crc kubenswrapper[4789]: I1122 08:15:02.109119 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-d5fc-account-create-tbdhr" Nov 22 08:15:02 crc kubenswrapper[4789]: I1122 08:15:02.111075 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-b867h" event={"ID":"95dc84f1-08e8-4e12-8181-4a52830326c1","Type":"ContainerDied","Data":"a74bc1a8cbd9a4a35de96e242e51cdc3f0d7f6a353883bb747148f07b20b0455"} Nov 22 08:15:02 crc kubenswrapper[4789]: I1122 08:15:02.111230 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a74bc1a8cbd9a4a35de96e242e51cdc3f0d7f6a353883bb747148f07b20b0455" Nov 22 08:15:02 crc kubenswrapper[4789]: I1122 08:15:02.111098 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-b867h" Nov 22 08:15:02 crc kubenswrapper[4789]: I1122 08:15:02.113648 4789 generic.go:334] "Generic (PLEG): container finished" podID="2a71b7c2-042d-4b11-8b37-867c4dddddf2" containerID="254a79e2a772d7a8f8af426f899fa26e8d86e738a8471b562508261c46095767" exitCode=0 Nov 22 08:15:02 crc kubenswrapper[4789]: I1122 08:15:02.113713 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29396655-pm4q9" event={"ID":"2a71b7c2-042d-4b11-8b37-867c4dddddf2","Type":"ContainerDied","Data":"254a79e2a772d7a8f8af426f899fa26e8d86e738a8471b562508261c46095767"} Nov 22 08:15:02 crc kubenswrapper[4789]: I1122 08:15:02.116668 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-e4c3-account-create-679fs" event={"ID":"56293959-d97b-4df2-948e-d8422250945f","Type":"ContainerDied","Data":"2a0c199a57b52847a700f00232e5f9948cd62a357d894166d84156e2c90a9cb2"} Nov 22 08:15:02 crc kubenswrapper[4789]: I1122 08:15:02.116706 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2a0c199a57b52847a700f00232e5f9948cd62a357d894166d84156e2c90a9cb2" Nov 22 08:15:02 crc kubenswrapper[4789]: I1122 08:15:02.116731 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-e4c3-account-create-679fs" Nov 22 08:15:02 crc kubenswrapper[4789]: I1122 08:15:02.623496 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-65c9b8d4f7-kfsdh" Nov 22 08:15:03 crc kubenswrapper[4789]: I1122 08:15:03.063970 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5c476d78c5-djb2g" Nov 22 08:15:03 crc kubenswrapper[4789]: I1122 08:15:03.114050 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-65c9b8d4f7-kfsdh"] Nov 22 08:15:03 crc kubenswrapper[4789]: I1122 08:15:03.125095 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-65c9b8d4f7-kfsdh" podUID="2e9862f9-e69a-472f-9db9-638c95ee3912" containerName="dnsmasq-dns" containerID="cri-o://449fe80a53cf4a823fa98663c8779562ef32de7a57fd4ceac3350994ab0ee9d8" gracePeriod=10 Nov 22 08:15:03 crc kubenswrapper[4789]: I1122 08:15:03.458686 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396655-pm4q9" Nov 22 08:15:03 crc kubenswrapper[4789]: I1122 08:15:03.493601 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-xhfrw"] Nov 22 08:15:03 crc kubenswrapper[4789]: E1122 08:15:03.493979 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dab42b3a-5fd5-43cf-801d-eeb4eef0ac98" containerName="mariadb-database-create" Nov 22 08:15:03 crc kubenswrapper[4789]: I1122 08:15:03.494001 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="dab42b3a-5fd5-43cf-801d-eeb4eef0ac98" containerName="mariadb-database-create" Nov 22 08:15:03 crc kubenswrapper[4789]: E1122 08:15:03.494016 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5b8f813-91e7-4b56-803d-44167089fdf3" containerName="mariadb-database-create" Nov 22 08:15:03 crc kubenswrapper[4789]: I1122 08:15:03.494023 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5b8f813-91e7-4b56-803d-44167089fdf3" containerName="mariadb-database-create" Nov 22 08:15:03 crc kubenswrapper[4789]: E1122 08:15:03.494033 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a71b7c2-042d-4b11-8b37-867c4dddddf2" containerName="collect-profiles" Nov 22 08:15:03 crc kubenswrapper[4789]: I1122 08:15:03.494039 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a71b7c2-042d-4b11-8b37-867c4dddddf2" containerName="collect-profiles" Nov 22 08:15:03 crc kubenswrapper[4789]: E1122 08:15:03.494054 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95dc84f1-08e8-4e12-8181-4a52830326c1" containerName="mariadb-database-create" Nov 22 08:15:03 crc kubenswrapper[4789]: I1122 08:15:03.494061 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="95dc84f1-08e8-4e12-8181-4a52830326c1" containerName="mariadb-database-create" Nov 22 08:15:03 crc kubenswrapper[4789]: E1122 08:15:03.494076 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28b7b6e7-2620-4b4b-b3ab-2da093176b81" containerName="mariadb-account-create" Nov 22 08:15:03 crc kubenswrapper[4789]: I1122 08:15:03.494082 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="28b7b6e7-2620-4b4b-b3ab-2da093176b81" containerName="mariadb-account-create" Nov 22 08:15:03 crc kubenswrapper[4789]: E1122 08:15:03.494093 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="033d4c4c-73bc-4aa4-a124-aff465a011d3" containerName="mariadb-account-create" Nov 22 08:15:03 crc kubenswrapper[4789]: I1122 08:15:03.494099 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="033d4c4c-73bc-4aa4-a124-aff465a011d3" containerName="mariadb-account-create" Nov 22 08:15:03 crc kubenswrapper[4789]: E1122 08:15:03.494111 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56293959-d97b-4df2-948e-d8422250945f" containerName="mariadb-account-create" Nov 22 08:15:03 crc kubenswrapper[4789]: I1122 08:15:03.494117 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="56293959-d97b-4df2-948e-d8422250945f" containerName="mariadb-account-create" Nov 22 08:15:03 crc kubenswrapper[4789]: I1122 08:15:03.494256 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="56293959-d97b-4df2-948e-d8422250945f" containerName="mariadb-account-create" Nov 22 08:15:03 crc kubenswrapper[4789]: I1122 08:15:03.494276 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="dab42b3a-5fd5-43cf-801d-eeb4eef0ac98" containerName="mariadb-database-create" Nov 22 08:15:03 crc kubenswrapper[4789]: I1122 08:15:03.494294 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a71b7c2-042d-4b11-8b37-867c4dddddf2" containerName="collect-profiles" Nov 22 08:15:03 crc kubenswrapper[4789]: I1122 08:15:03.494302 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="28b7b6e7-2620-4b4b-b3ab-2da093176b81" containerName="mariadb-account-create" Nov 22 08:15:03 crc kubenswrapper[4789]: I1122 08:15:03.494310 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="033d4c4c-73bc-4aa4-a124-aff465a011d3" containerName="mariadb-account-create" Nov 22 08:15:03 crc kubenswrapper[4789]: I1122 08:15:03.494318 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5b8f813-91e7-4b56-803d-44167089fdf3" containerName="mariadb-database-create" Nov 22 08:15:03 crc kubenswrapper[4789]: I1122 08:15:03.494327 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="95dc84f1-08e8-4e12-8181-4a52830326c1" containerName="mariadb-database-create" Nov 22 08:15:03 crc kubenswrapper[4789]: I1122 08:15:03.494827 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-xhfrw" Nov 22 08:15:03 crc kubenswrapper[4789]: I1122 08:15:03.499455 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-pd2j7" Nov 22 08:15:03 crc kubenswrapper[4789]: I1122 08:15:03.501585 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Nov 22 08:15:03 crc kubenswrapper[4789]: I1122 08:15:03.503924 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2a71b7c2-042d-4b11-8b37-867c4dddddf2-config-volume\") pod \"2a71b7c2-042d-4b11-8b37-867c4dddddf2\" (UID: \"2a71b7c2-042d-4b11-8b37-867c4dddddf2\") " Nov 22 08:15:03 crc kubenswrapper[4789]: I1122 08:15:03.504086 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2a71b7c2-042d-4b11-8b37-867c4dddddf2-secret-volume\") pod \"2a71b7c2-042d-4b11-8b37-867c4dddddf2\" (UID: \"2a71b7c2-042d-4b11-8b37-867c4dddddf2\") " Nov 22 08:15:03 crc kubenswrapper[4789]: I1122 08:15:03.504162 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l25fh\" (UniqueName: \"kubernetes.io/projected/2a71b7c2-042d-4b11-8b37-867c4dddddf2-kube-api-access-l25fh\") pod \"2a71b7c2-042d-4b11-8b37-867c4dddddf2\" (UID: \"2a71b7c2-042d-4b11-8b37-867c4dddddf2\") " Nov 22 08:15:03 crc kubenswrapper[4789]: I1122 08:15:03.505220 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2a71b7c2-042d-4b11-8b37-867c4dddddf2-config-volume" (OuterVolumeSpecName: "config-volume") pod "2a71b7c2-042d-4b11-8b37-867c4dddddf2" (UID: "2a71b7c2-042d-4b11-8b37-867c4dddddf2"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:15:03 crc kubenswrapper[4789]: I1122 08:15:03.510431 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a71b7c2-042d-4b11-8b37-867c4dddddf2-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "2a71b7c2-042d-4b11-8b37-867c4dddddf2" (UID: "2a71b7c2-042d-4b11-8b37-867c4dddddf2"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:15:03 crc kubenswrapper[4789]: I1122 08:15:03.524335 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a71b7c2-042d-4b11-8b37-867c4dddddf2-kube-api-access-l25fh" (OuterVolumeSpecName: "kube-api-access-l25fh") pod "2a71b7c2-042d-4b11-8b37-867c4dddddf2" (UID: "2a71b7c2-042d-4b11-8b37-867c4dddddf2"). InnerVolumeSpecName "kube-api-access-l25fh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:15:03 crc kubenswrapper[4789]: I1122 08:15:03.541501 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-xhfrw"] Nov 22 08:15:03 crc kubenswrapper[4789]: I1122 08:15:03.605577 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/afee6663-d80e-41bc-af23-7d64a23d4b03-db-sync-config-data\") pod \"glance-db-sync-xhfrw\" (UID: \"afee6663-d80e-41bc-af23-7d64a23d4b03\") " pod="openstack/glance-db-sync-xhfrw" Nov 22 08:15:03 crc kubenswrapper[4789]: I1122 08:15:03.605633 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/afee6663-d80e-41bc-af23-7d64a23d4b03-config-data\") pod \"glance-db-sync-xhfrw\" (UID: \"afee6663-d80e-41bc-af23-7d64a23d4b03\") " pod="openstack/glance-db-sync-xhfrw" Nov 22 08:15:03 crc kubenswrapper[4789]: I1122 08:15:03.605657 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/afee6663-d80e-41bc-af23-7d64a23d4b03-combined-ca-bundle\") pod \"glance-db-sync-xhfrw\" (UID: \"afee6663-d80e-41bc-af23-7d64a23d4b03\") " pod="openstack/glance-db-sync-xhfrw" Nov 22 08:15:03 crc kubenswrapper[4789]: I1122 08:15:03.605712 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-78d92\" (UniqueName: \"kubernetes.io/projected/afee6663-d80e-41bc-af23-7d64a23d4b03-kube-api-access-78d92\") pod \"glance-db-sync-xhfrw\" (UID: \"afee6663-d80e-41bc-af23-7d64a23d4b03\") " pod="openstack/glance-db-sync-xhfrw" Nov 22 08:15:03 crc kubenswrapper[4789]: I1122 08:15:03.605843 4789 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2a71b7c2-042d-4b11-8b37-867c4dddddf2-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 22 08:15:03 crc kubenswrapper[4789]: I1122 08:15:03.605862 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l25fh\" (UniqueName: \"kubernetes.io/projected/2a71b7c2-042d-4b11-8b37-867c4dddddf2-kube-api-access-l25fh\") on node \"crc\" DevicePath \"\"" Nov 22 08:15:03 crc kubenswrapper[4789]: I1122 08:15:03.605873 4789 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2a71b7c2-042d-4b11-8b37-867c4dddddf2-config-volume\") on node \"crc\" DevicePath \"\"" Nov 22 08:15:03 crc kubenswrapper[4789]: I1122 08:15:03.706725 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/afee6663-d80e-41bc-af23-7d64a23d4b03-db-sync-config-data\") pod \"glance-db-sync-xhfrw\" (UID: \"afee6663-d80e-41bc-af23-7d64a23d4b03\") " pod="openstack/glance-db-sync-xhfrw" Nov 22 08:15:03 crc kubenswrapper[4789]: I1122 08:15:03.706809 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/afee6663-d80e-41bc-af23-7d64a23d4b03-config-data\") pod \"glance-db-sync-xhfrw\" (UID: \"afee6663-d80e-41bc-af23-7d64a23d4b03\") " pod="openstack/glance-db-sync-xhfrw" Nov 22 08:15:03 crc kubenswrapper[4789]: I1122 08:15:03.706836 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/afee6663-d80e-41bc-af23-7d64a23d4b03-combined-ca-bundle\") pod \"glance-db-sync-xhfrw\" (UID: \"afee6663-d80e-41bc-af23-7d64a23d4b03\") " pod="openstack/glance-db-sync-xhfrw" Nov 22 08:15:03 crc kubenswrapper[4789]: I1122 08:15:03.706894 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-78d92\" (UniqueName: \"kubernetes.io/projected/afee6663-d80e-41bc-af23-7d64a23d4b03-kube-api-access-78d92\") pod \"glance-db-sync-xhfrw\" (UID: \"afee6663-d80e-41bc-af23-7d64a23d4b03\") " pod="openstack/glance-db-sync-xhfrw" Nov 22 08:15:03 crc kubenswrapper[4789]: I1122 08:15:03.711087 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/afee6663-d80e-41bc-af23-7d64a23d4b03-db-sync-config-data\") pod \"glance-db-sync-xhfrw\" (UID: \"afee6663-d80e-41bc-af23-7d64a23d4b03\") " pod="openstack/glance-db-sync-xhfrw" Nov 22 08:15:03 crc kubenswrapper[4789]: I1122 08:15:03.716503 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/afee6663-d80e-41bc-af23-7d64a23d4b03-combined-ca-bundle\") pod \"glance-db-sync-xhfrw\" (UID: \"afee6663-d80e-41bc-af23-7d64a23d4b03\") " pod="openstack/glance-db-sync-xhfrw" Nov 22 08:15:03 crc kubenswrapper[4789]: I1122 08:15:03.717024 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/afee6663-d80e-41bc-af23-7d64a23d4b03-config-data\") pod \"glance-db-sync-xhfrw\" (UID: \"afee6663-d80e-41bc-af23-7d64a23d4b03\") " pod="openstack/glance-db-sync-xhfrw" Nov 22 08:15:03 crc kubenswrapper[4789]: I1122 08:15:03.729295 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-78d92\" (UniqueName: \"kubernetes.io/projected/afee6663-d80e-41bc-af23-7d64a23d4b03-kube-api-access-78d92\") pod \"glance-db-sync-xhfrw\" (UID: \"afee6663-d80e-41bc-af23-7d64a23d4b03\") " pod="openstack/glance-db-sync-xhfrw" Nov 22 08:15:03 crc kubenswrapper[4789]: I1122 08:15:03.945689 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-xhfrw" Nov 22 08:15:04 crc kubenswrapper[4789]: I1122 08:15:04.134932 4789 generic.go:334] "Generic (PLEG): container finished" podID="2e9862f9-e69a-472f-9db9-638c95ee3912" containerID="449fe80a53cf4a823fa98663c8779562ef32de7a57fd4ceac3350994ab0ee9d8" exitCode=0 Nov 22 08:15:04 crc kubenswrapper[4789]: I1122 08:15:04.135011 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65c9b8d4f7-kfsdh" event={"ID":"2e9862f9-e69a-472f-9db9-638c95ee3912","Type":"ContainerDied","Data":"449fe80a53cf4a823fa98663c8779562ef32de7a57fd4ceac3350994ab0ee9d8"} Nov 22 08:15:04 crc kubenswrapper[4789]: I1122 08:15:04.137347 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29396655-pm4q9" event={"ID":"2a71b7c2-042d-4b11-8b37-867c4dddddf2","Type":"ContainerDied","Data":"e758d1b9f70d2a137e96e20e65a64765e04cef7d9a279eb27f9958bb68c00536"} Nov 22 08:15:04 crc kubenswrapper[4789]: I1122 08:15:04.137390 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e758d1b9f70d2a137e96e20e65a64765e04cef7d9a279eb27f9958bb68c00536" Nov 22 08:15:04 crc kubenswrapper[4789]: I1122 08:15:04.137482 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396655-pm4q9" Nov 22 08:15:04 crc kubenswrapper[4789]: I1122 08:15:04.270237 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-65c9b8d4f7-kfsdh" Nov 22 08:15:04 crc kubenswrapper[4789]: I1122 08:15:04.314635 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2e9862f9-e69a-472f-9db9-638c95ee3912-dns-svc\") pod \"2e9862f9-e69a-472f-9db9-638c95ee3912\" (UID: \"2e9862f9-e69a-472f-9db9-638c95ee3912\") " Nov 22 08:15:04 crc kubenswrapper[4789]: I1122 08:15:04.314722 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2e9862f9-e69a-472f-9db9-638c95ee3912-ovsdbserver-sb\") pod \"2e9862f9-e69a-472f-9db9-638c95ee3912\" (UID: \"2e9862f9-e69a-472f-9db9-638c95ee3912\") " Nov 22 08:15:04 crc kubenswrapper[4789]: I1122 08:15:04.314822 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h5hkh\" (UniqueName: \"kubernetes.io/projected/2e9862f9-e69a-472f-9db9-638c95ee3912-kube-api-access-h5hkh\") pod \"2e9862f9-e69a-472f-9db9-638c95ee3912\" (UID: \"2e9862f9-e69a-472f-9db9-638c95ee3912\") " Nov 22 08:15:04 crc kubenswrapper[4789]: I1122 08:15:04.314881 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2e9862f9-e69a-472f-9db9-638c95ee3912-config\") pod \"2e9862f9-e69a-472f-9db9-638c95ee3912\" (UID: \"2e9862f9-e69a-472f-9db9-638c95ee3912\") " Nov 22 08:15:04 crc kubenswrapper[4789]: I1122 08:15:04.320282 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e9862f9-e69a-472f-9db9-638c95ee3912-kube-api-access-h5hkh" (OuterVolumeSpecName: "kube-api-access-h5hkh") pod "2e9862f9-e69a-472f-9db9-638c95ee3912" (UID: "2e9862f9-e69a-472f-9db9-638c95ee3912"). InnerVolumeSpecName "kube-api-access-h5hkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:15:04 crc kubenswrapper[4789]: I1122 08:15:04.361241 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2e9862f9-e69a-472f-9db9-638c95ee3912-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "2e9862f9-e69a-472f-9db9-638c95ee3912" (UID: "2e9862f9-e69a-472f-9db9-638c95ee3912"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:15:04 crc kubenswrapper[4789]: I1122 08:15:04.367002 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2e9862f9-e69a-472f-9db9-638c95ee3912-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "2e9862f9-e69a-472f-9db9-638c95ee3912" (UID: "2e9862f9-e69a-472f-9db9-638c95ee3912"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:15:04 crc kubenswrapper[4789]: I1122 08:15:04.372071 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2e9862f9-e69a-472f-9db9-638c95ee3912-config" (OuterVolumeSpecName: "config") pod "2e9862f9-e69a-472f-9db9-638c95ee3912" (UID: "2e9862f9-e69a-472f-9db9-638c95ee3912"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:15:04 crc kubenswrapper[4789]: I1122 08:15:04.420722 4789 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2e9862f9-e69a-472f-9db9-638c95ee3912-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 22 08:15:04 crc kubenswrapper[4789]: I1122 08:15:04.420787 4789 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2e9862f9-e69a-472f-9db9-638c95ee3912-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 22 08:15:04 crc kubenswrapper[4789]: I1122 08:15:04.420802 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h5hkh\" (UniqueName: \"kubernetes.io/projected/2e9862f9-e69a-472f-9db9-638c95ee3912-kube-api-access-h5hkh\") on node \"crc\" DevicePath \"\"" Nov 22 08:15:04 crc kubenswrapper[4789]: I1122 08:15:04.420812 4789 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2e9862f9-e69a-472f-9db9-638c95ee3912-config\") on node \"crc\" DevicePath \"\"" Nov 22 08:15:04 crc kubenswrapper[4789]: I1122 08:15:04.486319 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-xhfrw"] Nov 22 08:15:04 crc kubenswrapper[4789]: W1122 08:15:04.490973 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podafee6663_d80e_41bc_af23_7d64a23d4b03.slice/crio-1d64ac414476e3ab051cfab249ebdfbbc852d987e6367ba307f5068368bdc726 WatchSource:0}: Error finding container 1d64ac414476e3ab051cfab249ebdfbbc852d987e6367ba307f5068368bdc726: Status 404 returned error can't find the container with id 1d64ac414476e3ab051cfab249ebdfbbc852d987e6367ba307f5068368bdc726 Nov 22 08:15:05 crc kubenswrapper[4789]: I1122 08:15:05.144907 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-xhfrw" event={"ID":"afee6663-d80e-41bc-af23-7d64a23d4b03","Type":"ContainerStarted","Data":"1d64ac414476e3ab051cfab249ebdfbbc852d987e6367ba307f5068368bdc726"} Nov 22 08:15:05 crc kubenswrapper[4789]: I1122 08:15:05.146629 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65c9b8d4f7-kfsdh" event={"ID":"2e9862f9-e69a-472f-9db9-638c95ee3912","Type":"ContainerDied","Data":"5c6eecd336e919c7bef4462ee52c51e3602fe0cc8c361f4b47e2e0b7fe6c6f79"} Nov 22 08:15:05 crc kubenswrapper[4789]: I1122 08:15:05.146658 4789 scope.go:117] "RemoveContainer" containerID="449fe80a53cf4a823fa98663c8779562ef32de7a57fd4ceac3350994ab0ee9d8" Nov 22 08:15:05 crc kubenswrapper[4789]: I1122 08:15:05.146702 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-65c9b8d4f7-kfsdh" Nov 22 08:15:05 crc kubenswrapper[4789]: I1122 08:15:05.166407 4789 scope.go:117] "RemoveContainer" containerID="e9e5f2aec8b3d285a6e9e963b44ca3fa3a8b87702afb2dd44b4d4baf6c4a3391" Nov 22 08:15:05 crc kubenswrapper[4789]: I1122 08:15:05.178289 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-65c9b8d4f7-kfsdh"] Nov 22 08:15:05 crc kubenswrapper[4789]: I1122 08:15:05.184143 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-65c9b8d4f7-kfsdh"] Nov 22 08:15:05 crc kubenswrapper[4789]: I1122 08:15:05.372537 4789 patch_prober.go:28] interesting pod/machine-config-daemon-pxvrm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 08:15:05 crc kubenswrapper[4789]: I1122 08:15:05.372594 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 08:15:05 crc kubenswrapper[4789]: I1122 08:15:05.981406 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e9862f9-e69a-472f-9db9-638c95ee3912" path="/var/lib/kubelet/pods/2e9862f9-e69a-472f-9db9-638c95ee3912/volumes" Nov 22 08:15:09 crc kubenswrapper[4789]: I1122 08:15:09.728839 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Nov 22 08:15:15 crc kubenswrapper[4789]: I1122 08:15:15.227820 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-xhfrw" event={"ID":"afee6663-d80e-41bc-af23-7d64a23d4b03","Type":"ContainerStarted","Data":"93ffeb8fcbfbb327e457f3753e9f8e777b7a951794760ed8c26507beb61b1a70"} Nov 22 08:15:15 crc kubenswrapper[4789]: I1122 08:15:15.248290 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-xhfrw" podStartSLOduration=2.181022234 podStartE2EDuration="12.248266416s" podCreationTimestamp="2025-11-22 08:15:03 +0000 UTC" firstStartedPulling="2025-11-22 08:15:04.493812582 +0000 UTC m=+1118.728212845" lastFinishedPulling="2025-11-22 08:15:14.561056754 +0000 UTC m=+1128.795457027" observedRunningTime="2025-11-22 08:15:15.246456476 +0000 UTC m=+1129.480856739" watchObservedRunningTime="2025-11-22 08:15:15.248266416 +0000 UTC m=+1129.482666699" Nov 22 08:15:16 crc kubenswrapper[4789]: I1122 08:15:16.241070 4789 generic.go:334] "Generic (PLEG): container finished" podID="df1ccc5e-c20c-40ce-bb76-17da02c4bc8c" containerID="b52c35865cafb51202b9dc7fd7ca36b2d1afa82daff2e035973c3e4405cd63b2" exitCode=0 Nov 22 08:15:16 crc kubenswrapper[4789]: I1122 08:15:16.241152 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"df1ccc5e-c20c-40ce-bb76-17da02c4bc8c","Type":"ContainerDied","Data":"b52c35865cafb51202b9dc7fd7ca36b2d1afa82daff2e035973c3e4405cd63b2"} Nov 22 08:15:16 crc kubenswrapper[4789]: I1122 08:15:16.243494 4789 generic.go:334] "Generic (PLEG): container finished" podID="af27dae6-c617-4d16-bb53-19057df7689b" containerID="e7ed5333a86c250ae63dba829dea4c481af2d77e17c98c355f0fc1115c16d37c" exitCode=0 Nov 22 08:15:16 crc kubenswrapper[4789]: I1122 08:15:16.244597 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"af27dae6-c617-4d16-bb53-19057df7689b","Type":"ContainerDied","Data":"e7ed5333a86c250ae63dba829dea4c481af2d77e17c98c355f0fc1115c16d37c"} Nov 22 08:15:17 crc kubenswrapper[4789]: I1122 08:15:17.253215 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"af27dae6-c617-4d16-bb53-19057df7689b","Type":"ContainerStarted","Data":"5d2637566554df256e507507e52737776868ccf6d6cf34ad3f5ce06caa06a1c8"} Nov 22 08:15:17 crc kubenswrapper[4789]: I1122 08:15:17.253487 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Nov 22 08:15:17 crc kubenswrapper[4789]: I1122 08:15:17.257170 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"df1ccc5e-c20c-40ce-bb76-17da02c4bc8c","Type":"ContainerStarted","Data":"2b2f9cf2f918a4e8b984cfe694a19fd982ad7500f88d820af5e9d08ce511ffe4"} Nov 22 08:15:17 crc kubenswrapper[4789]: I1122 08:15:17.257405 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:15:17 crc kubenswrapper[4789]: I1122 08:15:17.283325 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=38.238814344 podStartE2EDuration="1m4.283303318s" podCreationTimestamp="2025-11-22 08:14:13 +0000 UTC" firstStartedPulling="2025-11-22 08:14:15.84248647 +0000 UTC m=+1070.076886743" lastFinishedPulling="2025-11-22 08:14:41.886975434 +0000 UTC m=+1096.121375717" observedRunningTime="2025-11-22 08:15:17.280580883 +0000 UTC m=+1131.514981156" watchObservedRunningTime="2025-11-22 08:15:17.283303318 +0000 UTC m=+1131.517703591" Nov 22 08:15:17 crc kubenswrapper[4789]: I1122 08:15:17.307355 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=45.266730723 podStartE2EDuration="1m4.307333802s" podCreationTimestamp="2025-11-22 08:14:13 +0000 UTC" firstStartedPulling="2025-11-22 08:14:15.546039018 +0000 UTC m=+1069.780439291" lastFinishedPulling="2025-11-22 08:14:34.586642097 +0000 UTC m=+1088.821042370" observedRunningTime="2025-11-22 08:15:17.297697629 +0000 UTC m=+1131.532097912" watchObservedRunningTime="2025-11-22 08:15:17.307333802 +0000 UTC m=+1131.541734075" Nov 22 08:15:19 crc kubenswrapper[4789]: I1122 08:15:19.187427 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-cl8m4" podUID="e6f04e32-3dca-4d6d-8c8d-19ea15f46875" containerName="ovn-controller" probeResult="failure" output=< Nov 22 08:15:19 crc kubenswrapper[4789]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Nov 22 08:15:19 crc kubenswrapper[4789]: > Nov 22 08:15:19 crc kubenswrapper[4789]: I1122 08:15:19.201067 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-gjjwv" Nov 22 08:15:19 crc kubenswrapper[4789]: I1122 08:15:19.206220 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-gjjwv" Nov 22 08:15:19 crc kubenswrapper[4789]: I1122 08:15:19.437554 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-cl8m4-config-9zqs8"] Nov 22 08:15:19 crc kubenswrapper[4789]: E1122 08:15:19.438226 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e9862f9-e69a-472f-9db9-638c95ee3912" containerName="dnsmasq-dns" Nov 22 08:15:19 crc kubenswrapper[4789]: I1122 08:15:19.438367 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e9862f9-e69a-472f-9db9-638c95ee3912" containerName="dnsmasq-dns" Nov 22 08:15:19 crc kubenswrapper[4789]: E1122 08:15:19.438460 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e9862f9-e69a-472f-9db9-638c95ee3912" containerName="init" Nov 22 08:15:19 crc kubenswrapper[4789]: I1122 08:15:19.438531 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e9862f9-e69a-472f-9db9-638c95ee3912" containerName="init" Nov 22 08:15:19 crc kubenswrapper[4789]: I1122 08:15:19.440785 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e9862f9-e69a-472f-9db9-638c95ee3912" containerName="dnsmasq-dns" Nov 22 08:15:19 crc kubenswrapper[4789]: I1122 08:15:19.441994 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-cl8m4-config-9zqs8" Nov 22 08:15:19 crc kubenswrapper[4789]: I1122 08:15:19.456626 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Nov 22 08:15:19 crc kubenswrapper[4789]: I1122 08:15:19.480805 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-cl8m4-config-9zqs8"] Nov 22 08:15:19 crc kubenswrapper[4789]: I1122 08:15:19.569684 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/fb5af810-c839-40d7-a4d4-cb53a7fa464b-var-run\") pod \"ovn-controller-cl8m4-config-9zqs8\" (UID: \"fb5af810-c839-40d7-a4d4-cb53a7fa464b\") " pod="openstack/ovn-controller-cl8m4-config-9zqs8" Nov 22 08:15:19 crc kubenswrapper[4789]: I1122 08:15:19.569802 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/fb5af810-c839-40d7-a4d4-cb53a7fa464b-var-log-ovn\") pod \"ovn-controller-cl8m4-config-9zqs8\" (UID: \"fb5af810-c839-40d7-a4d4-cb53a7fa464b\") " pod="openstack/ovn-controller-cl8m4-config-9zqs8" Nov 22 08:15:19 crc kubenswrapper[4789]: I1122 08:15:19.569839 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/fb5af810-c839-40d7-a4d4-cb53a7fa464b-var-run-ovn\") pod \"ovn-controller-cl8m4-config-9zqs8\" (UID: \"fb5af810-c839-40d7-a4d4-cb53a7fa464b\") " pod="openstack/ovn-controller-cl8m4-config-9zqs8" Nov 22 08:15:19 crc kubenswrapper[4789]: I1122 08:15:19.570017 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-54wfq\" (UniqueName: \"kubernetes.io/projected/fb5af810-c839-40d7-a4d4-cb53a7fa464b-kube-api-access-54wfq\") pod \"ovn-controller-cl8m4-config-9zqs8\" (UID: \"fb5af810-c839-40d7-a4d4-cb53a7fa464b\") " pod="openstack/ovn-controller-cl8m4-config-9zqs8" Nov 22 08:15:19 crc kubenswrapper[4789]: I1122 08:15:19.570083 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fb5af810-c839-40d7-a4d4-cb53a7fa464b-scripts\") pod \"ovn-controller-cl8m4-config-9zqs8\" (UID: \"fb5af810-c839-40d7-a4d4-cb53a7fa464b\") " pod="openstack/ovn-controller-cl8m4-config-9zqs8" Nov 22 08:15:19 crc kubenswrapper[4789]: I1122 08:15:19.570133 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/fb5af810-c839-40d7-a4d4-cb53a7fa464b-additional-scripts\") pod \"ovn-controller-cl8m4-config-9zqs8\" (UID: \"fb5af810-c839-40d7-a4d4-cb53a7fa464b\") " pod="openstack/ovn-controller-cl8m4-config-9zqs8" Nov 22 08:15:19 crc kubenswrapper[4789]: I1122 08:15:19.671844 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fb5af810-c839-40d7-a4d4-cb53a7fa464b-scripts\") pod \"ovn-controller-cl8m4-config-9zqs8\" (UID: \"fb5af810-c839-40d7-a4d4-cb53a7fa464b\") " pod="openstack/ovn-controller-cl8m4-config-9zqs8" Nov 22 08:15:19 crc kubenswrapper[4789]: I1122 08:15:19.671920 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/fb5af810-c839-40d7-a4d4-cb53a7fa464b-additional-scripts\") pod \"ovn-controller-cl8m4-config-9zqs8\" (UID: \"fb5af810-c839-40d7-a4d4-cb53a7fa464b\") " pod="openstack/ovn-controller-cl8m4-config-9zqs8" Nov 22 08:15:19 crc kubenswrapper[4789]: I1122 08:15:19.672009 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/fb5af810-c839-40d7-a4d4-cb53a7fa464b-var-run\") pod \"ovn-controller-cl8m4-config-9zqs8\" (UID: \"fb5af810-c839-40d7-a4d4-cb53a7fa464b\") " pod="openstack/ovn-controller-cl8m4-config-9zqs8" Nov 22 08:15:19 crc kubenswrapper[4789]: I1122 08:15:19.672060 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/fb5af810-c839-40d7-a4d4-cb53a7fa464b-var-log-ovn\") pod \"ovn-controller-cl8m4-config-9zqs8\" (UID: \"fb5af810-c839-40d7-a4d4-cb53a7fa464b\") " pod="openstack/ovn-controller-cl8m4-config-9zqs8" Nov 22 08:15:19 crc kubenswrapper[4789]: I1122 08:15:19.672095 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/fb5af810-c839-40d7-a4d4-cb53a7fa464b-var-run-ovn\") pod \"ovn-controller-cl8m4-config-9zqs8\" (UID: \"fb5af810-c839-40d7-a4d4-cb53a7fa464b\") " pod="openstack/ovn-controller-cl8m4-config-9zqs8" Nov 22 08:15:19 crc kubenswrapper[4789]: I1122 08:15:19.672151 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-54wfq\" (UniqueName: \"kubernetes.io/projected/fb5af810-c839-40d7-a4d4-cb53a7fa464b-kube-api-access-54wfq\") pod \"ovn-controller-cl8m4-config-9zqs8\" (UID: \"fb5af810-c839-40d7-a4d4-cb53a7fa464b\") " pod="openstack/ovn-controller-cl8m4-config-9zqs8" Nov 22 08:15:19 crc kubenswrapper[4789]: I1122 08:15:19.672306 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/fb5af810-c839-40d7-a4d4-cb53a7fa464b-var-run\") pod \"ovn-controller-cl8m4-config-9zqs8\" (UID: \"fb5af810-c839-40d7-a4d4-cb53a7fa464b\") " pod="openstack/ovn-controller-cl8m4-config-9zqs8" Nov 22 08:15:19 crc kubenswrapper[4789]: I1122 08:15:19.672388 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/fb5af810-c839-40d7-a4d4-cb53a7fa464b-var-log-ovn\") pod \"ovn-controller-cl8m4-config-9zqs8\" (UID: \"fb5af810-c839-40d7-a4d4-cb53a7fa464b\") " pod="openstack/ovn-controller-cl8m4-config-9zqs8" Nov 22 08:15:19 crc kubenswrapper[4789]: I1122 08:15:19.672425 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/fb5af810-c839-40d7-a4d4-cb53a7fa464b-var-run-ovn\") pod \"ovn-controller-cl8m4-config-9zqs8\" (UID: \"fb5af810-c839-40d7-a4d4-cb53a7fa464b\") " pod="openstack/ovn-controller-cl8m4-config-9zqs8" Nov 22 08:15:19 crc kubenswrapper[4789]: I1122 08:15:19.672647 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/fb5af810-c839-40d7-a4d4-cb53a7fa464b-additional-scripts\") pod \"ovn-controller-cl8m4-config-9zqs8\" (UID: \"fb5af810-c839-40d7-a4d4-cb53a7fa464b\") " pod="openstack/ovn-controller-cl8m4-config-9zqs8" Nov 22 08:15:19 crc kubenswrapper[4789]: I1122 08:15:19.674021 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fb5af810-c839-40d7-a4d4-cb53a7fa464b-scripts\") pod \"ovn-controller-cl8m4-config-9zqs8\" (UID: \"fb5af810-c839-40d7-a4d4-cb53a7fa464b\") " pod="openstack/ovn-controller-cl8m4-config-9zqs8" Nov 22 08:15:19 crc kubenswrapper[4789]: I1122 08:15:19.711880 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-54wfq\" (UniqueName: \"kubernetes.io/projected/fb5af810-c839-40d7-a4d4-cb53a7fa464b-kube-api-access-54wfq\") pod \"ovn-controller-cl8m4-config-9zqs8\" (UID: \"fb5af810-c839-40d7-a4d4-cb53a7fa464b\") " pod="openstack/ovn-controller-cl8m4-config-9zqs8" Nov 22 08:15:19 crc kubenswrapper[4789]: I1122 08:15:19.766999 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-cl8m4-config-9zqs8" Nov 22 08:15:20 crc kubenswrapper[4789]: I1122 08:15:20.186036 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-cl8m4-config-9zqs8"] Nov 22 08:15:20 crc kubenswrapper[4789]: I1122 08:15:20.278793 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-cl8m4-config-9zqs8" event={"ID":"fb5af810-c839-40d7-a4d4-cb53a7fa464b","Type":"ContainerStarted","Data":"bf708353db0da252150a9d11c4523d8ac786c1374f1cf98fa2713c73f93ad177"} Nov 22 08:15:24 crc kubenswrapper[4789]: I1122 08:15:24.176580 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-cl8m4" podUID="e6f04e32-3dca-4d6d-8c8d-19ea15f46875" containerName="ovn-controller" probeResult="failure" output=< Nov 22 08:15:24 crc kubenswrapper[4789]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Nov 22 08:15:24 crc kubenswrapper[4789]: > Nov 22 08:15:29 crc kubenswrapper[4789]: I1122 08:15:29.181795 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-cl8m4" podUID="e6f04e32-3dca-4d6d-8c8d-19ea15f46875" containerName="ovn-controller" probeResult="failure" output=< Nov 22 08:15:29 crc kubenswrapper[4789]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Nov 22 08:15:29 crc kubenswrapper[4789]: > Nov 22 08:15:31 crc kubenswrapper[4789]: I1122 08:15:31.381568 4789 generic.go:334] "Generic (PLEG): container finished" podID="fb5af810-c839-40d7-a4d4-cb53a7fa464b" containerID="404f9e8585b31f57145fa7f61ae33da7eacd4bcd8fd95b769ea20c94366662cb" exitCode=0 Nov 22 08:15:31 crc kubenswrapper[4789]: I1122 08:15:31.381631 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-cl8m4-config-9zqs8" event={"ID":"fb5af810-c839-40d7-a4d4-cb53a7fa464b","Type":"ContainerDied","Data":"404f9e8585b31f57145fa7f61ae33da7eacd4bcd8fd95b769ea20c94366662cb"} Nov 22 08:15:32 crc kubenswrapper[4789]: I1122 08:15:32.647088 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-cl8m4-config-9zqs8" Nov 22 08:15:32 crc kubenswrapper[4789]: I1122 08:15:32.666593 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/fb5af810-c839-40d7-a4d4-cb53a7fa464b-var-run-ovn\") pod \"fb5af810-c839-40d7-a4d4-cb53a7fa464b\" (UID: \"fb5af810-c839-40d7-a4d4-cb53a7fa464b\") " Nov 22 08:15:32 crc kubenswrapper[4789]: I1122 08:15:32.666674 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-54wfq\" (UniqueName: \"kubernetes.io/projected/fb5af810-c839-40d7-a4d4-cb53a7fa464b-kube-api-access-54wfq\") pod \"fb5af810-c839-40d7-a4d4-cb53a7fa464b\" (UID: \"fb5af810-c839-40d7-a4d4-cb53a7fa464b\") " Nov 22 08:15:32 crc kubenswrapper[4789]: I1122 08:15:32.666765 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/fb5af810-c839-40d7-a4d4-cb53a7fa464b-additional-scripts\") pod \"fb5af810-c839-40d7-a4d4-cb53a7fa464b\" (UID: \"fb5af810-c839-40d7-a4d4-cb53a7fa464b\") " Nov 22 08:15:32 crc kubenswrapper[4789]: I1122 08:15:32.666725 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fb5af810-c839-40d7-a4d4-cb53a7fa464b-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "fb5af810-c839-40d7-a4d4-cb53a7fa464b" (UID: "fb5af810-c839-40d7-a4d4-cb53a7fa464b"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 08:15:32 crc kubenswrapper[4789]: I1122 08:15:32.666802 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/fb5af810-c839-40d7-a4d4-cb53a7fa464b-var-run\") pod \"fb5af810-c839-40d7-a4d4-cb53a7fa464b\" (UID: \"fb5af810-c839-40d7-a4d4-cb53a7fa464b\") " Nov 22 08:15:32 crc kubenswrapper[4789]: I1122 08:15:32.666980 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fb5af810-c839-40d7-a4d4-cb53a7fa464b-scripts\") pod \"fb5af810-c839-40d7-a4d4-cb53a7fa464b\" (UID: \"fb5af810-c839-40d7-a4d4-cb53a7fa464b\") " Nov 22 08:15:32 crc kubenswrapper[4789]: I1122 08:15:32.667009 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/fb5af810-c839-40d7-a4d4-cb53a7fa464b-var-log-ovn\") pod \"fb5af810-c839-40d7-a4d4-cb53a7fa464b\" (UID: \"fb5af810-c839-40d7-a4d4-cb53a7fa464b\") " Nov 22 08:15:32 crc kubenswrapper[4789]: I1122 08:15:32.667295 4789 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/fb5af810-c839-40d7-a4d4-cb53a7fa464b-var-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 22 08:15:32 crc kubenswrapper[4789]: I1122 08:15:32.667340 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fb5af810-c839-40d7-a4d4-cb53a7fa464b-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "fb5af810-c839-40d7-a4d4-cb53a7fa464b" (UID: "fb5af810-c839-40d7-a4d4-cb53a7fa464b"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 08:15:32 crc kubenswrapper[4789]: I1122 08:15:32.667367 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fb5af810-c839-40d7-a4d4-cb53a7fa464b-var-run" (OuterVolumeSpecName: "var-run") pod "fb5af810-c839-40d7-a4d4-cb53a7fa464b" (UID: "fb5af810-c839-40d7-a4d4-cb53a7fa464b"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 08:15:32 crc kubenswrapper[4789]: I1122 08:15:32.667706 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fb5af810-c839-40d7-a4d4-cb53a7fa464b-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "fb5af810-c839-40d7-a4d4-cb53a7fa464b" (UID: "fb5af810-c839-40d7-a4d4-cb53a7fa464b"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:15:32 crc kubenswrapper[4789]: I1122 08:15:32.668132 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fb5af810-c839-40d7-a4d4-cb53a7fa464b-scripts" (OuterVolumeSpecName: "scripts") pod "fb5af810-c839-40d7-a4d4-cb53a7fa464b" (UID: "fb5af810-c839-40d7-a4d4-cb53a7fa464b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:15:32 crc kubenswrapper[4789]: I1122 08:15:32.673111 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb5af810-c839-40d7-a4d4-cb53a7fa464b-kube-api-access-54wfq" (OuterVolumeSpecName: "kube-api-access-54wfq") pod "fb5af810-c839-40d7-a4d4-cb53a7fa464b" (UID: "fb5af810-c839-40d7-a4d4-cb53a7fa464b"). InnerVolumeSpecName "kube-api-access-54wfq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:15:32 crc kubenswrapper[4789]: I1122 08:15:32.768507 4789 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fb5af810-c839-40d7-a4d4-cb53a7fa464b-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:15:32 crc kubenswrapper[4789]: I1122 08:15:32.768545 4789 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/fb5af810-c839-40d7-a4d4-cb53a7fa464b-var-log-ovn\") on node \"crc\" DevicePath \"\"" Nov 22 08:15:32 crc kubenswrapper[4789]: I1122 08:15:32.768560 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-54wfq\" (UniqueName: \"kubernetes.io/projected/fb5af810-c839-40d7-a4d4-cb53a7fa464b-kube-api-access-54wfq\") on node \"crc\" DevicePath \"\"" Nov 22 08:15:32 crc kubenswrapper[4789]: I1122 08:15:32.768574 4789 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/fb5af810-c839-40d7-a4d4-cb53a7fa464b-additional-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:15:32 crc kubenswrapper[4789]: I1122 08:15:32.768585 4789 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/fb5af810-c839-40d7-a4d4-cb53a7fa464b-var-run\") on node \"crc\" DevicePath \"\"" Nov 22 08:15:33 crc kubenswrapper[4789]: I1122 08:15:33.400908 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-cl8m4-config-9zqs8" event={"ID":"fb5af810-c839-40d7-a4d4-cb53a7fa464b","Type":"ContainerDied","Data":"bf708353db0da252150a9d11c4523d8ac786c1374f1cf98fa2713c73f93ad177"} Nov 22 08:15:33 crc kubenswrapper[4789]: I1122 08:15:33.400967 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bf708353db0da252150a9d11c4523d8ac786c1374f1cf98fa2713c73f93ad177" Nov 22 08:15:33 crc kubenswrapper[4789]: I1122 08:15:33.401051 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-cl8m4-config-9zqs8" Nov 22 08:15:33 crc kubenswrapper[4789]: E1122 08:15:33.527037 4789 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfb5af810_c839_40d7_a4d4_cb53a7fa464b.slice/crio-bf708353db0da252150a9d11c4523d8ac786c1374f1cf98fa2713c73f93ad177\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfb5af810_c839_40d7_a4d4_cb53a7fa464b.slice\": RecentStats: unable to find data in memory cache]" Nov 22 08:15:33 crc kubenswrapper[4789]: I1122 08:15:33.751366 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-cl8m4-config-9zqs8"] Nov 22 08:15:33 crc kubenswrapper[4789]: I1122 08:15:33.757446 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-cl8m4-config-9zqs8"] Nov 22 08:15:33 crc kubenswrapper[4789]: I1122 08:15:33.974494 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fb5af810-c839-40d7-a4d4-cb53a7fa464b" path="/var/lib/kubelet/pods/fb5af810-c839-40d7-a4d4-cb53a7fa464b/volumes" Nov 22 08:15:34 crc kubenswrapper[4789]: I1122 08:15:34.195004 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-cl8m4" Nov 22 08:15:34 crc kubenswrapper[4789]: I1122 08:15:34.990932 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:15:35 crc kubenswrapper[4789]: I1122 08:15:35.230019 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Nov 22 08:15:35 crc kubenswrapper[4789]: I1122 08:15:35.371938 4789 patch_prober.go:28] interesting pod/machine-config-daemon-pxvrm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 08:15:35 crc kubenswrapper[4789]: I1122 08:15:35.372021 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 08:15:36 crc kubenswrapper[4789]: I1122 08:15:36.457868 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-nsptz"] Nov 22 08:15:36 crc kubenswrapper[4789]: E1122 08:15:36.458621 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb5af810-c839-40d7-a4d4-cb53a7fa464b" containerName="ovn-config" Nov 22 08:15:36 crc kubenswrapper[4789]: I1122 08:15:36.458637 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb5af810-c839-40d7-a4d4-cb53a7fa464b" containerName="ovn-config" Nov 22 08:15:36 crc kubenswrapper[4789]: I1122 08:15:36.458842 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb5af810-c839-40d7-a4d4-cb53a7fa464b" containerName="ovn-config" Nov 22 08:15:36 crc kubenswrapper[4789]: I1122 08:15:36.459437 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-nsptz" Nov 22 08:15:36 crc kubenswrapper[4789]: I1122 08:15:36.470537 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-nsptz"] Nov 22 08:15:36 crc kubenswrapper[4789]: I1122 08:15:36.547410 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-64ckc"] Nov 22 08:15:36 crc kubenswrapper[4789]: I1122 08:15:36.548602 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-64ckc" Nov 22 08:15:36 crc kubenswrapper[4789]: I1122 08:15:36.560473 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-64ckc"] Nov 22 08:15:36 crc kubenswrapper[4789]: I1122 08:15:36.568795 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-5a05-account-create-dcglb"] Nov 22 08:15:36 crc kubenswrapper[4789]: I1122 08:15:36.570060 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-5a05-account-create-dcglb" Nov 22 08:15:36 crc kubenswrapper[4789]: I1122 08:15:36.572960 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Nov 22 08:15:36 crc kubenswrapper[4789]: I1122 08:15:36.594102 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-5a05-account-create-dcglb"] Nov 22 08:15:36 crc kubenswrapper[4789]: I1122 08:15:36.622906 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0d6e8982-c10f-441a-be60-34888bc0acf2-operator-scripts\") pod \"cinder-db-create-nsptz\" (UID: \"0d6e8982-c10f-441a-be60-34888bc0acf2\") " pod="openstack/cinder-db-create-nsptz" Nov 22 08:15:36 crc kubenswrapper[4789]: I1122 08:15:36.623015 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-27cwh\" (UniqueName: \"kubernetes.io/projected/0d6e8982-c10f-441a-be60-34888bc0acf2-kube-api-access-27cwh\") pod \"cinder-db-create-nsptz\" (UID: \"0d6e8982-c10f-441a-be60-34888bc0acf2\") " pod="openstack/cinder-db-create-nsptz" Nov 22 08:15:36 crc kubenswrapper[4789]: I1122 08:15:36.663496 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-2375-account-create-chvgz"] Nov 22 08:15:36 crc kubenswrapper[4789]: I1122 08:15:36.664683 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-2375-account-create-chvgz" Nov 22 08:15:36 crc kubenswrapper[4789]: I1122 08:15:36.669167 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Nov 22 08:15:36 crc kubenswrapper[4789]: I1122 08:15:36.678943 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-2375-account-create-chvgz"] Nov 22 08:15:36 crc kubenswrapper[4789]: I1122 08:15:36.724679 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kgv85\" (UniqueName: \"kubernetes.io/projected/0c9e2b87-ef93-4176-9362-046e849033a5-kube-api-access-kgv85\") pod \"barbican-db-create-64ckc\" (UID: \"0c9e2b87-ef93-4176-9362-046e849033a5\") " pod="openstack/barbican-db-create-64ckc" Nov 22 08:15:36 crc kubenswrapper[4789]: I1122 08:15:36.724740 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0d6e8982-c10f-441a-be60-34888bc0acf2-operator-scripts\") pod \"cinder-db-create-nsptz\" (UID: \"0d6e8982-c10f-441a-be60-34888bc0acf2\") " pod="openstack/cinder-db-create-nsptz" Nov 22 08:15:36 crc kubenswrapper[4789]: I1122 08:15:36.724867 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nbj26\" (UniqueName: \"kubernetes.io/projected/56bd722c-eb14-4063-9ddb-9c010f461238-kube-api-access-nbj26\") pod \"barbican-5a05-account-create-dcglb\" (UID: \"56bd722c-eb14-4063-9ddb-9c010f461238\") " pod="openstack/barbican-5a05-account-create-dcglb" Nov 22 08:15:36 crc kubenswrapper[4789]: I1122 08:15:36.724898 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-27cwh\" (UniqueName: \"kubernetes.io/projected/0d6e8982-c10f-441a-be60-34888bc0acf2-kube-api-access-27cwh\") pod \"cinder-db-create-nsptz\" (UID: \"0d6e8982-c10f-441a-be60-34888bc0acf2\") " pod="openstack/cinder-db-create-nsptz" Nov 22 08:15:36 crc kubenswrapper[4789]: I1122 08:15:36.724921 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0c9e2b87-ef93-4176-9362-046e849033a5-operator-scripts\") pod \"barbican-db-create-64ckc\" (UID: \"0c9e2b87-ef93-4176-9362-046e849033a5\") " pod="openstack/barbican-db-create-64ckc" Nov 22 08:15:36 crc kubenswrapper[4789]: I1122 08:15:36.724939 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/56bd722c-eb14-4063-9ddb-9c010f461238-operator-scripts\") pod \"barbican-5a05-account-create-dcglb\" (UID: \"56bd722c-eb14-4063-9ddb-9c010f461238\") " pod="openstack/barbican-5a05-account-create-dcglb" Nov 22 08:15:36 crc kubenswrapper[4789]: I1122 08:15:36.725697 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0d6e8982-c10f-441a-be60-34888bc0acf2-operator-scripts\") pod \"cinder-db-create-nsptz\" (UID: \"0d6e8982-c10f-441a-be60-34888bc0acf2\") " pod="openstack/cinder-db-create-nsptz" Nov 22 08:15:36 crc kubenswrapper[4789]: I1122 08:15:36.750553 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-27cwh\" (UniqueName: \"kubernetes.io/projected/0d6e8982-c10f-441a-be60-34888bc0acf2-kube-api-access-27cwh\") pod \"cinder-db-create-nsptz\" (UID: \"0d6e8982-c10f-441a-be60-34888bc0acf2\") " pod="openstack/cinder-db-create-nsptz" Nov 22 08:15:36 crc kubenswrapper[4789]: I1122 08:15:36.785927 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-nsptz" Nov 22 08:15:36 crc kubenswrapper[4789]: I1122 08:15:36.826429 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jtjr4\" (UniqueName: \"kubernetes.io/projected/41a9281e-b846-4bd9-9c3a-085fe7d8e6f2-kube-api-access-jtjr4\") pod \"cinder-2375-account-create-chvgz\" (UID: \"41a9281e-b846-4bd9-9c3a-085fe7d8e6f2\") " pod="openstack/cinder-2375-account-create-chvgz" Nov 22 08:15:36 crc kubenswrapper[4789]: I1122 08:15:36.826522 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kgv85\" (UniqueName: \"kubernetes.io/projected/0c9e2b87-ef93-4176-9362-046e849033a5-kube-api-access-kgv85\") pod \"barbican-db-create-64ckc\" (UID: \"0c9e2b87-ef93-4176-9362-046e849033a5\") " pod="openstack/barbican-db-create-64ckc" Nov 22 08:15:36 crc kubenswrapper[4789]: I1122 08:15:36.826602 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nbj26\" (UniqueName: \"kubernetes.io/projected/56bd722c-eb14-4063-9ddb-9c010f461238-kube-api-access-nbj26\") pod \"barbican-5a05-account-create-dcglb\" (UID: \"56bd722c-eb14-4063-9ddb-9c010f461238\") " pod="openstack/barbican-5a05-account-create-dcglb" Nov 22 08:15:36 crc kubenswrapper[4789]: I1122 08:15:36.826706 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/41a9281e-b846-4bd9-9c3a-085fe7d8e6f2-operator-scripts\") pod \"cinder-2375-account-create-chvgz\" (UID: \"41a9281e-b846-4bd9-9c3a-085fe7d8e6f2\") " pod="openstack/cinder-2375-account-create-chvgz" Nov 22 08:15:36 crc kubenswrapper[4789]: I1122 08:15:36.826797 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0c9e2b87-ef93-4176-9362-046e849033a5-operator-scripts\") pod \"barbican-db-create-64ckc\" (UID: \"0c9e2b87-ef93-4176-9362-046e849033a5\") " pod="openstack/barbican-db-create-64ckc" Nov 22 08:15:36 crc kubenswrapper[4789]: I1122 08:15:36.826823 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/56bd722c-eb14-4063-9ddb-9c010f461238-operator-scripts\") pod \"barbican-5a05-account-create-dcglb\" (UID: \"56bd722c-eb14-4063-9ddb-9c010f461238\") " pod="openstack/barbican-5a05-account-create-dcglb" Nov 22 08:15:36 crc kubenswrapper[4789]: I1122 08:15:36.827671 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/56bd722c-eb14-4063-9ddb-9c010f461238-operator-scripts\") pod \"barbican-5a05-account-create-dcglb\" (UID: \"56bd722c-eb14-4063-9ddb-9c010f461238\") " pod="openstack/barbican-5a05-account-create-dcglb" Nov 22 08:15:36 crc kubenswrapper[4789]: I1122 08:15:36.827947 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0c9e2b87-ef93-4176-9362-046e849033a5-operator-scripts\") pod \"barbican-db-create-64ckc\" (UID: \"0c9e2b87-ef93-4176-9362-046e849033a5\") " pod="openstack/barbican-db-create-64ckc" Nov 22 08:15:36 crc kubenswrapper[4789]: I1122 08:15:36.853075 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kgv85\" (UniqueName: \"kubernetes.io/projected/0c9e2b87-ef93-4176-9362-046e849033a5-kube-api-access-kgv85\") pod \"barbican-db-create-64ckc\" (UID: \"0c9e2b87-ef93-4176-9362-046e849033a5\") " pod="openstack/barbican-db-create-64ckc" Nov 22 08:15:36 crc kubenswrapper[4789]: I1122 08:15:36.857402 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nbj26\" (UniqueName: \"kubernetes.io/projected/56bd722c-eb14-4063-9ddb-9c010f461238-kube-api-access-nbj26\") pod \"barbican-5a05-account-create-dcglb\" (UID: \"56bd722c-eb14-4063-9ddb-9c010f461238\") " pod="openstack/barbican-5a05-account-create-dcglb" Nov 22 08:15:36 crc kubenswrapper[4789]: I1122 08:15:36.861309 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-rnxhf"] Nov 22 08:15:36 crc kubenswrapper[4789]: I1122 08:15:36.862712 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-rnxhf" Nov 22 08:15:36 crc kubenswrapper[4789]: I1122 08:15:36.863411 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-64ckc" Nov 22 08:15:36 crc kubenswrapper[4789]: I1122 08:15:36.871171 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-411a-account-create-dh7ln"] Nov 22 08:15:36 crc kubenswrapper[4789]: I1122 08:15:36.872518 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-411a-account-create-dh7ln" Nov 22 08:15:36 crc kubenswrapper[4789]: I1122 08:15:36.874620 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Nov 22 08:15:36 crc kubenswrapper[4789]: I1122 08:15:36.882958 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-rnxhf"] Nov 22 08:15:36 crc kubenswrapper[4789]: I1122 08:15:36.888910 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-5a05-account-create-dcglb" Nov 22 08:15:36 crc kubenswrapper[4789]: I1122 08:15:36.893649 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-411a-account-create-dh7ln"] Nov 22 08:15:36 crc kubenswrapper[4789]: I1122 08:15:36.929329 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/41a9281e-b846-4bd9-9c3a-085fe7d8e6f2-operator-scripts\") pod \"cinder-2375-account-create-chvgz\" (UID: \"41a9281e-b846-4bd9-9c3a-085fe7d8e6f2\") " pod="openstack/cinder-2375-account-create-chvgz" Nov 22 08:15:36 crc kubenswrapper[4789]: I1122 08:15:36.929413 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jtjr4\" (UniqueName: \"kubernetes.io/projected/41a9281e-b846-4bd9-9c3a-085fe7d8e6f2-kube-api-access-jtjr4\") pod \"cinder-2375-account-create-chvgz\" (UID: \"41a9281e-b846-4bd9-9c3a-085fe7d8e6f2\") " pod="openstack/cinder-2375-account-create-chvgz" Nov 22 08:15:36 crc kubenswrapper[4789]: I1122 08:15:36.930371 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/41a9281e-b846-4bd9-9c3a-085fe7d8e6f2-operator-scripts\") pod \"cinder-2375-account-create-chvgz\" (UID: \"41a9281e-b846-4bd9-9c3a-085fe7d8e6f2\") " pod="openstack/cinder-2375-account-create-chvgz" Nov 22 08:15:36 crc kubenswrapper[4789]: I1122 08:15:36.946523 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-tx8g5"] Nov 22 08:15:36 crc kubenswrapper[4789]: I1122 08:15:36.951771 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-tx8g5" Nov 22 08:15:36 crc kubenswrapper[4789]: I1122 08:15:36.954920 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 22 08:15:36 crc kubenswrapper[4789]: I1122 08:15:36.955138 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 22 08:15:36 crc kubenswrapper[4789]: I1122 08:15:36.955394 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-wnnts" Nov 22 08:15:36 crc kubenswrapper[4789]: I1122 08:15:36.956912 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 22 08:15:36 crc kubenswrapper[4789]: I1122 08:15:36.958044 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-tx8g5"] Nov 22 08:15:36 crc kubenswrapper[4789]: I1122 08:15:36.961629 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jtjr4\" (UniqueName: \"kubernetes.io/projected/41a9281e-b846-4bd9-9c3a-085fe7d8e6f2-kube-api-access-jtjr4\") pod \"cinder-2375-account-create-chvgz\" (UID: \"41a9281e-b846-4bd9-9c3a-085fe7d8e6f2\") " pod="openstack/cinder-2375-account-create-chvgz" Nov 22 08:15:36 crc kubenswrapper[4789]: I1122 08:15:36.985146 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-2375-account-create-chvgz" Nov 22 08:15:37 crc kubenswrapper[4789]: I1122 08:15:37.030864 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mvdm2\" (UniqueName: \"kubernetes.io/projected/201449e6-debb-4b67-80c6-524b309dac1f-kube-api-access-mvdm2\") pod \"neutron-411a-account-create-dh7ln\" (UID: \"201449e6-debb-4b67-80c6-524b309dac1f\") " pod="openstack/neutron-411a-account-create-dh7ln" Nov 22 08:15:37 crc kubenswrapper[4789]: I1122 08:15:37.030922 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b56702e5-c634-4162-9aab-75c723993be4-operator-scripts\") pod \"neutron-db-create-rnxhf\" (UID: \"b56702e5-c634-4162-9aab-75c723993be4\") " pod="openstack/neutron-db-create-rnxhf" Nov 22 08:15:37 crc kubenswrapper[4789]: I1122 08:15:37.030983 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/201449e6-debb-4b67-80c6-524b309dac1f-operator-scripts\") pod \"neutron-411a-account-create-dh7ln\" (UID: \"201449e6-debb-4b67-80c6-524b309dac1f\") " pod="openstack/neutron-411a-account-create-dh7ln" Nov 22 08:15:37 crc kubenswrapper[4789]: I1122 08:15:37.031023 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-64lp4\" (UniqueName: \"kubernetes.io/projected/b56702e5-c634-4162-9aab-75c723993be4-kube-api-access-64lp4\") pod \"neutron-db-create-rnxhf\" (UID: \"b56702e5-c634-4162-9aab-75c723993be4\") " pod="openstack/neutron-db-create-rnxhf" Nov 22 08:15:37 crc kubenswrapper[4789]: I1122 08:15:37.132913 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c80081f-1494-4fd4-ba90-da18af310075-config-data\") pod \"keystone-db-sync-tx8g5\" (UID: \"4c80081f-1494-4fd4-ba90-da18af310075\") " pod="openstack/keystone-db-sync-tx8g5" Nov 22 08:15:37 crc kubenswrapper[4789]: I1122 08:15:37.132991 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mvdm2\" (UniqueName: \"kubernetes.io/projected/201449e6-debb-4b67-80c6-524b309dac1f-kube-api-access-mvdm2\") pod \"neutron-411a-account-create-dh7ln\" (UID: \"201449e6-debb-4b67-80c6-524b309dac1f\") " pod="openstack/neutron-411a-account-create-dh7ln" Nov 22 08:15:37 crc kubenswrapper[4789]: I1122 08:15:37.133026 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b56702e5-c634-4162-9aab-75c723993be4-operator-scripts\") pod \"neutron-db-create-rnxhf\" (UID: \"b56702e5-c634-4162-9aab-75c723993be4\") " pod="openstack/neutron-db-create-rnxhf" Nov 22 08:15:37 crc kubenswrapper[4789]: I1122 08:15:37.133084 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/201449e6-debb-4b67-80c6-524b309dac1f-operator-scripts\") pod \"neutron-411a-account-create-dh7ln\" (UID: \"201449e6-debb-4b67-80c6-524b309dac1f\") " pod="openstack/neutron-411a-account-create-dh7ln" Nov 22 08:15:37 crc kubenswrapper[4789]: I1122 08:15:37.133111 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c80081f-1494-4fd4-ba90-da18af310075-combined-ca-bundle\") pod \"keystone-db-sync-tx8g5\" (UID: \"4c80081f-1494-4fd4-ba90-da18af310075\") " pod="openstack/keystone-db-sync-tx8g5" Nov 22 08:15:37 crc kubenswrapper[4789]: I1122 08:15:37.133150 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-64lp4\" (UniqueName: \"kubernetes.io/projected/b56702e5-c634-4162-9aab-75c723993be4-kube-api-access-64lp4\") pod \"neutron-db-create-rnxhf\" (UID: \"b56702e5-c634-4162-9aab-75c723993be4\") " pod="openstack/neutron-db-create-rnxhf" Nov 22 08:15:37 crc kubenswrapper[4789]: I1122 08:15:37.134470 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b56702e5-c634-4162-9aab-75c723993be4-operator-scripts\") pod \"neutron-db-create-rnxhf\" (UID: \"b56702e5-c634-4162-9aab-75c723993be4\") " pod="openstack/neutron-db-create-rnxhf" Nov 22 08:15:37 crc kubenswrapper[4789]: I1122 08:15:37.134538 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/201449e6-debb-4b67-80c6-524b309dac1f-operator-scripts\") pod \"neutron-411a-account-create-dh7ln\" (UID: \"201449e6-debb-4b67-80c6-524b309dac1f\") " pod="openstack/neutron-411a-account-create-dh7ln" Nov 22 08:15:37 crc kubenswrapper[4789]: I1122 08:15:37.134702 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p47bc\" (UniqueName: \"kubernetes.io/projected/4c80081f-1494-4fd4-ba90-da18af310075-kube-api-access-p47bc\") pod \"keystone-db-sync-tx8g5\" (UID: \"4c80081f-1494-4fd4-ba90-da18af310075\") " pod="openstack/keystone-db-sync-tx8g5" Nov 22 08:15:37 crc kubenswrapper[4789]: I1122 08:15:37.156177 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-64lp4\" (UniqueName: \"kubernetes.io/projected/b56702e5-c634-4162-9aab-75c723993be4-kube-api-access-64lp4\") pod \"neutron-db-create-rnxhf\" (UID: \"b56702e5-c634-4162-9aab-75c723993be4\") " pod="openstack/neutron-db-create-rnxhf" Nov 22 08:15:37 crc kubenswrapper[4789]: I1122 08:15:37.160535 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mvdm2\" (UniqueName: \"kubernetes.io/projected/201449e6-debb-4b67-80c6-524b309dac1f-kube-api-access-mvdm2\") pod \"neutron-411a-account-create-dh7ln\" (UID: \"201449e6-debb-4b67-80c6-524b309dac1f\") " pod="openstack/neutron-411a-account-create-dh7ln" Nov 22 08:15:37 crc kubenswrapper[4789]: I1122 08:15:37.178232 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-nsptz"] Nov 22 08:15:37 crc kubenswrapper[4789]: I1122 08:15:37.236874 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c80081f-1494-4fd4-ba90-da18af310075-combined-ca-bundle\") pod \"keystone-db-sync-tx8g5\" (UID: \"4c80081f-1494-4fd4-ba90-da18af310075\") " pod="openstack/keystone-db-sync-tx8g5" Nov 22 08:15:37 crc kubenswrapper[4789]: I1122 08:15:37.236963 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p47bc\" (UniqueName: \"kubernetes.io/projected/4c80081f-1494-4fd4-ba90-da18af310075-kube-api-access-p47bc\") pod \"keystone-db-sync-tx8g5\" (UID: \"4c80081f-1494-4fd4-ba90-da18af310075\") " pod="openstack/keystone-db-sync-tx8g5" Nov 22 08:15:37 crc kubenswrapper[4789]: I1122 08:15:37.237050 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c80081f-1494-4fd4-ba90-da18af310075-config-data\") pod \"keystone-db-sync-tx8g5\" (UID: \"4c80081f-1494-4fd4-ba90-da18af310075\") " pod="openstack/keystone-db-sync-tx8g5" Nov 22 08:15:37 crc kubenswrapper[4789]: I1122 08:15:37.241452 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c80081f-1494-4fd4-ba90-da18af310075-combined-ca-bundle\") pod \"keystone-db-sync-tx8g5\" (UID: \"4c80081f-1494-4fd4-ba90-da18af310075\") " pod="openstack/keystone-db-sync-tx8g5" Nov 22 08:15:37 crc kubenswrapper[4789]: I1122 08:15:37.242087 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c80081f-1494-4fd4-ba90-da18af310075-config-data\") pod \"keystone-db-sync-tx8g5\" (UID: \"4c80081f-1494-4fd4-ba90-da18af310075\") " pod="openstack/keystone-db-sync-tx8g5" Nov 22 08:15:37 crc kubenswrapper[4789]: I1122 08:15:37.256659 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p47bc\" (UniqueName: \"kubernetes.io/projected/4c80081f-1494-4fd4-ba90-da18af310075-kube-api-access-p47bc\") pod \"keystone-db-sync-tx8g5\" (UID: \"4c80081f-1494-4fd4-ba90-da18af310075\") " pod="openstack/keystone-db-sync-tx8g5" Nov 22 08:15:37 crc kubenswrapper[4789]: I1122 08:15:37.274963 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-rnxhf" Nov 22 08:15:37 crc kubenswrapper[4789]: I1122 08:15:37.287419 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-411a-account-create-dh7ln" Nov 22 08:15:37 crc kubenswrapper[4789]: I1122 08:15:37.296426 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-2375-account-create-chvgz"] Nov 22 08:15:37 crc kubenswrapper[4789]: I1122 08:15:37.297648 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-tx8g5" Nov 22 08:15:37 crc kubenswrapper[4789]: I1122 08:15:37.459193 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-5a05-account-create-dcglb"] Nov 22 08:15:37 crc kubenswrapper[4789]: I1122 08:15:37.461867 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-nsptz" event={"ID":"0d6e8982-c10f-441a-be60-34888bc0acf2","Type":"ContainerStarted","Data":"b1483ede4e565a0a93afedd8132895003b4f2a08d1541b3d3df1683ab24a7407"} Nov 22 08:15:37 crc kubenswrapper[4789]: I1122 08:15:37.461908 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-nsptz" event={"ID":"0d6e8982-c10f-441a-be60-34888bc0acf2","Type":"ContainerStarted","Data":"f36aec206d0d79323906928bb7454a8e7498aea41a96791b535964f97222e300"} Nov 22 08:15:37 crc kubenswrapper[4789]: I1122 08:15:37.463274 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-2375-account-create-chvgz" event={"ID":"41a9281e-b846-4bd9-9c3a-085fe7d8e6f2","Type":"ContainerStarted","Data":"0f9c5d457c61c04370148a8fabdbd567860e323257c44bff250797b3586f1c20"} Nov 22 08:15:37 crc kubenswrapper[4789]: I1122 08:15:37.475079 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-64ckc"] Nov 22 08:15:37 crc kubenswrapper[4789]: I1122 08:15:37.480973 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-create-nsptz" podStartSLOduration=1.480953812 podStartE2EDuration="1.480953812s" podCreationTimestamp="2025-11-22 08:15:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:15:37.478830663 +0000 UTC m=+1151.713230946" watchObservedRunningTime="2025-11-22 08:15:37.480953812 +0000 UTC m=+1151.715354085" Nov 22 08:15:37 crc kubenswrapper[4789]: I1122 08:15:37.733294 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-tx8g5"] Nov 22 08:15:37 crc kubenswrapper[4789]: I1122 08:15:37.806699 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-rnxhf"] Nov 22 08:15:37 crc kubenswrapper[4789]: W1122 08:15:37.857180 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb56702e5_c634_4162_9aab_75c723993be4.slice/crio-3a636d03abc106705a3d25820b0309a22439288896e5142a372b09f44fc2e660 WatchSource:0}: Error finding container 3a636d03abc106705a3d25820b0309a22439288896e5142a372b09f44fc2e660: Status 404 returned error can't find the container with id 3a636d03abc106705a3d25820b0309a22439288896e5142a372b09f44fc2e660 Nov 22 08:15:37 crc kubenswrapper[4789]: I1122 08:15:37.861094 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-411a-account-create-dh7ln"] Nov 22 08:15:37 crc kubenswrapper[4789]: W1122 08:15:37.865421 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod201449e6_debb_4b67_80c6_524b309dac1f.slice/crio-4a6d4013035f81ae9b3817796a9360b081d5c119cf3a535ec59efe0fd12a5e7e WatchSource:0}: Error finding container 4a6d4013035f81ae9b3817796a9360b081d5c119cf3a535ec59efe0fd12a5e7e: Status 404 returned error can't find the container with id 4a6d4013035f81ae9b3817796a9360b081d5c119cf3a535ec59efe0fd12a5e7e Nov 22 08:15:38 crc kubenswrapper[4789]: I1122 08:15:38.474057 4789 generic.go:334] "Generic (PLEG): container finished" podID="41a9281e-b846-4bd9-9c3a-085fe7d8e6f2" containerID="1dfe678ead9e68e4225b9febfd4b0a95e66b0d0f04f4d2e710036df1d295f976" exitCode=0 Nov 22 08:15:38 crc kubenswrapper[4789]: I1122 08:15:38.474109 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-2375-account-create-chvgz" event={"ID":"41a9281e-b846-4bd9-9c3a-085fe7d8e6f2","Type":"ContainerDied","Data":"1dfe678ead9e68e4225b9febfd4b0a95e66b0d0f04f4d2e710036df1d295f976"} Nov 22 08:15:38 crc kubenswrapper[4789]: I1122 08:15:38.476107 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-rnxhf" event={"ID":"b56702e5-c634-4162-9aab-75c723993be4","Type":"ContainerStarted","Data":"aac635889d3cbe5eeec4bde45f41b0bd395e0ca68c5bb3cff673bbf27978bee3"} Nov 22 08:15:38 crc kubenswrapper[4789]: I1122 08:15:38.476139 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-rnxhf" event={"ID":"b56702e5-c634-4162-9aab-75c723993be4","Type":"ContainerStarted","Data":"3a636d03abc106705a3d25820b0309a22439288896e5142a372b09f44fc2e660"} Nov 22 08:15:38 crc kubenswrapper[4789]: I1122 08:15:38.477597 4789 generic.go:334] "Generic (PLEG): container finished" podID="0c9e2b87-ef93-4176-9362-046e849033a5" containerID="5c5f0e394886a566012dcde583f9244b73eb7e8f80bc29c6ec5ea672207f15bc" exitCode=0 Nov 22 08:15:38 crc kubenswrapper[4789]: I1122 08:15:38.477667 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-64ckc" event={"ID":"0c9e2b87-ef93-4176-9362-046e849033a5","Type":"ContainerDied","Data":"5c5f0e394886a566012dcde583f9244b73eb7e8f80bc29c6ec5ea672207f15bc"} Nov 22 08:15:38 crc kubenswrapper[4789]: I1122 08:15:38.477694 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-64ckc" event={"ID":"0c9e2b87-ef93-4176-9362-046e849033a5","Type":"ContainerStarted","Data":"2017a2f5fa6bff574c4e3e4ee9678cc52eedbbb335845664ef9734a2b25ccc4c"} Nov 22 08:15:38 crc kubenswrapper[4789]: I1122 08:15:38.479675 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-tx8g5" event={"ID":"4c80081f-1494-4fd4-ba90-da18af310075","Type":"ContainerStarted","Data":"98f965b1787cf989365213bef3e4f62fa5f12da534bbec60cd171697b6c98848"} Nov 22 08:15:38 crc kubenswrapper[4789]: I1122 08:15:38.481167 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-411a-account-create-dh7ln" event={"ID":"201449e6-debb-4b67-80c6-524b309dac1f","Type":"ContainerStarted","Data":"1cb148dd74abf9675a681dd2ec67770bd47fe31a37cdf5da5b1fc713bb4ffebc"} Nov 22 08:15:38 crc kubenswrapper[4789]: I1122 08:15:38.481206 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-411a-account-create-dh7ln" event={"ID":"201449e6-debb-4b67-80c6-524b309dac1f","Type":"ContainerStarted","Data":"4a6d4013035f81ae9b3817796a9360b081d5c119cf3a535ec59efe0fd12a5e7e"} Nov 22 08:15:38 crc kubenswrapper[4789]: I1122 08:15:38.482267 4789 generic.go:334] "Generic (PLEG): container finished" podID="56bd722c-eb14-4063-9ddb-9c010f461238" containerID="165aaecc3c7daba27ab8232bc51f5d633e1f9d811f9fe7aab330b0a5a7ed65a2" exitCode=0 Nov 22 08:15:38 crc kubenswrapper[4789]: I1122 08:15:38.482321 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-5a05-account-create-dcglb" event={"ID":"56bd722c-eb14-4063-9ddb-9c010f461238","Type":"ContainerDied","Data":"165aaecc3c7daba27ab8232bc51f5d633e1f9d811f9fe7aab330b0a5a7ed65a2"} Nov 22 08:15:38 crc kubenswrapper[4789]: I1122 08:15:38.482339 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-5a05-account-create-dcglb" event={"ID":"56bd722c-eb14-4063-9ddb-9c010f461238","Type":"ContainerStarted","Data":"ce2102086904fda084b67280ebafdb5a6e5d44cc7a118e3ee8345041ae637eed"} Nov 22 08:15:38 crc kubenswrapper[4789]: I1122 08:15:38.485383 4789 generic.go:334] "Generic (PLEG): container finished" podID="0d6e8982-c10f-441a-be60-34888bc0acf2" containerID="b1483ede4e565a0a93afedd8132895003b4f2a08d1541b3d3df1683ab24a7407" exitCode=0 Nov 22 08:15:38 crc kubenswrapper[4789]: I1122 08:15:38.485436 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-nsptz" event={"ID":"0d6e8982-c10f-441a-be60-34888bc0acf2","Type":"ContainerDied","Data":"b1483ede4e565a0a93afedd8132895003b4f2a08d1541b3d3df1683ab24a7407"} Nov 22 08:15:38 crc kubenswrapper[4789]: I1122 08:15:38.546977 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-411a-account-create-dh7ln" podStartSLOduration=2.546955564 podStartE2EDuration="2.546955564s" podCreationTimestamp="2025-11-22 08:15:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:15:38.536872819 +0000 UTC m=+1152.771273102" watchObservedRunningTime="2025-11-22 08:15:38.546955564 +0000 UTC m=+1152.781355837" Nov 22 08:15:38 crc kubenswrapper[4789]: I1122 08:15:38.578838 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-create-rnxhf" podStartSLOduration=2.578822592 podStartE2EDuration="2.578822592s" podCreationTimestamp="2025-11-22 08:15:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:15:38.576362945 +0000 UTC m=+1152.810763218" watchObservedRunningTime="2025-11-22 08:15:38.578822592 +0000 UTC m=+1152.813222865" Nov 22 08:15:39 crc kubenswrapper[4789]: I1122 08:15:39.497668 4789 generic.go:334] "Generic (PLEG): container finished" podID="afee6663-d80e-41bc-af23-7d64a23d4b03" containerID="93ffeb8fcbfbb327e457f3753e9f8e777b7a951794760ed8c26507beb61b1a70" exitCode=0 Nov 22 08:15:39 crc kubenswrapper[4789]: I1122 08:15:39.497767 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-xhfrw" event={"ID":"afee6663-d80e-41bc-af23-7d64a23d4b03","Type":"ContainerDied","Data":"93ffeb8fcbfbb327e457f3753e9f8e777b7a951794760ed8c26507beb61b1a70"} Nov 22 08:15:39 crc kubenswrapper[4789]: I1122 08:15:39.500156 4789 generic.go:334] "Generic (PLEG): container finished" podID="201449e6-debb-4b67-80c6-524b309dac1f" containerID="1cb148dd74abf9675a681dd2ec67770bd47fe31a37cdf5da5b1fc713bb4ffebc" exitCode=0 Nov 22 08:15:39 crc kubenswrapper[4789]: I1122 08:15:39.500250 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-411a-account-create-dh7ln" event={"ID":"201449e6-debb-4b67-80c6-524b309dac1f","Type":"ContainerDied","Data":"1cb148dd74abf9675a681dd2ec67770bd47fe31a37cdf5da5b1fc713bb4ffebc"} Nov 22 08:15:39 crc kubenswrapper[4789]: I1122 08:15:39.502036 4789 generic.go:334] "Generic (PLEG): container finished" podID="b56702e5-c634-4162-9aab-75c723993be4" containerID="aac635889d3cbe5eeec4bde45f41b0bd395e0ca68c5bb3cff673bbf27978bee3" exitCode=0 Nov 22 08:15:39 crc kubenswrapper[4789]: I1122 08:15:39.502132 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-rnxhf" event={"ID":"b56702e5-c634-4162-9aab-75c723993be4","Type":"ContainerDied","Data":"aac635889d3cbe5eeec4bde45f41b0bd395e0ca68c5bb3cff673bbf27978bee3"} Nov 22 08:15:42 crc kubenswrapper[4789]: I1122 08:15:42.004477 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-2375-account-create-chvgz" Nov 22 08:15:42 crc kubenswrapper[4789]: I1122 08:15:42.010617 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-rnxhf" Nov 22 08:15:42 crc kubenswrapper[4789]: I1122 08:15:42.021065 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-64ckc" Nov 22 08:15:42 crc kubenswrapper[4789]: I1122 08:15:42.042978 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-411a-account-create-dh7ln" Nov 22 08:15:42 crc kubenswrapper[4789]: I1122 08:15:42.049089 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-5a05-account-create-dcglb" Nov 22 08:15:42 crc kubenswrapper[4789]: I1122 08:15:42.056565 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-nsptz" Nov 22 08:15:42 crc kubenswrapper[4789]: I1122 08:15:42.063727 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-xhfrw" Nov 22 08:15:42 crc kubenswrapper[4789]: I1122 08:15:42.131456 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/41a9281e-b846-4bd9-9c3a-085fe7d8e6f2-operator-scripts\") pod \"41a9281e-b846-4bd9-9c3a-085fe7d8e6f2\" (UID: \"41a9281e-b846-4bd9-9c3a-085fe7d8e6f2\") " Nov 22 08:15:42 crc kubenswrapper[4789]: I1122 08:15:42.131806 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jtjr4\" (UniqueName: \"kubernetes.io/projected/41a9281e-b846-4bd9-9c3a-085fe7d8e6f2-kube-api-access-jtjr4\") pod \"41a9281e-b846-4bd9-9c3a-085fe7d8e6f2\" (UID: \"41a9281e-b846-4bd9-9c3a-085fe7d8e6f2\") " Nov 22 08:15:42 crc kubenswrapper[4789]: I1122 08:15:42.131910 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/201449e6-debb-4b67-80c6-524b309dac1f-operator-scripts\") pod \"201449e6-debb-4b67-80c6-524b309dac1f\" (UID: \"201449e6-debb-4b67-80c6-524b309dac1f\") " Nov 22 08:15:42 crc kubenswrapper[4789]: I1122 08:15:42.132001 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/56bd722c-eb14-4063-9ddb-9c010f461238-operator-scripts\") pod \"56bd722c-eb14-4063-9ddb-9c010f461238\" (UID: \"56bd722c-eb14-4063-9ddb-9c010f461238\") " Nov 22 08:15:42 crc kubenswrapper[4789]: I1122 08:15:42.132101 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0c9e2b87-ef93-4176-9362-046e849033a5-operator-scripts\") pod \"0c9e2b87-ef93-4176-9362-046e849033a5\" (UID: \"0c9e2b87-ef93-4176-9362-046e849033a5\") " Nov 22 08:15:42 crc kubenswrapper[4789]: I1122 08:15:42.132206 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kgv85\" (UniqueName: \"kubernetes.io/projected/0c9e2b87-ef93-4176-9362-046e849033a5-kube-api-access-kgv85\") pod \"0c9e2b87-ef93-4176-9362-046e849033a5\" (UID: \"0c9e2b87-ef93-4176-9362-046e849033a5\") " Nov 22 08:15:42 crc kubenswrapper[4789]: I1122 08:15:42.132340 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-64lp4\" (UniqueName: \"kubernetes.io/projected/b56702e5-c634-4162-9aab-75c723993be4-kube-api-access-64lp4\") pod \"b56702e5-c634-4162-9aab-75c723993be4\" (UID: \"b56702e5-c634-4162-9aab-75c723993be4\") " Nov 22 08:15:42 crc kubenswrapper[4789]: I1122 08:15:42.132425 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b56702e5-c634-4162-9aab-75c723993be4-operator-scripts\") pod \"b56702e5-c634-4162-9aab-75c723993be4\" (UID: \"b56702e5-c634-4162-9aab-75c723993be4\") " Nov 22 08:15:42 crc kubenswrapper[4789]: I1122 08:15:42.132511 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nbj26\" (UniqueName: \"kubernetes.io/projected/56bd722c-eb14-4063-9ddb-9c010f461238-kube-api-access-nbj26\") pod \"56bd722c-eb14-4063-9ddb-9c010f461238\" (UID: \"56bd722c-eb14-4063-9ddb-9c010f461238\") " Nov 22 08:15:42 crc kubenswrapper[4789]: I1122 08:15:42.132589 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mvdm2\" (UniqueName: \"kubernetes.io/projected/201449e6-debb-4b67-80c6-524b309dac1f-kube-api-access-mvdm2\") pod \"201449e6-debb-4b67-80c6-524b309dac1f\" (UID: \"201449e6-debb-4b67-80c6-524b309dac1f\") " Nov 22 08:15:42 crc kubenswrapper[4789]: I1122 08:15:42.132637 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/41a9281e-b846-4bd9-9c3a-085fe7d8e6f2-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "41a9281e-b846-4bd9-9c3a-085fe7d8e6f2" (UID: "41a9281e-b846-4bd9-9c3a-085fe7d8e6f2"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:15:42 crc kubenswrapper[4789]: I1122 08:15:42.132858 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b56702e5-c634-4162-9aab-75c723993be4-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b56702e5-c634-4162-9aab-75c723993be4" (UID: "b56702e5-c634-4162-9aab-75c723993be4"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:15:42 crc kubenswrapper[4789]: I1122 08:15:42.132875 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/56bd722c-eb14-4063-9ddb-9c010f461238-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "56bd722c-eb14-4063-9ddb-9c010f461238" (UID: "56bd722c-eb14-4063-9ddb-9c010f461238"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:15:42 crc kubenswrapper[4789]: I1122 08:15:42.133233 4789 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/56bd722c-eb14-4063-9ddb-9c010f461238-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:15:42 crc kubenswrapper[4789]: I1122 08:15:42.133307 4789 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b56702e5-c634-4162-9aab-75c723993be4-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:15:42 crc kubenswrapper[4789]: I1122 08:15:42.133373 4789 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/41a9281e-b846-4bd9-9c3a-085fe7d8e6f2-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:15:42 crc kubenswrapper[4789]: I1122 08:15:42.133323 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0c9e2b87-ef93-4176-9362-046e849033a5-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0c9e2b87-ef93-4176-9362-046e849033a5" (UID: "0c9e2b87-ef93-4176-9362-046e849033a5"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:15:42 crc kubenswrapper[4789]: I1122 08:15:42.133390 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/201449e6-debb-4b67-80c6-524b309dac1f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "201449e6-debb-4b67-80c6-524b309dac1f" (UID: "201449e6-debb-4b67-80c6-524b309dac1f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:15:42 crc kubenswrapper[4789]: I1122 08:15:42.137931 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/41a9281e-b846-4bd9-9c3a-085fe7d8e6f2-kube-api-access-jtjr4" (OuterVolumeSpecName: "kube-api-access-jtjr4") pod "41a9281e-b846-4bd9-9c3a-085fe7d8e6f2" (UID: "41a9281e-b846-4bd9-9c3a-085fe7d8e6f2"). InnerVolumeSpecName "kube-api-access-jtjr4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:15:42 crc kubenswrapper[4789]: I1122 08:15:42.141469 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56bd722c-eb14-4063-9ddb-9c010f461238-kube-api-access-nbj26" (OuterVolumeSpecName: "kube-api-access-nbj26") pod "56bd722c-eb14-4063-9ddb-9c010f461238" (UID: "56bd722c-eb14-4063-9ddb-9c010f461238"). InnerVolumeSpecName "kube-api-access-nbj26". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:15:42 crc kubenswrapper[4789]: I1122 08:15:42.141553 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c9e2b87-ef93-4176-9362-046e849033a5-kube-api-access-kgv85" (OuterVolumeSpecName: "kube-api-access-kgv85") pod "0c9e2b87-ef93-4176-9362-046e849033a5" (UID: "0c9e2b87-ef93-4176-9362-046e849033a5"). InnerVolumeSpecName "kube-api-access-kgv85". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:15:42 crc kubenswrapper[4789]: I1122 08:15:42.141629 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b56702e5-c634-4162-9aab-75c723993be4-kube-api-access-64lp4" (OuterVolumeSpecName: "kube-api-access-64lp4") pod "b56702e5-c634-4162-9aab-75c723993be4" (UID: "b56702e5-c634-4162-9aab-75c723993be4"). InnerVolumeSpecName "kube-api-access-64lp4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:15:42 crc kubenswrapper[4789]: I1122 08:15:42.142866 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/201449e6-debb-4b67-80c6-524b309dac1f-kube-api-access-mvdm2" (OuterVolumeSpecName: "kube-api-access-mvdm2") pod "201449e6-debb-4b67-80c6-524b309dac1f" (UID: "201449e6-debb-4b67-80c6-524b309dac1f"). InnerVolumeSpecName "kube-api-access-mvdm2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:15:42 crc kubenswrapper[4789]: I1122 08:15:42.234984 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0d6e8982-c10f-441a-be60-34888bc0acf2-operator-scripts\") pod \"0d6e8982-c10f-441a-be60-34888bc0acf2\" (UID: \"0d6e8982-c10f-441a-be60-34888bc0acf2\") " Nov 22 08:15:42 crc kubenswrapper[4789]: I1122 08:15:42.235460 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/afee6663-d80e-41bc-af23-7d64a23d4b03-config-data\") pod \"afee6663-d80e-41bc-af23-7d64a23d4b03\" (UID: \"afee6663-d80e-41bc-af23-7d64a23d4b03\") " Nov 22 08:15:42 crc kubenswrapper[4789]: I1122 08:15:42.235490 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-27cwh\" (UniqueName: \"kubernetes.io/projected/0d6e8982-c10f-441a-be60-34888bc0acf2-kube-api-access-27cwh\") pod \"0d6e8982-c10f-441a-be60-34888bc0acf2\" (UID: \"0d6e8982-c10f-441a-be60-34888bc0acf2\") " Nov 22 08:15:42 crc kubenswrapper[4789]: I1122 08:15:42.235579 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/afee6663-d80e-41bc-af23-7d64a23d4b03-db-sync-config-data\") pod \"afee6663-d80e-41bc-af23-7d64a23d4b03\" (UID: \"afee6663-d80e-41bc-af23-7d64a23d4b03\") " Nov 22 08:15:42 crc kubenswrapper[4789]: I1122 08:15:42.235593 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0d6e8982-c10f-441a-be60-34888bc0acf2-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0d6e8982-c10f-441a-be60-34888bc0acf2" (UID: "0d6e8982-c10f-441a-be60-34888bc0acf2"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:15:42 crc kubenswrapper[4789]: I1122 08:15:42.235713 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-78d92\" (UniqueName: \"kubernetes.io/projected/afee6663-d80e-41bc-af23-7d64a23d4b03-kube-api-access-78d92\") pod \"afee6663-d80e-41bc-af23-7d64a23d4b03\" (UID: \"afee6663-d80e-41bc-af23-7d64a23d4b03\") " Nov 22 08:15:42 crc kubenswrapper[4789]: I1122 08:15:42.235781 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/afee6663-d80e-41bc-af23-7d64a23d4b03-combined-ca-bundle\") pod \"afee6663-d80e-41bc-af23-7d64a23d4b03\" (UID: \"afee6663-d80e-41bc-af23-7d64a23d4b03\") " Nov 22 08:15:42 crc kubenswrapper[4789]: I1122 08:15:42.236135 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-64lp4\" (UniqueName: \"kubernetes.io/projected/b56702e5-c634-4162-9aab-75c723993be4-kube-api-access-64lp4\") on node \"crc\" DevicePath \"\"" Nov 22 08:15:42 crc kubenswrapper[4789]: I1122 08:15:42.236157 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nbj26\" (UniqueName: \"kubernetes.io/projected/56bd722c-eb14-4063-9ddb-9c010f461238-kube-api-access-nbj26\") on node \"crc\" DevicePath \"\"" Nov 22 08:15:42 crc kubenswrapper[4789]: I1122 08:15:42.236167 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mvdm2\" (UniqueName: \"kubernetes.io/projected/201449e6-debb-4b67-80c6-524b309dac1f-kube-api-access-mvdm2\") on node \"crc\" DevicePath \"\"" Nov 22 08:15:42 crc kubenswrapper[4789]: I1122 08:15:42.236175 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jtjr4\" (UniqueName: \"kubernetes.io/projected/41a9281e-b846-4bd9-9c3a-085fe7d8e6f2-kube-api-access-jtjr4\") on node \"crc\" DevicePath \"\"" Nov 22 08:15:42 crc kubenswrapper[4789]: I1122 08:15:42.236184 4789 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/201449e6-debb-4b67-80c6-524b309dac1f-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:15:42 crc kubenswrapper[4789]: I1122 08:15:42.236195 4789 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0d6e8982-c10f-441a-be60-34888bc0acf2-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:15:42 crc kubenswrapper[4789]: I1122 08:15:42.236206 4789 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0c9e2b87-ef93-4176-9362-046e849033a5-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:15:42 crc kubenswrapper[4789]: I1122 08:15:42.236218 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kgv85\" (UniqueName: \"kubernetes.io/projected/0c9e2b87-ef93-4176-9362-046e849033a5-kube-api-access-kgv85\") on node \"crc\" DevicePath \"\"" Nov 22 08:15:42 crc kubenswrapper[4789]: I1122 08:15:42.238266 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d6e8982-c10f-441a-be60-34888bc0acf2-kube-api-access-27cwh" (OuterVolumeSpecName: "kube-api-access-27cwh") pod "0d6e8982-c10f-441a-be60-34888bc0acf2" (UID: "0d6e8982-c10f-441a-be60-34888bc0acf2"). InnerVolumeSpecName "kube-api-access-27cwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:15:42 crc kubenswrapper[4789]: I1122 08:15:42.239323 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/afee6663-d80e-41bc-af23-7d64a23d4b03-kube-api-access-78d92" (OuterVolumeSpecName: "kube-api-access-78d92") pod "afee6663-d80e-41bc-af23-7d64a23d4b03" (UID: "afee6663-d80e-41bc-af23-7d64a23d4b03"). InnerVolumeSpecName "kube-api-access-78d92". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:15:42 crc kubenswrapper[4789]: I1122 08:15:42.240841 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/afee6663-d80e-41bc-af23-7d64a23d4b03-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "afee6663-d80e-41bc-af23-7d64a23d4b03" (UID: "afee6663-d80e-41bc-af23-7d64a23d4b03"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:15:42 crc kubenswrapper[4789]: I1122 08:15:42.258970 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/afee6663-d80e-41bc-af23-7d64a23d4b03-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "afee6663-d80e-41bc-af23-7d64a23d4b03" (UID: "afee6663-d80e-41bc-af23-7d64a23d4b03"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:15:42 crc kubenswrapper[4789]: I1122 08:15:42.273007 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/afee6663-d80e-41bc-af23-7d64a23d4b03-config-data" (OuterVolumeSpecName: "config-data") pod "afee6663-d80e-41bc-af23-7d64a23d4b03" (UID: "afee6663-d80e-41bc-af23-7d64a23d4b03"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:15:42 crc kubenswrapper[4789]: I1122 08:15:42.337775 4789 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/afee6663-d80e-41bc-af23-7d64a23d4b03-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:15:42 crc kubenswrapper[4789]: I1122 08:15:42.337824 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-78d92\" (UniqueName: \"kubernetes.io/projected/afee6663-d80e-41bc-af23-7d64a23d4b03-kube-api-access-78d92\") on node \"crc\" DevicePath \"\"" Nov 22 08:15:42 crc kubenswrapper[4789]: I1122 08:15:42.337838 4789 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/afee6663-d80e-41bc-af23-7d64a23d4b03-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:15:42 crc kubenswrapper[4789]: I1122 08:15:42.337849 4789 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/afee6663-d80e-41bc-af23-7d64a23d4b03-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:15:42 crc kubenswrapper[4789]: I1122 08:15:42.337859 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-27cwh\" (UniqueName: \"kubernetes.io/projected/0d6e8982-c10f-441a-be60-34888bc0acf2-kube-api-access-27cwh\") on node \"crc\" DevicePath \"\"" Nov 22 08:15:42 crc kubenswrapper[4789]: I1122 08:15:42.525588 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-2375-account-create-chvgz" event={"ID":"41a9281e-b846-4bd9-9c3a-085fe7d8e6f2","Type":"ContainerDied","Data":"0f9c5d457c61c04370148a8fabdbd567860e323257c44bff250797b3586f1c20"} Nov 22 08:15:42 crc kubenswrapper[4789]: I1122 08:15:42.525631 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0f9c5d457c61c04370148a8fabdbd567860e323257c44bff250797b3586f1c20" Nov 22 08:15:42 crc kubenswrapper[4789]: I1122 08:15:42.526105 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-2375-account-create-chvgz" Nov 22 08:15:42 crc kubenswrapper[4789]: I1122 08:15:42.527035 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-rnxhf" Nov 22 08:15:42 crc kubenswrapper[4789]: I1122 08:15:42.527059 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-rnxhf" event={"ID":"b56702e5-c634-4162-9aab-75c723993be4","Type":"ContainerDied","Data":"3a636d03abc106705a3d25820b0309a22439288896e5142a372b09f44fc2e660"} Nov 22 08:15:42 crc kubenswrapper[4789]: I1122 08:15:42.527109 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3a636d03abc106705a3d25820b0309a22439288896e5142a372b09f44fc2e660" Nov 22 08:15:42 crc kubenswrapper[4789]: I1122 08:15:42.529643 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-64ckc" event={"ID":"0c9e2b87-ef93-4176-9362-046e849033a5","Type":"ContainerDied","Data":"2017a2f5fa6bff574c4e3e4ee9678cc52eedbbb335845664ef9734a2b25ccc4c"} Nov 22 08:15:42 crc kubenswrapper[4789]: I1122 08:15:42.529680 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2017a2f5fa6bff574c4e3e4ee9678cc52eedbbb335845664ef9734a2b25ccc4c" Nov 22 08:15:42 crc kubenswrapper[4789]: I1122 08:15:42.529655 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-64ckc" Nov 22 08:15:42 crc kubenswrapper[4789]: I1122 08:15:42.531385 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-xhfrw" event={"ID":"afee6663-d80e-41bc-af23-7d64a23d4b03","Type":"ContainerDied","Data":"1d64ac414476e3ab051cfab249ebdfbbc852d987e6367ba307f5068368bdc726"} Nov 22 08:15:42 crc kubenswrapper[4789]: I1122 08:15:42.531424 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1d64ac414476e3ab051cfab249ebdfbbc852d987e6367ba307f5068368bdc726" Nov 22 08:15:42 crc kubenswrapper[4789]: I1122 08:15:42.531427 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-xhfrw" Nov 22 08:15:42 crc kubenswrapper[4789]: I1122 08:15:42.534583 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-411a-account-create-dh7ln" Nov 22 08:15:42 crc kubenswrapper[4789]: I1122 08:15:42.534973 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-411a-account-create-dh7ln" event={"ID":"201449e6-debb-4b67-80c6-524b309dac1f","Type":"ContainerDied","Data":"4a6d4013035f81ae9b3817796a9360b081d5c119cf3a535ec59efe0fd12a5e7e"} Nov 22 08:15:42 crc kubenswrapper[4789]: I1122 08:15:42.535026 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4a6d4013035f81ae9b3817796a9360b081d5c119cf3a535ec59efe0fd12a5e7e" Nov 22 08:15:42 crc kubenswrapper[4789]: I1122 08:15:42.537234 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-5a05-account-create-dcglb" event={"ID":"56bd722c-eb14-4063-9ddb-9c010f461238","Type":"ContainerDied","Data":"ce2102086904fda084b67280ebafdb5a6e5d44cc7a118e3ee8345041ae637eed"} Nov 22 08:15:42 crc kubenswrapper[4789]: I1122 08:15:42.537258 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-5a05-account-create-dcglb" Nov 22 08:15:42 crc kubenswrapper[4789]: I1122 08:15:42.537271 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ce2102086904fda084b67280ebafdb5a6e5d44cc7a118e3ee8345041ae637eed" Nov 22 08:15:42 crc kubenswrapper[4789]: I1122 08:15:42.538950 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-nsptz" event={"ID":"0d6e8982-c10f-441a-be60-34888bc0acf2","Type":"ContainerDied","Data":"f36aec206d0d79323906928bb7454a8e7498aea41a96791b535964f97222e300"} Nov 22 08:15:42 crc kubenswrapper[4789]: I1122 08:15:42.538987 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f36aec206d0d79323906928bb7454a8e7498aea41a96791b535964f97222e300" Nov 22 08:15:42 crc kubenswrapper[4789]: I1122 08:15:42.539550 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-nsptz" Nov 22 08:15:43 crc kubenswrapper[4789]: I1122 08:15:43.417731 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-569d458467-hxxbh"] Nov 22 08:15:43 crc kubenswrapper[4789]: E1122 08:15:43.418719 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56bd722c-eb14-4063-9ddb-9c010f461238" containerName="mariadb-account-create" Nov 22 08:15:43 crc kubenswrapper[4789]: I1122 08:15:43.418737 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="56bd722c-eb14-4063-9ddb-9c010f461238" containerName="mariadb-account-create" Nov 22 08:15:43 crc kubenswrapper[4789]: E1122 08:15:43.418790 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="201449e6-debb-4b67-80c6-524b309dac1f" containerName="mariadb-account-create" Nov 22 08:15:43 crc kubenswrapper[4789]: I1122 08:15:43.418801 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="201449e6-debb-4b67-80c6-524b309dac1f" containerName="mariadb-account-create" Nov 22 08:15:43 crc kubenswrapper[4789]: E1122 08:15:43.418814 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afee6663-d80e-41bc-af23-7d64a23d4b03" containerName="glance-db-sync" Nov 22 08:15:43 crc kubenswrapper[4789]: I1122 08:15:43.418824 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="afee6663-d80e-41bc-af23-7d64a23d4b03" containerName="glance-db-sync" Nov 22 08:15:43 crc kubenswrapper[4789]: E1122 08:15:43.418873 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d6e8982-c10f-441a-be60-34888bc0acf2" containerName="mariadb-database-create" Nov 22 08:15:43 crc kubenswrapper[4789]: I1122 08:15:43.418883 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d6e8982-c10f-441a-be60-34888bc0acf2" containerName="mariadb-database-create" Nov 22 08:15:43 crc kubenswrapper[4789]: E1122 08:15:43.418895 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41a9281e-b846-4bd9-9c3a-085fe7d8e6f2" containerName="mariadb-account-create" Nov 22 08:15:43 crc kubenswrapper[4789]: I1122 08:15:43.418903 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="41a9281e-b846-4bd9-9c3a-085fe7d8e6f2" containerName="mariadb-account-create" Nov 22 08:15:43 crc kubenswrapper[4789]: E1122 08:15:43.418958 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c9e2b87-ef93-4176-9362-046e849033a5" containerName="mariadb-database-create" Nov 22 08:15:43 crc kubenswrapper[4789]: I1122 08:15:43.418968 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c9e2b87-ef93-4176-9362-046e849033a5" containerName="mariadb-database-create" Nov 22 08:15:43 crc kubenswrapper[4789]: E1122 08:15:43.418981 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b56702e5-c634-4162-9aab-75c723993be4" containerName="mariadb-database-create" Nov 22 08:15:43 crc kubenswrapper[4789]: I1122 08:15:43.418989 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="b56702e5-c634-4162-9aab-75c723993be4" containerName="mariadb-database-create" Nov 22 08:15:43 crc kubenswrapper[4789]: I1122 08:15:43.419298 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="56bd722c-eb14-4063-9ddb-9c010f461238" containerName="mariadb-account-create" Nov 22 08:15:43 crc kubenswrapper[4789]: I1122 08:15:43.419314 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="b56702e5-c634-4162-9aab-75c723993be4" containerName="mariadb-database-create" Nov 22 08:15:43 crc kubenswrapper[4789]: I1122 08:15:43.419324 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="201449e6-debb-4b67-80c6-524b309dac1f" containerName="mariadb-account-create" Nov 22 08:15:43 crc kubenswrapper[4789]: I1122 08:15:43.419366 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="41a9281e-b846-4bd9-9c3a-085fe7d8e6f2" containerName="mariadb-account-create" Nov 22 08:15:43 crc kubenswrapper[4789]: I1122 08:15:43.419383 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d6e8982-c10f-441a-be60-34888bc0acf2" containerName="mariadb-database-create" Nov 22 08:15:43 crc kubenswrapper[4789]: I1122 08:15:43.419395 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c9e2b87-ef93-4176-9362-046e849033a5" containerName="mariadb-database-create" Nov 22 08:15:43 crc kubenswrapper[4789]: I1122 08:15:43.419405 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="afee6663-d80e-41bc-af23-7d64a23d4b03" containerName="glance-db-sync" Nov 22 08:15:43 crc kubenswrapper[4789]: I1122 08:15:43.421543 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-569d458467-hxxbh" Nov 22 08:15:43 crc kubenswrapper[4789]: I1122 08:15:43.442682 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-569d458467-hxxbh"] Nov 22 08:15:43 crc kubenswrapper[4789]: I1122 08:15:43.548052 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-tx8g5" event={"ID":"4c80081f-1494-4fd4-ba90-da18af310075","Type":"ContainerStarted","Data":"578155d67afcd1f4a2024c85e97167323cd513d97c07c4883d838855bb132d55"} Nov 22 08:15:43 crc kubenswrapper[4789]: I1122 08:15:43.556300 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b48e8695-47fa-4a3b-af9f-847184c147fd-ovsdbserver-sb\") pod \"dnsmasq-dns-569d458467-hxxbh\" (UID: \"b48e8695-47fa-4a3b-af9f-847184c147fd\") " pod="openstack/dnsmasq-dns-569d458467-hxxbh" Nov 22 08:15:43 crc kubenswrapper[4789]: I1122 08:15:43.556390 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b48e8695-47fa-4a3b-af9f-847184c147fd-config\") pod \"dnsmasq-dns-569d458467-hxxbh\" (UID: \"b48e8695-47fa-4a3b-af9f-847184c147fd\") " pod="openstack/dnsmasq-dns-569d458467-hxxbh" Nov 22 08:15:43 crc kubenswrapper[4789]: I1122 08:15:43.556524 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b48e8695-47fa-4a3b-af9f-847184c147fd-ovsdbserver-nb\") pod \"dnsmasq-dns-569d458467-hxxbh\" (UID: \"b48e8695-47fa-4a3b-af9f-847184c147fd\") " pod="openstack/dnsmasq-dns-569d458467-hxxbh" Nov 22 08:15:43 crc kubenswrapper[4789]: I1122 08:15:43.556607 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rfg9x\" (UniqueName: \"kubernetes.io/projected/b48e8695-47fa-4a3b-af9f-847184c147fd-kube-api-access-rfg9x\") pod \"dnsmasq-dns-569d458467-hxxbh\" (UID: \"b48e8695-47fa-4a3b-af9f-847184c147fd\") " pod="openstack/dnsmasq-dns-569d458467-hxxbh" Nov 22 08:15:43 crc kubenswrapper[4789]: I1122 08:15:43.556667 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b48e8695-47fa-4a3b-af9f-847184c147fd-dns-svc\") pod \"dnsmasq-dns-569d458467-hxxbh\" (UID: \"b48e8695-47fa-4a3b-af9f-847184c147fd\") " pod="openstack/dnsmasq-dns-569d458467-hxxbh" Nov 22 08:15:43 crc kubenswrapper[4789]: I1122 08:15:43.568939 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-tx8g5" podStartSLOduration=2.786011798 podStartE2EDuration="7.568915523s" podCreationTimestamp="2025-11-22 08:15:36 +0000 UTC" firstStartedPulling="2025-11-22 08:15:37.745144429 +0000 UTC m=+1151.979544702" lastFinishedPulling="2025-11-22 08:15:42.528048144 +0000 UTC m=+1156.762448427" observedRunningTime="2025-11-22 08:15:43.560792781 +0000 UTC m=+1157.795193074" watchObservedRunningTime="2025-11-22 08:15:43.568915523 +0000 UTC m=+1157.803315796" Nov 22 08:15:43 crc kubenswrapper[4789]: I1122 08:15:43.658615 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b48e8695-47fa-4a3b-af9f-847184c147fd-ovsdbserver-sb\") pod \"dnsmasq-dns-569d458467-hxxbh\" (UID: \"b48e8695-47fa-4a3b-af9f-847184c147fd\") " pod="openstack/dnsmasq-dns-569d458467-hxxbh" Nov 22 08:15:43 crc kubenswrapper[4789]: I1122 08:15:43.658720 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b48e8695-47fa-4a3b-af9f-847184c147fd-config\") pod \"dnsmasq-dns-569d458467-hxxbh\" (UID: \"b48e8695-47fa-4a3b-af9f-847184c147fd\") " pod="openstack/dnsmasq-dns-569d458467-hxxbh" Nov 22 08:15:43 crc kubenswrapper[4789]: I1122 08:15:43.658799 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b48e8695-47fa-4a3b-af9f-847184c147fd-ovsdbserver-nb\") pod \"dnsmasq-dns-569d458467-hxxbh\" (UID: \"b48e8695-47fa-4a3b-af9f-847184c147fd\") " pod="openstack/dnsmasq-dns-569d458467-hxxbh" Nov 22 08:15:43 crc kubenswrapper[4789]: I1122 08:15:43.658830 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rfg9x\" (UniqueName: \"kubernetes.io/projected/b48e8695-47fa-4a3b-af9f-847184c147fd-kube-api-access-rfg9x\") pod \"dnsmasq-dns-569d458467-hxxbh\" (UID: \"b48e8695-47fa-4a3b-af9f-847184c147fd\") " pod="openstack/dnsmasq-dns-569d458467-hxxbh" Nov 22 08:15:43 crc kubenswrapper[4789]: I1122 08:15:43.658853 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b48e8695-47fa-4a3b-af9f-847184c147fd-dns-svc\") pod \"dnsmasq-dns-569d458467-hxxbh\" (UID: \"b48e8695-47fa-4a3b-af9f-847184c147fd\") " pod="openstack/dnsmasq-dns-569d458467-hxxbh" Nov 22 08:15:43 crc kubenswrapper[4789]: I1122 08:15:43.659674 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b48e8695-47fa-4a3b-af9f-847184c147fd-ovsdbserver-sb\") pod \"dnsmasq-dns-569d458467-hxxbh\" (UID: \"b48e8695-47fa-4a3b-af9f-847184c147fd\") " pod="openstack/dnsmasq-dns-569d458467-hxxbh" Nov 22 08:15:43 crc kubenswrapper[4789]: I1122 08:15:43.659779 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b48e8695-47fa-4a3b-af9f-847184c147fd-dns-svc\") pod \"dnsmasq-dns-569d458467-hxxbh\" (UID: \"b48e8695-47fa-4a3b-af9f-847184c147fd\") " pod="openstack/dnsmasq-dns-569d458467-hxxbh" Nov 22 08:15:43 crc kubenswrapper[4789]: I1122 08:15:43.659988 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b48e8695-47fa-4a3b-af9f-847184c147fd-ovsdbserver-nb\") pod \"dnsmasq-dns-569d458467-hxxbh\" (UID: \"b48e8695-47fa-4a3b-af9f-847184c147fd\") " pod="openstack/dnsmasq-dns-569d458467-hxxbh" Nov 22 08:15:43 crc kubenswrapper[4789]: I1122 08:15:43.660145 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b48e8695-47fa-4a3b-af9f-847184c147fd-config\") pod \"dnsmasq-dns-569d458467-hxxbh\" (UID: \"b48e8695-47fa-4a3b-af9f-847184c147fd\") " pod="openstack/dnsmasq-dns-569d458467-hxxbh" Nov 22 08:15:43 crc kubenswrapper[4789]: I1122 08:15:43.680609 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rfg9x\" (UniqueName: \"kubernetes.io/projected/b48e8695-47fa-4a3b-af9f-847184c147fd-kube-api-access-rfg9x\") pod \"dnsmasq-dns-569d458467-hxxbh\" (UID: \"b48e8695-47fa-4a3b-af9f-847184c147fd\") " pod="openstack/dnsmasq-dns-569d458467-hxxbh" Nov 22 08:15:43 crc kubenswrapper[4789]: I1122 08:15:43.745640 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-569d458467-hxxbh" Nov 22 08:15:44 crc kubenswrapper[4789]: I1122 08:15:44.163559 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-569d458467-hxxbh"] Nov 22 08:15:44 crc kubenswrapper[4789]: W1122 08:15:44.175402 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb48e8695_47fa_4a3b_af9f_847184c147fd.slice/crio-b7b2f39d7a7a2aedce672cb740da7f0ffd2de1c748289746f031ddeef52a0df0 WatchSource:0}: Error finding container b7b2f39d7a7a2aedce672cb740da7f0ffd2de1c748289746f031ddeef52a0df0: Status 404 returned error can't find the container with id b7b2f39d7a7a2aedce672cb740da7f0ffd2de1c748289746f031ddeef52a0df0 Nov 22 08:15:44 crc kubenswrapper[4789]: I1122 08:15:44.557689 4789 generic.go:334] "Generic (PLEG): container finished" podID="b48e8695-47fa-4a3b-af9f-847184c147fd" containerID="6ac1e3b8a17f8aa8ccc8ad6418df1dabe4253068f96afc0d213f79064b5c2e23" exitCode=0 Nov 22 08:15:44 crc kubenswrapper[4789]: I1122 08:15:44.557827 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-569d458467-hxxbh" event={"ID":"b48e8695-47fa-4a3b-af9f-847184c147fd","Type":"ContainerDied","Data":"6ac1e3b8a17f8aa8ccc8ad6418df1dabe4253068f96afc0d213f79064b5c2e23"} Nov 22 08:15:44 crc kubenswrapper[4789]: I1122 08:15:44.558256 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-569d458467-hxxbh" event={"ID":"b48e8695-47fa-4a3b-af9f-847184c147fd","Type":"ContainerStarted","Data":"b7b2f39d7a7a2aedce672cb740da7f0ffd2de1c748289746f031ddeef52a0df0"} Nov 22 08:15:45 crc kubenswrapper[4789]: I1122 08:15:45.566831 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-569d458467-hxxbh" event={"ID":"b48e8695-47fa-4a3b-af9f-847184c147fd","Type":"ContainerStarted","Data":"904da12081f37b85040b7953040498e9abae99a90d641670f730e4916c15e0a9"} Nov 22 08:15:45 crc kubenswrapper[4789]: I1122 08:15:45.567232 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-569d458467-hxxbh" Nov 22 08:15:45 crc kubenswrapper[4789]: I1122 08:15:45.582668 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-569d458467-hxxbh" podStartSLOduration=2.582647475 podStartE2EDuration="2.582647475s" podCreationTimestamp="2025-11-22 08:15:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:15:45.581077612 +0000 UTC m=+1159.815477895" watchObservedRunningTime="2025-11-22 08:15:45.582647475 +0000 UTC m=+1159.817047768" Nov 22 08:15:46 crc kubenswrapper[4789]: I1122 08:15:46.576235 4789 generic.go:334] "Generic (PLEG): container finished" podID="4c80081f-1494-4fd4-ba90-da18af310075" containerID="578155d67afcd1f4a2024c85e97167323cd513d97c07c4883d838855bb132d55" exitCode=0 Nov 22 08:15:46 crc kubenswrapper[4789]: I1122 08:15:46.576344 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-tx8g5" event={"ID":"4c80081f-1494-4fd4-ba90-da18af310075","Type":"ContainerDied","Data":"578155d67afcd1f4a2024c85e97167323cd513d97c07c4883d838855bb132d55"} Nov 22 08:15:47 crc kubenswrapper[4789]: I1122 08:15:47.844200 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-tx8g5" Nov 22 08:15:48 crc kubenswrapper[4789]: I1122 08:15:48.032806 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p47bc\" (UniqueName: \"kubernetes.io/projected/4c80081f-1494-4fd4-ba90-da18af310075-kube-api-access-p47bc\") pod \"4c80081f-1494-4fd4-ba90-da18af310075\" (UID: \"4c80081f-1494-4fd4-ba90-da18af310075\") " Nov 22 08:15:48 crc kubenswrapper[4789]: I1122 08:15:48.032917 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c80081f-1494-4fd4-ba90-da18af310075-config-data\") pod \"4c80081f-1494-4fd4-ba90-da18af310075\" (UID: \"4c80081f-1494-4fd4-ba90-da18af310075\") " Nov 22 08:15:48 crc kubenswrapper[4789]: I1122 08:15:48.033024 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c80081f-1494-4fd4-ba90-da18af310075-combined-ca-bundle\") pod \"4c80081f-1494-4fd4-ba90-da18af310075\" (UID: \"4c80081f-1494-4fd4-ba90-da18af310075\") " Nov 22 08:15:48 crc kubenswrapper[4789]: I1122 08:15:48.041037 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c80081f-1494-4fd4-ba90-da18af310075-kube-api-access-p47bc" (OuterVolumeSpecName: "kube-api-access-p47bc") pod "4c80081f-1494-4fd4-ba90-da18af310075" (UID: "4c80081f-1494-4fd4-ba90-da18af310075"). InnerVolumeSpecName "kube-api-access-p47bc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:15:48 crc kubenswrapper[4789]: I1122 08:15:48.060052 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c80081f-1494-4fd4-ba90-da18af310075-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4c80081f-1494-4fd4-ba90-da18af310075" (UID: "4c80081f-1494-4fd4-ba90-da18af310075"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:15:48 crc kubenswrapper[4789]: I1122 08:15:48.078319 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c80081f-1494-4fd4-ba90-da18af310075-config-data" (OuterVolumeSpecName: "config-data") pod "4c80081f-1494-4fd4-ba90-da18af310075" (UID: "4c80081f-1494-4fd4-ba90-da18af310075"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:15:48 crc kubenswrapper[4789]: I1122 08:15:48.134826 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p47bc\" (UniqueName: \"kubernetes.io/projected/4c80081f-1494-4fd4-ba90-da18af310075-kube-api-access-p47bc\") on node \"crc\" DevicePath \"\"" Nov 22 08:15:48 crc kubenswrapper[4789]: I1122 08:15:48.134870 4789 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c80081f-1494-4fd4-ba90-da18af310075-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:15:48 crc kubenswrapper[4789]: I1122 08:15:48.134886 4789 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c80081f-1494-4fd4-ba90-da18af310075-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:15:48 crc kubenswrapper[4789]: I1122 08:15:48.592059 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-tx8g5" event={"ID":"4c80081f-1494-4fd4-ba90-da18af310075","Type":"ContainerDied","Data":"98f965b1787cf989365213bef3e4f62fa5f12da534bbec60cd171697b6c98848"} Nov 22 08:15:48 crc kubenswrapper[4789]: I1122 08:15:48.592402 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="98f965b1787cf989365213bef3e4f62fa5f12da534bbec60cd171697b6c98848" Nov 22 08:15:48 crc kubenswrapper[4789]: I1122 08:15:48.592115 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-tx8g5" Nov 22 08:15:48 crc kubenswrapper[4789]: I1122 08:15:48.768204 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-569d458467-hxxbh"] Nov 22 08:15:48 crc kubenswrapper[4789]: I1122 08:15:48.768554 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-569d458467-hxxbh" podUID="b48e8695-47fa-4a3b-af9f-847184c147fd" containerName="dnsmasq-dns" containerID="cri-o://904da12081f37b85040b7953040498e9abae99a90d641670f730e4916c15e0a9" gracePeriod=10 Nov 22 08:15:48 crc kubenswrapper[4789]: I1122 08:15:48.802498 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-b76c757b7-z85gc"] Nov 22 08:15:48 crc kubenswrapper[4789]: E1122 08:15:48.802948 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c80081f-1494-4fd4-ba90-da18af310075" containerName="keystone-db-sync" Nov 22 08:15:48 crc kubenswrapper[4789]: I1122 08:15:48.802972 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c80081f-1494-4fd4-ba90-da18af310075" containerName="keystone-db-sync" Nov 22 08:15:48 crc kubenswrapper[4789]: I1122 08:15:48.803163 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c80081f-1494-4fd4-ba90-da18af310075" containerName="keystone-db-sync" Nov 22 08:15:48 crc kubenswrapper[4789]: I1122 08:15:48.804194 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b76c757b7-z85gc" Nov 22 08:15:48 crc kubenswrapper[4789]: I1122 08:15:48.815699 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b76c757b7-z85gc"] Nov 22 08:15:48 crc kubenswrapper[4789]: I1122 08:15:48.848823 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-9ts56"] Nov 22 08:15:48 crc kubenswrapper[4789]: I1122 08:15:48.849938 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-9ts56" Nov 22 08:15:48 crc kubenswrapper[4789]: I1122 08:15:48.852515 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 22 08:15:48 crc kubenswrapper[4789]: I1122 08:15:48.852799 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 22 08:15:48 crc kubenswrapper[4789]: I1122 08:15:48.854325 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-wnnts" Nov 22 08:15:48 crc kubenswrapper[4789]: I1122 08:15:48.854408 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 22 08:15:48 crc kubenswrapper[4789]: I1122 08:15:48.854625 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 22 08:15:48 crc kubenswrapper[4789]: I1122 08:15:48.870837 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-9ts56"] Nov 22 08:15:48 crc kubenswrapper[4789]: I1122 08:15:48.946932 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1a8be067-2aaa-4459-8dfb-2ebc7f21b258-config\") pod \"dnsmasq-dns-b76c757b7-z85gc\" (UID: \"1a8be067-2aaa-4459-8dfb-2ebc7f21b258\") " pod="openstack/dnsmasq-dns-b76c757b7-z85gc" Nov 22 08:15:48 crc kubenswrapper[4789]: I1122 08:15:48.946988 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/09e8392f-f97d-4922-be5a-fb6229549344-config-data\") pod \"keystone-bootstrap-9ts56\" (UID: \"09e8392f-f97d-4922-be5a-fb6229549344\") " pod="openstack/keystone-bootstrap-9ts56" Nov 22 08:15:48 crc kubenswrapper[4789]: I1122 08:15:48.947017 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h4brh\" (UniqueName: \"kubernetes.io/projected/1a8be067-2aaa-4459-8dfb-2ebc7f21b258-kube-api-access-h4brh\") pod \"dnsmasq-dns-b76c757b7-z85gc\" (UID: \"1a8be067-2aaa-4459-8dfb-2ebc7f21b258\") " pod="openstack/dnsmasq-dns-b76c757b7-z85gc" Nov 22 08:15:48 crc kubenswrapper[4789]: I1122 08:15:48.947046 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gs4k9\" (UniqueName: \"kubernetes.io/projected/09e8392f-f97d-4922-be5a-fb6229549344-kube-api-access-gs4k9\") pod \"keystone-bootstrap-9ts56\" (UID: \"09e8392f-f97d-4922-be5a-fb6229549344\") " pod="openstack/keystone-bootstrap-9ts56" Nov 22 08:15:48 crc kubenswrapper[4789]: I1122 08:15:48.947068 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1a8be067-2aaa-4459-8dfb-2ebc7f21b258-dns-svc\") pod \"dnsmasq-dns-b76c757b7-z85gc\" (UID: \"1a8be067-2aaa-4459-8dfb-2ebc7f21b258\") " pod="openstack/dnsmasq-dns-b76c757b7-z85gc" Nov 22 08:15:48 crc kubenswrapper[4789]: I1122 08:15:48.947092 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09e8392f-f97d-4922-be5a-fb6229549344-combined-ca-bundle\") pod \"keystone-bootstrap-9ts56\" (UID: \"09e8392f-f97d-4922-be5a-fb6229549344\") " pod="openstack/keystone-bootstrap-9ts56" Nov 22 08:15:48 crc kubenswrapper[4789]: I1122 08:15:48.947107 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1a8be067-2aaa-4459-8dfb-2ebc7f21b258-ovsdbserver-nb\") pod \"dnsmasq-dns-b76c757b7-z85gc\" (UID: \"1a8be067-2aaa-4459-8dfb-2ebc7f21b258\") " pod="openstack/dnsmasq-dns-b76c757b7-z85gc" Nov 22 08:15:48 crc kubenswrapper[4789]: I1122 08:15:48.947145 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1a8be067-2aaa-4459-8dfb-2ebc7f21b258-ovsdbserver-sb\") pod \"dnsmasq-dns-b76c757b7-z85gc\" (UID: \"1a8be067-2aaa-4459-8dfb-2ebc7f21b258\") " pod="openstack/dnsmasq-dns-b76c757b7-z85gc" Nov 22 08:15:48 crc kubenswrapper[4789]: I1122 08:15:48.947179 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/09e8392f-f97d-4922-be5a-fb6229549344-scripts\") pod \"keystone-bootstrap-9ts56\" (UID: \"09e8392f-f97d-4922-be5a-fb6229549344\") " pod="openstack/keystone-bootstrap-9ts56" Nov 22 08:15:48 crc kubenswrapper[4789]: I1122 08:15:48.947201 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/09e8392f-f97d-4922-be5a-fb6229549344-credential-keys\") pod \"keystone-bootstrap-9ts56\" (UID: \"09e8392f-f97d-4922-be5a-fb6229549344\") " pod="openstack/keystone-bootstrap-9ts56" Nov 22 08:15:48 crc kubenswrapper[4789]: I1122 08:15:48.947220 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/09e8392f-f97d-4922-be5a-fb6229549344-fernet-keys\") pod \"keystone-bootstrap-9ts56\" (UID: \"09e8392f-f97d-4922-be5a-fb6229549344\") " pod="openstack/keystone-bootstrap-9ts56" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.003942 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-64b577d9f-n8w7t"] Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.005921 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-64b577d9f-n8w7t" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.009675 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-n4s7g"] Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.013482 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.014027 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.014295 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-cnkvq" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.018629 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-n4s7g" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.021294 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-zhkq9" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.021521 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.021640 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.021743 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.025133 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-64b577d9f-n8w7t"] Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.040768 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-n4s7g"] Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.049472 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gs4k9\" (UniqueName: \"kubernetes.io/projected/09e8392f-f97d-4922-be5a-fb6229549344-kube-api-access-gs4k9\") pod \"keystone-bootstrap-9ts56\" (UID: \"09e8392f-f97d-4922-be5a-fb6229549344\") " pod="openstack/keystone-bootstrap-9ts56" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.049527 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1a8be067-2aaa-4459-8dfb-2ebc7f21b258-dns-svc\") pod \"dnsmasq-dns-b76c757b7-z85gc\" (UID: \"1a8be067-2aaa-4459-8dfb-2ebc7f21b258\") " pod="openstack/dnsmasq-dns-b76c757b7-z85gc" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.049560 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09e8392f-f97d-4922-be5a-fb6229549344-combined-ca-bundle\") pod \"keystone-bootstrap-9ts56\" (UID: \"09e8392f-f97d-4922-be5a-fb6229549344\") " pod="openstack/keystone-bootstrap-9ts56" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.049583 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1a8be067-2aaa-4459-8dfb-2ebc7f21b258-ovsdbserver-nb\") pod \"dnsmasq-dns-b76c757b7-z85gc\" (UID: \"1a8be067-2aaa-4459-8dfb-2ebc7f21b258\") " pod="openstack/dnsmasq-dns-b76c757b7-z85gc" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.049614 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1a8be067-2aaa-4459-8dfb-2ebc7f21b258-ovsdbserver-sb\") pod \"dnsmasq-dns-b76c757b7-z85gc\" (UID: \"1a8be067-2aaa-4459-8dfb-2ebc7f21b258\") " pod="openstack/dnsmasq-dns-b76c757b7-z85gc" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.049665 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/09e8392f-f97d-4922-be5a-fb6229549344-scripts\") pod \"keystone-bootstrap-9ts56\" (UID: \"09e8392f-f97d-4922-be5a-fb6229549344\") " pod="openstack/keystone-bootstrap-9ts56" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.049700 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/09e8392f-f97d-4922-be5a-fb6229549344-credential-keys\") pod \"keystone-bootstrap-9ts56\" (UID: \"09e8392f-f97d-4922-be5a-fb6229549344\") " pod="openstack/keystone-bootstrap-9ts56" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.049726 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/09e8392f-f97d-4922-be5a-fb6229549344-fernet-keys\") pod \"keystone-bootstrap-9ts56\" (UID: \"09e8392f-f97d-4922-be5a-fb6229549344\") " pod="openstack/keystone-bootstrap-9ts56" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.049793 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1a8be067-2aaa-4459-8dfb-2ebc7f21b258-config\") pod \"dnsmasq-dns-b76c757b7-z85gc\" (UID: \"1a8be067-2aaa-4459-8dfb-2ebc7f21b258\") " pod="openstack/dnsmasq-dns-b76c757b7-z85gc" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.049835 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/09e8392f-f97d-4922-be5a-fb6229549344-config-data\") pod \"keystone-bootstrap-9ts56\" (UID: \"09e8392f-f97d-4922-be5a-fb6229549344\") " pod="openstack/keystone-bootstrap-9ts56" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.049871 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h4brh\" (UniqueName: \"kubernetes.io/projected/1a8be067-2aaa-4459-8dfb-2ebc7f21b258-kube-api-access-h4brh\") pod \"dnsmasq-dns-b76c757b7-z85gc\" (UID: \"1a8be067-2aaa-4459-8dfb-2ebc7f21b258\") " pod="openstack/dnsmasq-dns-b76c757b7-z85gc" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.051385 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1a8be067-2aaa-4459-8dfb-2ebc7f21b258-dns-svc\") pod \"dnsmasq-dns-b76c757b7-z85gc\" (UID: \"1a8be067-2aaa-4459-8dfb-2ebc7f21b258\") " pod="openstack/dnsmasq-dns-b76c757b7-z85gc" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.053745 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1a8be067-2aaa-4459-8dfb-2ebc7f21b258-ovsdbserver-sb\") pod \"dnsmasq-dns-b76c757b7-z85gc\" (UID: \"1a8be067-2aaa-4459-8dfb-2ebc7f21b258\") " pod="openstack/dnsmasq-dns-b76c757b7-z85gc" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.055544 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1a8be067-2aaa-4459-8dfb-2ebc7f21b258-config\") pod \"dnsmasq-dns-b76c757b7-z85gc\" (UID: \"1a8be067-2aaa-4459-8dfb-2ebc7f21b258\") " pod="openstack/dnsmasq-dns-b76c757b7-z85gc" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.056203 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1a8be067-2aaa-4459-8dfb-2ebc7f21b258-ovsdbserver-nb\") pod \"dnsmasq-dns-b76c757b7-z85gc\" (UID: \"1a8be067-2aaa-4459-8dfb-2ebc7f21b258\") " pod="openstack/dnsmasq-dns-b76c757b7-z85gc" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.070564 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/09e8392f-f97d-4922-be5a-fb6229549344-credential-keys\") pod \"keystone-bootstrap-9ts56\" (UID: \"09e8392f-f97d-4922-be5a-fb6229549344\") " pod="openstack/keystone-bootstrap-9ts56" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.070643 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.078610 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.082674 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.082917 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.084270 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/09e8392f-f97d-4922-be5a-fb6229549344-scripts\") pod \"keystone-bootstrap-9ts56\" (UID: \"09e8392f-f97d-4922-be5a-fb6229549344\") " pod="openstack/keystone-bootstrap-9ts56" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.088385 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/09e8392f-f97d-4922-be5a-fb6229549344-config-data\") pod \"keystone-bootstrap-9ts56\" (UID: \"09e8392f-f97d-4922-be5a-fb6229549344\") " pod="openstack/keystone-bootstrap-9ts56" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.088492 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h4brh\" (UniqueName: \"kubernetes.io/projected/1a8be067-2aaa-4459-8dfb-2ebc7f21b258-kube-api-access-h4brh\") pod \"dnsmasq-dns-b76c757b7-z85gc\" (UID: \"1a8be067-2aaa-4459-8dfb-2ebc7f21b258\") " pod="openstack/dnsmasq-dns-b76c757b7-z85gc" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.094298 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gs4k9\" (UniqueName: \"kubernetes.io/projected/09e8392f-f97d-4922-be5a-fb6229549344-kube-api-access-gs4k9\") pod \"keystone-bootstrap-9ts56\" (UID: \"09e8392f-f97d-4922-be5a-fb6229549344\") " pod="openstack/keystone-bootstrap-9ts56" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.101333 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/09e8392f-f97d-4922-be5a-fb6229549344-fernet-keys\") pod \"keystone-bootstrap-9ts56\" (UID: \"09e8392f-f97d-4922-be5a-fb6229549344\") " pod="openstack/keystone-bootstrap-9ts56" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.105463 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.106517 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09e8392f-f97d-4922-be5a-fb6229549344-combined-ca-bundle\") pod \"keystone-bootstrap-9ts56\" (UID: \"09e8392f-f97d-4922-be5a-fb6229549344\") " pod="openstack/keystone-bootstrap-9ts56" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.131426 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b76c757b7-z85gc" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.143065 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-lg59m"] Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.144432 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-lg59m" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.163376 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.163952 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-8mlsn" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.165990 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e3b7f50f-961d-472f-9423-18ab69fcdaef-logs\") pod \"horizon-64b577d9f-n8w7t\" (UID: \"e3b7f50f-961d-472f-9423-18ab69fcdaef\") " pod="openstack/horizon-64b577d9f-n8w7t" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.166028 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6bf8aa3e-a7e7-4f0c-a542-d668a5623707-scripts\") pod \"cinder-db-sync-lg59m\" (UID: \"6bf8aa3e-a7e7-4f0c-a542-d668a5623707\") " pod="openstack/cinder-db-sync-lg59m" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.166050 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/e3b7f50f-961d-472f-9423-18ab69fcdaef-horizon-secret-key\") pod \"horizon-64b577d9f-n8w7t\" (UID: \"e3b7f50f-961d-472f-9423-18ab69fcdaef\") " pod="openstack/horizon-64b577d9f-n8w7t" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.166071 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6bf8aa3e-a7e7-4f0c-a542-d668a5623707-db-sync-config-data\") pod \"cinder-db-sync-lg59m\" (UID: \"6bf8aa3e-a7e7-4f0c-a542-d668a5623707\") " pod="openstack/cinder-db-sync-lg59m" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.166087 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/94d7bb01-64a7-442a-aee4-e958d8511b76-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"94d7bb01-64a7-442a-aee4-e958d8511b76\") " pod="openstack/ceilometer-0" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.166101 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/733e94b9-65c2-406e-9ebe-244585e4e15d-config\") pod \"neutron-db-sync-n4s7g\" (UID: \"733e94b9-65c2-406e-9ebe-244585e4e15d\") " pod="openstack/neutron-db-sync-n4s7g" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.166123 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t9qz8\" (UniqueName: \"kubernetes.io/projected/6bf8aa3e-a7e7-4f0c-a542-d668a5623707-kube-api-access-t9qz8\") pod \"cinder-db-sync-lg59m\" (UID: \"6bf8aa3e-a7e7-4f0c-a542-d668a5623707\") " pod="openstack/cinder-db-sync-lg59m" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.166139 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/94d7bb01-64a7-442a-aee4-e958d8511b76-run-httpd\") pod \"ceilometer-0\" (UID: \"94d7bb01-64a7-442a-aee4-e958d8511b76\") " pod="openstack/ceilometer-0" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.166161 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e3b7f50f-961d-472f-9423-18ab69fcdaef-scripts\") pod \"horizon-64b577d9f-n8w7t\" (UID: \"e3b7f50f-961d-472f-9423-18ab69fcdaef\") " pod="openstack/horizon-64b577d9f-n8w7t" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.166286 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e3b7f50f-961d-472f-9423-18ab69fcdaef-config-data\") pod \"horizon-64b577d9f-n8w7t\" (UID: \"e3b7f50f-961d-472f-9423-18ab69fcdaef\") " pod="openstack/horizon-64b577d9f-n8w7t" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.166319 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94d7bb01-64a7-442a-aee4-e958d8511b76-config-data\") pod \"ceilometer-0\" (UID: \"94d7bb01-64a7-442a-aee4-e958d8511b76\") " pod="openstack/ceilometer-0" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.166392 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/94d7bb01-64a7-442a-aee4-e958d8511b76-log-httpd\") pod \"ceilometer-0\" (UID: \"94d7bb01-64a7-442a-aee4-e958d8511b76\") " pod="openstack/ceilometer-0" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.166446 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94d7bb01-64a7-442a-aee4-e958d8511b76-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"94d7bb01-64a7-442a-aee4-e958d8511b76\") " pod="openstack/ceilometer-0" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.166460 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/94d7bb01-64a7-442a-aee4-e958d8511b76-scripts\") pod \"ceilometer-0\" (UID: \"94d7bb01-64a7-442a-aee4-e958d8511b76\") " pod="openstack/ceilometer-0" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.166512 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6bf8aa3e-a7e7-4f0c-a542-d668a5623707-combined-ca-bundle\") pod \"cinder-db-sync-lg59m\" (UID: \"6bf8aa3e-a7e7-4f0c-a542-d668a5623707\") " pod="openstack/cinder-db-sync-lg59m" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.166580 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cswpl\" (UniqueName: \"kubernetes.io/projected/e3b7f50f-961d-472f-9423-18ab69fcdaef-kube-api-access-cswpl\") pod \"horizon-64b577d9f-n8w7t\" (UID: \"e3b7f50f-961d-472f-9423-18ab69fcdaef\") " pod="openstack/horizon-64b577d9f-n8w7t" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.166597 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9mw6d\" (UniqueName: \"kubernetes.io/projected/733e94b9-65c2-406e-9ebe-244585e4e15d-kube-api-access-9mw6d\") pod \"neutron-db-sync-n4s7g\" (UID: \"733e94b9-65c2-406e-9ebe-244585e4e15d\") " pod="openstack/neutron-db-sync-n4s7g" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.166613 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/733e94b9-65c2-406e-9ebe-244585e4e15d-combined-ca-bundle\") pod \"neutron-db-sync-n4s7g\" (UID: \"733e94b9-65c2-406e-9ebe-244585e4e15d\") " pod="openstack/neutron-db-sync-n4s7g" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.167476 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rxhrx\" (UniqueName: \"kubernetes.io/projected/94d7bb01-64a7-442a-aee4-e958d8511b76-kube-api-access-rxhrx\") pod \"ceilometer-0\" (UID: \"94d7bb01-64a7-442a-aee4-e958d8511b76\") " pod="openstack/ceilometer-0" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.167513 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6bf8aa3e-a7e7-4f0c-a542-d668a5623707-etc-machine-id\") pod \"cinder-db-sync-lg59m\" (UID: \"6bf8aa3e-a7e7-4f0c-a542-d668a5623707\") " pod="openstack/cinder-db-sync-lg59m" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.167537 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6bf8aa3e-a7e7-4f0c-a542-d668a5623707-config-data\") pod \"cinder-db-sync-lg59m\" (UID: \"6bf8aa3e-a7e7-4f0c-a542-d668a5623707\") " pod="openstack/cinder-db-sync-lg59m" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.173774 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.226955 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-9ts56" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.258888 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-lg59m"] Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.270777 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6bf8aa3e-a7e7-4f0c-a542-d668a5623707-config-data\") pod \"cinder-db-sync-lg59m\" (UID: \"6bf8aa3e-a7e7-4f0c-a542-d668a5623707\") " pod="openstack/cinder-db-sync-lg59m" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.270825 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e3b7f50f-961d-472f-9423-18ab69fcdaef-logs\") pod \"horizon-64b577d9f-n8w7t\" (UID: \"e3b7f50f-961d-472f-9423-18ab69fcdaef\") " pod="openstack/horizon-64b577d9f-n8w7t" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.270846 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6bf8aa3e-a7e7-4f0c-a542-d668a5623707-scripts\") pod \"cinder-db-sync-lg59m\" (UID: \"6bf8aa3e-a7e7-4f0c-a542-d668a5623707\") " pod="openstack/cinder-db-sync-lg59m" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.270867 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/e3b7f50f-961d-472f-9423-18ab69fcdaef-horizon-secret-key\") pod \"horizon-64b577d9f-n8w7t\" (UID: \"e3b7f50f-961d-472f-9423-18ab69fcdaef\") " pod="openstack/horizon-64b577d9f-n8w7t" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.270888 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6bf8aa3e-a7e7-4f0c-a542-d668a5623707-db-sync-config-data\") pod \"cinder-db-sync-lg59m\" (UID: \"6bf8aa3e-a7e7-4f0c-a542-d668a5623707\") " pod="openstack/cinder-db-sync-lg59m" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.270903 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/94d7bb01-64a7-442a-aee4-e958d8511b76-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"94d7bb01-64a7-442a-aee4-e958d8511b76\") " pod="openstack/ceilometer-0" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.270936 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/733e94b9-65c2-406e-9ebe-244585e4e15d-config\") pod \"neutron-db-sync-n4s7g\" (UID: \"733e94b9-65c2-406e-9ebe-244585e4e15d\") " pod="openstack/neutron-db-sync-n4s7g" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.270957 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t9qz8\" (UniqueName: \"kubernetes.io/projected/6bf8aa3e-a7e7-4f0c-a542-d668a5623707-kube-api-access-t9qz8\") pod \"cinder-db-sync-lg59m\" (UID: \"6bf8aa3e-a7e7-4f0c-a542-d668a5623707\") " pod="openstack/cinder-db-sync-lg59m" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.271221 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/94d7bb01-64a7-442a-aee4-e958d8511b76-run-httpd\") pod \"ceilometer-0\" (UID: \"94d7bb01-64a7-442a-aee4-e958d8511b76\") " pod="openstack/ceilometer-0" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.271240 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e3b7f50f-961d-472f-9423-18ab69fcdaef-scripts\") pod \"horizon-64b577d9f-n8w7t\" (UID: \"e3b7f50f-961d-472f-9423-18ab69fcdaef\") " pod="openstack/horizon-64b577d9f-n8w7t" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.271260 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e3b7f50f-961d-472f-9423-18ab69fcdaef-config-data\") pod \"horizon-64b577d9f-n8w7t\" (UID: \"e3b7f50f-961d-472f-9423-18ab69fcdaef\") " pod="openstack/horizon-64b577d9f-n8w7t" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.271283 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94d7bb01-64a7-442a-aee4-e958d8511b76-config-data\") pod \"ceilometer-0\" (UID: \"94d7bb01-64a7-442a-aee4-e958d8511b76\") " pod="openstack/ceilometer-0" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.271330 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/94d7bb01-64a7-442a-aee4-e958d8511b76-log-httpd\") pod \"ceilometer-0\" (UID: \"94d7bb01-64a7-442a-aee4-e958d8511b76\") " pod="openstack/ceilometer-0" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.271356 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94d7bb01-64a7-442a-aee4-e958d8511b76-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"94d7bb01-64a7-442a-aee4-e958d8511b76\") " pod="openstack/ceilometer-0" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.271372 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/94d7bb01-64a7-442a-aee4-e958d8511b76-scripts\") pod \"ceilometer-0\" (UID: \"94d7bb01-64a7-442a-aee4-e958d8511b76\") " pod="openstack/ceilometer-0" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.271395 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6bf8aa3e-a7e7-4f0c-a542-d668a5623707-combined-ca-bundle\") pod \"cinder-db-sync-lg59m\" (UID: \"6bf8aa3e-a7e7-4f0c-a542-d668a5623707\") " pod="openstack/cinder-db-sync-lg59m" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.271412 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cswpl\" (UniqueName: \"kubernetes.io/projected/e3b7f50f-961d-472f-9423-18ab69fcdaef-kube-api-access-cswpl\") pod \"horizon-64b577d9f-n8w7t\" (UID: \"e3b7f50f-961d-472f-9423-18ab69fcdaef\") " pod="openstack/horizon-64b577d9f-n8w7t" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.271431 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9mw6d\" (UniqueName: \"kubernetes.io/projected/733e94b9-65c2-406e-9ebe-244585e4e15d-kube-api-access-9mw6d\") pod \"neutron-db-sync-n4s7g\" (UID: \"733e94b9-65c2-406e-9ebe-244585e4e15d\") " pod="openstack/neutron-db-sync-n4s7g" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.271451 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/733e94b9-65c2-406e-9ebe-244585e4e15d-combined-ca-bundle\") pod \"neutron-db-sync-n4s7g\" (UID: \"733e94b9-65c2-406e-9ebe-244585e4e15d\") " pod="openstack/neutron-db-sync-n4s7g" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.271475 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rxhrx\" (UniqueName: \"kubernetes.io/projected/94d7bb01-64a7-442a-aee4-e958d8511b76-kube-api-access-rxhrx\") pod \"ceilometer-0\" (UID: \"94d7bb01-64a7-442a-aee4-e958d8511b76\") " pod="openstack/ceilometer-0" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.271506 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6bf8aa3e-a7e7-4f0c-a542-d668a5623707-etc-machine-id\") pod \"cinder-db-sync-lg59m\" (UID: \"6bf8aa3e-a7e7-4f0c-a542-d668a5623707\") " pod="openstack/cinder-db-sync-lg59m" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.271580 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6bf8aa3e-a7e7-4f0c-a542-d668a5623707-etc-machine-id\") pod \"cinder-db-sync-lg59m\" (UID: \"6bf8aa3e-a7e7-4f0c-a542-d668a5623707\") " pod="openstack/cinder-db-sync-lg59m" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.277659 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6bf8aa3e-a7e7-4f0c-a542-d668a5623707-scripts\") pod \"cinder-db-sync-lg59m\" (UID: \"6bf8aa3e-a7e7-4f0c-a542-d668a5623707\") " pod="openstack/cinder-db-sync-lg59m" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.277969 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e3b7f50f-961d-472f-9423-18ab69fcdaef-logs\") pod \"horizon-64b577d9f-n8w7t\" (UID: \"e3b7f50f-961d-472f-9423-18ab69fcdaef\") " pod="openstack/horizon-64b577d9f-n8w7t" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.278898 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e3b7f50f-961d-472f-9423-18ab69fcdaef-scripts\") pod \"horizon-64b577d9f-n8w7t\" (UID: \"e3b7f50f-961d-472f-9423-18ab69fcdaef\") " pod="openstack/horizon-64b577d9f-n8w7t" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.279192 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/94d7bb01-64a7-442a-aee4-e958d8511b76-run-httpd\") pod \"ceilometer-0\" (UID: \"94d7bb01-64a7-442a-aee4-e958d8511b76\") " pod="openstack/ceilometer-0" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.282425 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e3b7f50f-961d-472f-9423-18ab69fcdaef-config-data\") pod \"horizon-64b577d9f-n8w7t\" (UID: \"e3b7f50f-961d-472f-9423-18ab69fcdaef\") " pod="openstack/horizon-64b577d9f-n8w7t" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.283340 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/94d7bb01-64a7-442a-aee4-e958d8511b76-log-httpd\") pod \"ceilometer-0\" (UID: \"94d7bb01-64a7-442a-aee4-e958d8511b76\") " pod="openstack/ceilometer-0" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.292771 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-r6nst"] Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.294171 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-r6nst" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.312237 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94d7bb01-64a7-442a-aee4-e958d8511b76-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"94d7bb01-64a7-442a-aee4-e958d8511b76\") " pod="openstack/ceilometer-0" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.312317 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-r6nst"] Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.313471 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6bf8aa3e-a7e7-4f0c-a542-d668a5623707-config-data\") pod \"cinder-db-sync-lg59m\" (UID: \"6bf8aa3e-a7e7-4f0c-a542-d668a5623707\") " pod="openstack/cinder-db-sync-lg59m" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.318554 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6bf8aa3e-a7e7-4f0c-a542-d668a5623707-db-sync-config-data\") pod \"cinder-db-sync-lg59m\" (UID: \"6bf8aa3e-a7e7-4f0c-a542-d668a5623707\") " pod="openstack/cinder-db-sync-lg59m" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.319627 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/94d7bb01-64a7-442a-aee4-e958d8511b76-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"94d7bb01-64a7-442a-aee4-e958d8511b76\") " pod="openstack/ceilometer-0" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.319827 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/733e94b9-65c2-406e-9ebe-244585e4e15d-config\") pod \"neutron-db-sync-n4s7g\" (UID: \"733e94b9-65c2-406e-9ebe-244585e4e15d\") " pod="openstack/neutron-db-sync-n4s7g" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.320074 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-hc559" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.320272 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6bf8aa3e-a7e7-4f0c-a542-d668a5623707-combined-ca-bundle\") pod \"cinder-db-sync-lg59m\" (UID: \"6bf8aa3e-a7e7-4f0c-a542-d668a5623707\") " pod="openstack/cinder-db-sync-lg59m" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.320327 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.320635 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/e3b7f50f-961d-472f-9423-18ab69fcdaef-horizon-secret-key\") pod \"horizon-64b577d9f-n8w7t\" (UID: \"e3b7f50f-961d-472f-9423-18ab69fcdaef\") " pod="openstack/horizon-64b577d9f-n8w7t" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.324243 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/733e94b9-65c2-406e-9ebe-244585e4e15d-combined-ca-bundle\") pod \"neutron-db-sync-n4s7g\" (UID: \"733e94b9-65c2-406e-9ebe-244585e4e15d\") " pod="openstack/neutron-db-sync-n4s7g" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.324304 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-6ddbf95fdc-64jxv"] Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.325580 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6ddbf95fdc-64jxv" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.329765 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/94d7bb01-64a7-442a-aee4-e958d8511b76-scripts\") pod \"ceilometer-0\" (UID: \"94d7bb01-64a7-442a-aee4-e958d8511b76\") " pod="openstack/ceilometer-0" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.330942 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6ddbf95fdc-64jxv"] Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.410943 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-2ss46"] Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.412141 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-2ss46" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.416640 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.416894 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.505519 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-vrlbv" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.512791 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cswpl\" (UniqueName: \"kubernetes.io/projected/e3b7f50f-961d-472f-9423-18ab69fcdaef-kube-api-access-cswpl\") pod \"horizon-64b577d9f-n8w7t\" (UID: \"e3b7f50f-961d-472f-9423-18ab69fcdaef\") " pod="openstack/horizon-64b577d9f-n8w7t" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.521320 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t9qz8\" (UniqueName: \"kubernetes.io/projected/6bf8aa3e-a7e7-4f0c-a542-d668a5623707-kube-api-access-t9qz8\") pod \"cinder-db-sync-lg59m\" (UID: \"6bf8aa3e-a7e7-4f0c-a542-d668a5623707\") " pod="openstack/cinder-db-sync-lg59m" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.522868 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3ee5cf5-1275-448f-8713-e2b5a0d24439-combined-ca-bundle\") pod \"barbican-db-sync-r6nst\" (UID: \"f3ee5cf5-1275-448f-8713-e2b5a0d24439\") " pod="openstack/barbican-db-sync-r6nst" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.522902 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eea1fd9b-942e-4d60-80ae-1fca761f5a32-logs\") pod \"horizon-6ddbf95fdc-64jxv\" (UID: \"eea1fd9b-942e-4d60-80ae-1fca761f5a32\") " pod="openstack/horizon-6ddbf95fdc-64jxv" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.522941 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/eea1fd9b-942e-4d60-80ae-1fca761f5a32-scripts\") pod \"horizon-6ddbf95fdc-64jxv\" (UID: \"eea1fd9b-942e-4d60-80ae-1fca761f5a32\") " pod="openstack/horizon-6ddbf95fdc-64jxv" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.522968 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rbbt8\" (UniqueName: \"kubernetes.io/projected/eea1fd9b-942e-4d60-80ae-1fca761f5a32-kube-api-access-rbbt8\") pod \"horizon-6ddbf95fdc-64jxv\" (UID: \"eea1fd9b-942e-4d60-80ae-1fca761f5a32\") " pod="openstack/horizon-6ddbf95fdc-64jxv" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.523005 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f3ee5cf5-1275-448f-8713-e2b5a0d24439-db-sync-config-data\") pod \"barbican-db-sync-r6nst\" (UID: \"f3ee5cf5-1275-448f-8713-e2b5a0d24439\") " pod="openstack/barbican-db-sync-r6nst" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.523069 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/eea1fd9b-942e-4d60-80ae-1fca761f5a32-horizon-secret-key\") pod \"horizon-6ddbf95fdc-64jxv\" (UID: \"eea1fd9b-942e-4d60-80ae-1fca761f5a32\") " pod="openstack/horizon-6ddbf95fdc-64jxv" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.523094 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/eea1fd9b-942e-4d60-80ae-1fca761f5a32-config-data\") pod \"horizon-6ddbf95fdc-64jxv\" (UID: \"eea1fd9b-942e-4d60-80ae-1fca761f5a32\") " pod="openstack/horizon-6ddbf95fdc-64jxv" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.523122 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9slr4\" (UniqueName: \"kubernetes.io/projected/f3ee5cf5-1275-448f-8713-e2b5a0d24439-kube-api-access-9slr4\") pod \"barbican-db-sync-r6nst\" (UID: \"f3ee5cf5-1275-448f-8713-e2b5a0d24439\") " pod="openstack/barbican-db-sync-r6nst" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.540654 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-2ss46"] Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.555659 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-lg59m" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.561310 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94d7bb01-64a7-442a-aee4-e958d8511b76-config-data\") pod \"ceilometer-0\" (UID: \"94d7bb01-64a7-442a-aee4-e958d8511b76\") " pod="openstack/ceilometer-0" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.603161 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rxhrx\" (UniqueName: \"kubernetes.io/projected/94d7bb01-64a7-442a-aee4-e958d8511b76-kube-api-access-rxhrx\") pod \"ceilometer-0\" (UID: \"94d7bb01-64a7-442a-aee4-e958d8511b76\") " pod="openstack/ceilometer-0" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.603948 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9mw6d\" (UniqueName: \"kubernetes.io/projected/733e94b9-65c2-406e-9ebe-244585e4e15d-kube-api-access-9mw6d\") pod \"neutron-db-sync-n4s7g\" (UID: \"733e94b9-65c2-406e-9ebe-244585e4e15d\") " pod="openstack/neutron-db-sync-n4s7g" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.628481 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2eb187f8-1892-4d72-b442-43592e75d489-scripts\") pod \"placement-db-sync-2ss46\" (UID: \"2eb187f8-1892-4d72-b442-43592e75d489\") " pod="openstack/placement-db-sync-2ss46" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.628541 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2eb187f8-1892-4d72-b442-43592e75d489-combined-ca-bundle\") pod \"placement-db-sync-2ss46\" (UID: \"2eb187f8-1892-4d72-b442-43592e75d489\") " pod="openstack/placement-db-sync-2ss46" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.628577 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/eea1fd9b-942e-4d60-80ae-1fca761f5a32-horizon-secret-key\") pod \"horizon-6ddbf95fdc-64jxv\" (UID: \"eea1fd9b-942e-4d60-80ae-1fca761f5a32\") " pod="openstack/horizon-6ddbf95fdc-64jxv" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.628617 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/eea1fd9b-942e-4d60-80ae-1fca761f5a32-config-data\") pod \"horizon-6ddbf95fdc-64jxv\" (UID: \"eea1fd9b-942e-4d60-80ae-1fca761f5a32\") " pod="openstack/horizon-6ddbf95fdc-64jxv" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.628645 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9slr4\" (UniqueName: \"kubernetes.io/projected/f3ee5cf5-1275-448f-8713-e2b5a0d24439-kube-api-access-9slr4\") pod \"barbican-db-sync-r6nst\" (UID: \"f3ee5cf5-1275-448f-8713-e2b5a0d24439\") " pod="openstack/barbican-db-sync-r6nst" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.628666 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2eb187f8-1892-4d72-b442-43592e75d489-logs\") pod \"placement-db-sync-2ss46\" (UID: \"2eb187f8-1892-4d72-b442-43592e75d489\") " pod="openstack/placement-db-sync-2ss46" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.628689 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2eb187f8-1892-4d72-b442-43592e75d489-config-data\") pod \"placement-db-sync-2ss46\" (UID: \"2eb187f8-1892-4d72-b442-43592e75d489\") " pod="openstack/placement-db-sync-2ss46" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.628725 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h867w\" (UniqueName: \"kubernetes.io/projected/2eb187f8-1892-4d72-b442-43592e75d489-kube-api-access-h867w\") pod \"placement-db-sync-2ss46\" (UID: \"2eb187f8-1892-4d72-b442-43592e75d489\") " pod="openstack/placement-db-sync-2ss46" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.628769 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3ee5cf5-1275-448f-8713-e2b5a0d24439-combined-ca-bundle\") pod \"barbican-db-sync-r6nst\" (UID: \"f3ee5cf5-1275-448f-8713-e2b5a0d24439\") " pod="openstack/barbican-db-sync-r6nst" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.628787 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eea1fd9b-942e-4d60-80ae-1fca761f5a32-logs\") pod \"horizon-6ddbf95fdc-64jxv\" (UID: \"eea1fd9b-942e-4d60-80ae-1fca761f5a32\") " pod="openstack/horizon-6ddbf95fdc-64jxv" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.628811 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/eea1fd9b-942e-4d60-80ae-1fca761f5a32-scripts\") pod \"horizon-6ddbf95fdc-64jxv\" (UID: \"eea1fd9b-942e-4d60-80ae-1fca761f5a32\") " pod="openstack/horizon-6ddbf95fdc-64jxv" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.629016 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rbbt8\" (UniqueName: \"kubernetes.io/projected/eea1fd9b-942e-4d60-80ae-1fca761f5a32-kube-api-access-rbbt8\") pod \"horizon-6ddbf95fdc-64jxv\" (UID: \"eea1fd9b-942e-4d60-80ae-1fca761f5a32\") " pod="openstack/horizon-6ddbf95fdc-64jxv" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.629042 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f3ee5cf5-1275-448f-8713-e2b5a0d24439-db-sync-config-data\") pod \"barbican-db-sync-r6nst\" (UID: \"f3ee5cf5-1275-448f-8713-e2b5a0d24439\") " pod="openstack/barbican-db-sync-r6nst" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.631272 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/eea1fd9b-942e-4d60-80ae-1fca761f5a32-config-data\") pod \"horizon-6ddbf95fdc-64jxv\" (UID: \"eea1fd9b-942e-4d60-80ae-1fca761f5a32\") " pod="openstack/horizon-6ddbf95fdc-64jxv" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.632188 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eea1fd9b-942e-4d60-80ae-1fca761f5a32-logs\") pod \"horizon-6ddbf95fdc-64jxv\" (UID: \"eea1fd9b-942e-4d60-80ae-1fca761f5a32\") " pod="openstack/horizon-6ddbf95fdc-64jxv" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.632927 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/eea1fd9b-942e-4d60-80ae-1fca761f5a32-horizon-secret-key\") pod \"horizon-6ddbf95fdc-64jxv\" (UID: \"eea1fd9b-942e-4d60-80ae-1fca761f5a32\") " pod="openstack/horizon-6ddbf95fdc-64jxv" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.633320 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/eea1fd9b-942e-4d60-80ae-1fca761f5a32-scripts\") pod \"horizon-6ddbf95fdc-64jxv\" (UID: \"eea1fd9b-942e-4d60-80ae-1fca761f5a32\") " pod="openstack/horizon-6ddbf95fdc-64jxv" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.636176 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f3ee5cf5-1275-448f-8713-e2b5a0d24439-db-sync-config-data\") pod \"barbican-db-sync-r6nst\" (UID: \"f3ee5cf5-1275-448f-8713-e2b5a0d24439\") " pod="openstack/barbican-db-sync-r6nst" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.637304 4789 generic.go:334] "Generic (PLEG): container finished" podID="b48e8695-47fa-4a3b-af9f-847184c147fd" containerID="904da12081f37b85040b7953040498e9abae99a90d641670f730e4916c15e0a9" exitCode=0 Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.637313 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3ee5cf5-1275-448f-8713-e2b5a0d24439-combined-ca-bundle\") pod \"barbican-db-sync-r6nst\" (UID: \"f3ee5cf5-1275-448f-8713-e2b5a0d24439\") " pod="openstack/barbican-db-sync-r6nst" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.637347 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-569d458467-hxxbh" event={"ID":"b48e8695-47fa-4a3b-af9f-847184c147fd","Type":"ContainerDied","Data":"904da12081f37b85040b7953040498e9abae99a90d641670f730e4916c15e0a9"} Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.642134 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-64b577d9f-n8w7t" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.649944 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b76c757b7-z85gc"] Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.664700 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-n4s7g" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.673516 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9slr4\" (UniqueName: \"kubernetes.io/projected/f3ee5cf5-1275-448f-8713-e2b5a0d24439-kube-api-access-9slr4\") pod \"barbican-db-sync-r6nst\" (UID: \"f3ee5cf5-1275-448f-8713-e2b5a0d24439\") " pod="openstack/barbican-db-sync-r6nst" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.678463 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rbbt8\" (UniqueName: \"kubernetes.io/projected/eea1fd9b-942e-4d60-80ae-1fca761f5a32-kube-api-access-rbbt8\") pod \"horizon-6ddbf95fdc-64jxv\" (UID: \"eea1fd9b-942e-4d60-80ae-1fca761f5a32\") " pod="openstack/horizon-6ddbf95fdc-64jxv" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.717106 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-66f4bdbdb7-tzwtz"] Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.719220 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-66f4bdbdb7-tzwtz" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.728682 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-66f4bdbdb7-tzwtz"] Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.730695 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2eb187f8-1892-4d72-b442-43592e75d489-scripts\") pod \"placement-db-sync-2ss46\" (UID: \"2eb187f8-1892-4d72-b442-43592e75d489\") " pod="openstack/placement-db-sync-2ss46" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.735909 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2eb187f8-1892-4d72-b442-43592e75d489-combined-ca-bundle\") pod \"placement-db-sync-2ss46\" (UID: \"2eb187f8-1892-4d72-b442-43592e75d489\") " pod="openstack/placement-db-sync-2ss46" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.736118 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2eb187f8-1892-4d72-b442-43592e75d489-logs\") pod \"placement-db-sync-2ss46\" (UID: \"2eb187f8-1892-4d72-b442-43592e75d489\") " pod="openstack/placement-db-sync-2ss46" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.734288 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2eb187f8-1892-4d72-b442-43592e75d489-scripts\") pod \"placement-db-sync-2ss46\" (UID: \"2eb187f8-1892-4d72-b442-43592e75d489\") " pod="openstack/placement-db-sync-2ss46" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.736248 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2eb187f8-1892-4d72-b442-43592e75d489-config-data\") pod \"placement-db-sync-2ss46\" (UID: \"2eb187f8-1892-4d72-b442-43592e75d489\") " pod="openstack/placement-db-sync-2ss46" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.736446 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h867w\" (UniqueName: \"kubernetes.io/projected/2eb187f8-1892-4d72-b442-43592e75d489-kube-api-access-h867w\") pod \"placement-db-sync-2ss46\" (UID: \"2eb187f8-1892-4d72-b442-43592e75d489\") " pod="openstack/placement-db-sync-2ss46" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.740813 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2eb187f8-1892-4d72-b442-43592e75d489-combined-ca-bundle\") pod \"placement-db-sync-2ss46\" (UID: \"2eb187f8-1892-4d72-b442-43592e75d489\") " pod="openstack/placement-db-sync-2ss46" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.741186 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2eb187f8-1892-4d72-b442-43592e75d489-logs\") pod \"placement-db-sync-2ss46\" (UID: \"2eb187f8-1892-4d72-b442-43592e75d489\") " pod="openstack/placement-db-sync-2ss46" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.745319 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2eb187f8-1892-4d72-b442-43592e75d489-config-data\") pod \"placement-db-sync-2ss46\" (UID: \"2eb187f8-1892-4d72-b442-43592e75d489\") " pod="openstack/placement-db-sync-2ss46" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.747779 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-r6nst" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.747880 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6ddbf95fdc-64jxv" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.768863 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h867w\" (UniqueName: \"kubernetes.io/projected/2eb187f8-1892-4d72-b442-43592e75d489-kube-api-access-h867w\") pod \"placement-db-sync-2ss46\" (UID: \"2eb187f8-1892-4d72-b442-43592e75d489\") " pod="openstack/placement-db-sync-2ss46" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.796931 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.821888 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-2ss46" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.838329 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/17e7e1f9-63dc-4917-b15c-c25cfca9ee85-ovsdbserver-nb\") pod \"dnsmasq-dns-66f4bdbdb7-tzwtz\" (UID: \"17e7e1f9-63dc-4917-b15c-c25cfca9ee85\") " pod="openstack/dnsmasq-dns-66f4bdbdb7-tzwtz" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.838377 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/17e7e1f9-63dc-4917-b15c-c25cfca9ee85-ovsdbserver-sb\") pod \"dnsmasq-dns-66f4bdbdb7-tzwtz\" (UID: \"17e7e1f9-63dc-4917-b15c-c25cfca9ee85\") " pod="openstack/dnsmasq-dns-66f4bdbdb7-tzwtz" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.838406 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6lj5x\" (UniqueName: \"kubernetes.io/projected/17e7e1f9-63dc-4917-b15c-c25cfca9ee85-kube-api-access-6lj5x\") pod \"dnsmasq-dns-66f4bdbdb7-tzwtz\" (UID: \"17e7e1f9-63dc-4917-b15c-c25cfca9ee85\") " pod="openstack/dnsmasq-dns-66f4bdbdb7-tzwtz" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.838426 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/17e7e1f9-63dc-4917-b15c-c25cfca9ee85-dns-svc\") pod \"dnsmasq-dns-66f4bdbdb7-tzwtz\" (UID: \"17e7e1f9-63dc-4917-b15c-c25cfca9ee85\") " pod="openstack/dnsmasq-dns-66f4bdbdb7-tzwtz" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.838486 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17e7e1f9-63dc-4917-b15c-c25cfca9ee85-config\") pod \"dnsmasq-dns-66f4bdbdb7-tzwtz\" (UID: \"17e7e1f9-63dc-4917-b15c-c25cfca9ee85\") " pod="openstack/dnsmasq-dns-66f4bdbdb7-tzwtz" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.838685 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-569d458467-hxxbh" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.942287 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rfg9x\" (UniqueName: \"kubernetes.io/projected/b48e8695-47fa-4a3b-af9f-847184c147fd-kube-api-access-rfg9x\") pod \"b48e8695-47fa-4a3b-af9f-847184c147fd\" (UID: \"b48e8695-47fa-4a3b-af9f-847184c147fd\") " Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.942590 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b48e8695-47fa-4a3b-af9f-847184c147fd-ovsdbserver-sb\") pod \"b48e8695-47fa-4a3b-af9f-847184c147fd\" (UID: \"b48e8695-47fa-4a3b-af9f-847184c147fd\") " Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.942648 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b48e8695-47fa-4a3b-af9f-847184c147fd-config\") pod \"b48e8695-47fa-4a3b-af9f-847184c147fd\" (UID: \"b48e8695-47fa-4a3b-af9f-847184c147fd\") " Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.942733 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b48e8695-47fa-4a3b-af9f-847184c147fd-ovsdbserver-nb\") pod \"b48e8695-47fa-4a3b-af9f-847184c147fd\" (UID: \"b48e8695-47fa-4a3b-af9f-847184c147fd\") " Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.942789 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b48e8695-47fa-4a3b-af9f-847184c147fd-dns-svc\") pod \"b48e8695-47fa-4a3b-af9f-847184c147fd\" (UID: \"b48e8695-47fa-4a3b-af9f-847184c147fd\") " Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.943049 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/17e7e1f9-63dc-4917-b15c-c25cfca9ee85-ovsdbserver-nb\") pod \"dnsmasq-dns-66f4bdbdb7-tzwtz\" (UID: \"17e7e1f9-63dc-4917-b15c-c25cfca9ee85\") " pod="openstack/dnsmasq-dns-66f4bdbdb7-tzwtz" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.943070 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/17e7e1f9-63dc-4917-b15c-c25cfca9ee85-ovsdbserver-sb\") pod \"dnsmasq-dns-66f4bdbdb7-tzwtz\" (UID: \"17e7e1f9-63dc-4917-b15c-c25cfca9ee85\") " pod="openstack/dnsmasq-dns-66f4bdbdb7-tzwtz" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.943092 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6lj5x\" (UniqueName: \"kubernetes.io/projected/17e7e1f9-63dc-4917-b15c-c25cfca9ee85-kube-api-access-6lj5x\") pod \"dnsmasq-dns-66f4bdbdb7-tzwtz\" (UID: \"17e7e1f9-63dc-4917-b15c-c25cfca9ee85\") " pod="openstack/dnsmasq-dns-66f4bdbdb7-tzwtz" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.943112 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/17e7e1f9-63dc-4917-b15c-c25cfca9ee85-dns-svc\") pod \"dnsmasq-dns-66f4bdbdb7-tzwtz\" (UID: \"17e7e1f9-63dc-4917-b15c-c25cfca9ee85\") " pod="openstack/dnsmasq-dns-66f4bdbdb7-tzwtz" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.943157 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17e7e1f9-63dc-4917-b15c-c25cfca9ee85-config\") pod \"dnsmasq-dns-66f4bdbdb7-tzwtz\" (UID: \"17e7e1f9-63dc-4917-b15c-c25cfca9ee85\") " pod="openstack/dnsmasq-dns-66f4bdbdb7-tzwtz" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.944184 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17e7e1f9-63dc-4917-b15c-c25cfca9ee85-config\") pod \"dnsmasq-dns-66f4bdbdb7-tzwtz\" (UID: \"17e7e1f9-63dc-4917-b15c-c25cfca9ee85\") " pod="openstack/dnsmasq-dns-66f4bdbdb7-tzwtz" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.952174 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/17e7e1f9-63dc-4917-b15c-c25cfca9ee85-ovsdbserver-nb\") pod \"dnsmasq-dns-66f4bdbdb7-tzwtz\" (UID: \"17e7e1f9-63dc-4917-b15c-c25cfca9ee85\") " pod="openstack/dnsmasq-dns-66f4bdbdb7-tzwtz" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.953532 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/17e7e1f9-63dc-4917-b15c-c25cfca9ee85-ovsdbserver-sb\") pod \"dnsmasq-dns-66f4bdbdb7-tzwtz\" (UID: \"17e7e1f9-63dc-4917-b15c-c25cfca9ee85\") " pod="openstack/dnsmasq-dns-66f4bdbdb7-tzwtz" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.954292 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/17e7e1f9-63dc-4917-b15c-c25cfca9ee85-dns-svc\") pod \"dnsmasq-dns-66f4bdbdb7-tzwtz\" (UID: \"17e7e1f9-63dc-4917-b15c-c25cfca9ee85\") " pod="openstack/dnsmasq-dns-66f4bdbdb7-tzwtz" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.973228 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b48e8695-47fa-4a3b-af9f-847184c147fd-kube-api-access-rfg9x" (OuterVolumeSpecName: "kube-api-access-rfg9x") pod "b48e8695-47fa-4a3b-af9f-847184c147fd" (UID: "b48e8695-47fa-4a3b-af9f-847184c147fd"). InnerVolumeSpecName "kube-api-access-rfg9x". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:15:49 crc kubenswrapper[4789]: I1122 08:15:49.996724 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6lj5x\" (UniqueName: \"kubernetes.io/projected/17e7e1f9-63dc-4917-b15c-c25cfca9ee85-kube-api-access-6lj5x\") pod \"dnsmasq-dns-66f4bdbdb7-tzwtz\" (UID: \"17e7e1f9-63dc-4917-b15c-c25cfca9ee85\") " pod="openstack/dnsmasq-dns-66f4bdbdb7-tzwtz" Nov 22 08:15:50 crc kubenswrapper[4789]: I1122 08:15:50.028736 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b48e8695-47fa-4a3b-af9f-847184c147fd-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b48e8695-47fa-4a3b-af9f-847184c147fd" (UID: "b48e8695-47fa-4a3b-af9f-847184c147fd"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:15:50 crc kubenswrapper[4789]: I1122 08:15:50.044860 4789 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b48e8695-47fa-4a3b-af9f-847184c147fd-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 22 08:15:50 crc kubenswrapper[4789]: I1122 08:15:50.044890 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rfg9x\" (UniqueName: \"kubernetes.io/projected/b48e8695-47fa-4a3b-af9f-847184c147fd-kube-api-access-rfg9x\") on node \"crc\" DevicePath \"\"" Nov 22 08:15:50 crc kubenswrapper[4789]: I1122 08:15:50.052299 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b48e8695-47fa-4a3b-af9f-847184c147fd-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b48e8695-47fa-4a3b-af9f-847184c147fd" (UID: "b48e8695-47fa-4a3b-af9f-847184c147fd"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:15:50 crc kubenswrapper[4789]: I1122 08:15:50.061461 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-66f4bdbdb7-tzwtz" Nov 22 08:15:50 crc kubenswrapper[4789]: I1122 08:15:50.065884 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b48e8695-47fa-4a3b-af9f-847184c147fd-config" (OuterVolumeSpecName: "config") pod "b48e8695-47fa-4a3b-af9f-847184c147fd" (UID: "b48e8695-47fa-4a3b-af9f-847184c147fd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:15:50 crc kubenswrapper[4789]: I1122 08:15:50.066841 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b48e8695-47fa-4a3b-af9f-847184c147fd-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b48e8695-47fa-4a3b-af9f-847184c147fd" (UID: "b48e8695-47fa-4a3b-af9f-847184c147fd"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:15:50 crc kubenswrapper[4789]: I1122 08:15:50.118862 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b76c757b7-z85gc"] Nov 22 08:15:50 crc kubenswrapper[4789]: I1122 08:15:50.146693 4789 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b48e8695-47fa-4a3b-af9f-847184c147fd-config\") on node \"crc\" DevicePath \"\"" Nov 22 08:15:50 crc kubenswrapper[4789]: I1122 08:15:50.146729 4789 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b48e8695-47fa-4a3b-af9f-847184c147fd-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 22 08:15:50 crc kubenswrapper[4789]: I1122 08:15:50.146739 4789 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b48e8695-47fa-4a3b-af9f-847184c147fd-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 22 08:15:50 crc kubenswrapper[4789]: I1122 08:15:50.299614 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-64b577d9f-n8w7t"] Nov 22 08:15:50 crc kubenswrapper[4789]: I1122 08:15:50.343448 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-9ts56"] Nov 22 08:15:50 crc kubenswrapper[4789]: I1122 08:15:50.462432 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-lg59m"] Nov 22 08:15:50 crc kubenswrapper[4789]: W1122 08:15:50.469393 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6bf8aa3e_a7e7_4f0c_a542_d668a5623707.slice/crio-835d05ccf6aeb44f1550e11dc4e9f94affbfd2055349d047da78d8dda6b514d6 WatchSource:0}: Error finding container 835d05ccf6aeb44f1550e11dc4e9f94affbfd2055349d047da78d8dda6b514d6: Status 404 returned error can't find the container with id 835d05ccf6aeb44f1550e11dc4e9f94affbfd2055349d047da78d8dda6b514d6 Nov 22 08:15:50 crc kubenswrapper[4789]: I1122 08:15:50.647259 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-64b577d9f-n8w7t" event={"ID":"e3b7f50f-961d-472f-9423-18ab69fcdaef","Type":"ContainerStarted","Data":"43b316398a79a1aad6168c1db602fd3d3dc6f7a04a903fb03d3ed9efdd69fcfd"} Nov 22 08:15:50 crc kubenswrapper[4789]: I1122 08:15:50.650078 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-n4s7g"] Nov 22 08:15:50 crc kubenswrapper[4789]: I1122 08:15:50.651956 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-lg59m" event={"ID":"6bf8aa3e-a7e7-4f0c-a542-d668a5623707","Type":"ContainerStarted","Data":"835d05ccf6aeb44f1550e11dc4e9f94affbfd2055349d047da78d8dda6b514d6"} Nov 22 08:15:50 crc kubenswrapper[4789]: I1122 08:15:50.653528 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b76c757b7-z85gc" event={"ID":"1a8be067-2aaa-4459-8dfb-2ebc7f21b258","Type":"ContainerStarted","Data":"abc41336de0da5305e37a1f3b3f0bfb48c521b8dc0a92a46179742a15a7826a5"} Nov 22 08:15:50 crc kubenswrapper[4789]: I1122 08:15:50.656257 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-9ts56" event={"ID":"09e8392f-f97d-4922-be5a-fb6229549344","Type":"ContainerStarted","Data":"7ebc87cea54f0af771db08c43b687c409d8174a725e89ad7b8e62fffbae9a118"} Nov 22 08:15:50 crc kubenswrapper[4789]: I1122 08:15:50.661856 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-569d458467-hxxbh" event={"ID":"b48e8695-47fa-4a3b-af9f-847184c147fd","Type":"ContainerDied","Data":"b7b2f39d7a7a2aedce672cb740da7f0ffd2de1c748289746f031ddeef52a0df0"} Nov 22 08:15:50 crc kubenswrapper[4789]: I1122 08:15:50.661910 4789 scope.go:117] "RemoveContainer" containerID="904da12081f37b85040b7953040498e9abae99a90d641670f730e4916c15e0a9" Nov 22 08:15:50 crc kubenswrapper[4789]: I1122 08:15:50.662123 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-569d458467-hxxbh" Nov 22 08:15:50 crc kubenswrapper[4789]: I1122 08:15:50.685867 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:15:50 crc kubenswrapper[4789]: I1122 08:15:50.708594 4789 scope.go:117] "RemoveContainer" containerID="6ac1e3b8a17f8aa8ccc8ad6418df1dabe4253068f96afc0d213f79064b5c2e23" Nov 22 08:15:50 crc kubenswrapper[4789]: I1122 08:15:50.710251 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-569d458467-hxxbh"] Nov 22 08:15:50 crc kubenswrapper[4789]: I1122 08:15:50.732534 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-569d458467-hxxbh"] Nov 22 08:15:50 crc kubenswrapper[4789]: I1122 08:15:50.753855 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-2ss46"] Nov 22 08:15:50 crc kubenswrapper[4789]: I1122 08:15:50.867384 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-r6nst"] Nov 22 08:15:50 crc kubenswrapper[4789]: W1122 08:15:50.875645 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf3ee5cf5_1275_448f_8713_e2b5a0d24439.slice/crio-09be797b1e863f926c538f47c129fcdbb99348630904565dc9c8607260421481 WatchSource:0}: Error finding container 09be797b1e863f926c538f47c129fcdbb99348630904565dc9c8607260421481: Status 404 returned error can't find the container with id 09be797b1e863f926c538f47c129fcdbb99348630904565dc9c8607260421481 Nov 22 08:15:50 crc kubenswrapper[4789]: I1122 08:15:50.887177 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6ddbf95fdc-64jxv"] Nov 22 08:15:50 crc kubenswrapper[4789]: I1122 08:15:50.895207 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-66f4bdbdb7-tzwtz"] Nov 22 08:15:51 crc kubenswrapper[4789]: I1122 08:15:51.412334 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-64b577d9f-n8w7t"] Nov 22 08:15:51 crc kubenswrapper[4789]: I1122 08:15:51.459638 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:15:51 crc kubenswrapper[4789]: I1122 08:15:51.467207 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-6d4f97b8d9-fqgm7"] Nov 22 08:15:51 crc kubenswrapper[4789]: E1122 08:15:51.467529 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b48e8695-47fa-4a3b-af9f-847184c147fd" containerName="init" Nov 22 08:15:51 crc kubenswrapper[4789]: I1122 08:15:51.467544 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="b48e8695-47fa-4a3b-af9f-847184c147fd" containerName="init" Nov 22 08:15:51 crc kubenswrapper[4789]: E1122 08:15:51.467566 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b48e8695-47fa-4a3b-af9f-847184c147fd" containerName="dnsmasq-dns" Nov 22 08:15:51 crc kubenswrapper[4789]: I1122 08:15:51.467573 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="b48e8695-47fa-4a3b-af9f-847184c147fd" containerName="dnsmasq-dns" Nov 22 08:15:51 crc kubenswrapper[4789]: I1122 08:15:51.467718 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="b48e8695-47fa-4a3b-af9f-847184c147fd" containerName="dnsmasq-dns" Nov 22 08:15:51 crc kubenswrapper[4789]: I1122 08:15:51.468507 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6d4f97b8d9-fqgm7" Nov 22 08:15:51 crc kubenswrapper[4789]: I1122 08:15:51.503394 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6d4f97b8d9-fqgm7"] Nov 22 08:15:51 crc kubenswrapper[4789]: I1122 08:15:51.577564 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/61bf72d1-ea9f-4ad5-9dd1-7fdb4a65e8e4-scripts\") pod \"horizon-6d4f97b8d9-fqgm7\" (UID: \"61bf72d1-ea9f-4ad5-9dd1-7fdb4a65e8e4\") " pod="openstack/horizon-6d4f97b8d9-fqgm7" Nov 22 08:15:51 crc kubenswrapper[4789]: I1122 08:15:51.577661 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-grm5s\" (UniqueName: \"kubernetes.io/projected/61bf72d1-ea9f-4ad5-9dd1-7fdb4a65e8e4-kube-api-access-grm5s\") pod \"horizon-6d4f97b8d9-fqgm7\" (UID: \"61bf72d1-ea9f-4ad5-9dd1-7fdb4a65e8e4\") " pod="openstack/horizon-6d4f97b8d9-fqgm7" Nov 22 08:15:51 crc kubenswrapper[4789]: I1122 08:15:51.577723 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/61bf72d1-ea9f-4ad5-9dd1-7fdb4a65e8e4-horizon-secret-key\") pod \"horizon-6d4f97b8d9-fqgm7\" (UID: \"61bf72d1-ea9f-4ad5-9dd1-7fdb4a65e8e4\") " pod="openstack/horizon-6d4f97b8d9-fqgm7" Nov 22 08:15:51 crc kubenswrapper[4789]: I1122 08:15:51.577782 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/61bf72d1-ea9f-4ad5-9dd1-7fdb4a65e8e4-logs\") pod \"horizon-6d4f97b8d9-fqgm7\" (UID: \"61bf72d1-ea9f-4ad5-9dd1-7fdb4a65e8e4\") " pod="openstack/horizon-6d4f97b8d9-fqgm7" Nov 22 08:15:51 crc kubenswrapper[4789]: I1122 08:15:51.577819 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/61bf72d1-ea9f-4ad5-9dd1-7fdb4a65e8e4-config-data\") pod \"horizon-6d4f97b8d9-fqgm7\" (UID: \"61bf72d1-ea9f-4ad5-9dd1-7fdb4a65e8e4\") " pod="openstack/horizon-6d4f97b8d9-fqgm7" Nov 22 08:15:51 crc kubenswrapper[4789]: I1122 08:15:51.679798 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/61bf72d1-ea9f-4ad5-9dd1-7fdb4a65e8e4-horizon-secret-key\") pod \"horizon-6d4f97b8d9-fqgm7\" (UID: \"61bf72d1-ea9f-4ad5-9dd1-7fdb4a65e8e4\") " pod="openstack/horizon-6d4f97b8d9-fqgm7" Nov 22 08:15:51 crc kubenswrapper[4789]: I1122 08:15:51.679866 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/61bf72d1-ea9f-4ad5-9dd1-7fdb4a65e8e4-logs\") pod \"horizon-6d4f97b8d9-fqgm7\" (UID: \"61bf72d1-ea9f-4ad5-9dd1-7fdb4a65e8e4\") " pod="openstack/horizon-6d4f97b8d9-fqgm7" Nov 22 08:15:51 crc kubenswrapper[4789]: I1122 08:15:51.679901 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/61bf72d1-ea9f-4ad5-9dd1-7fdb4a65e8e4-config-data\") pod \"horizon-6d4f97b8d9-fqgm7\" (UID: \"61bf72d1-ea9f-4ad5-9dd1-7fdb4a65e8e4\") " pod="openstack/horizon-6d4f97b8d9-fqgm7" Nov 22 08:15:51 crc kubenswrapper[4789]: I1122 08:15:51.679941 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/61bf72d1-ea9f-4ad5-9dd1-7fdb4a65e8e4-scripts\") pod \"horizon-6d4f97b8d9-fqgm7\" (UID: \"61bf72d1-ea9f-4ad5-9dd1-7fdb4a65e8e4\") " pod="openstack/horizon-6d4f97b8d9-fqgm7" Nov 22 08:15:51 crc kubenswrapper[4789]: I1122 08:15:51.679999 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-grm5s\" (UniqueName: \"kubernetes.io/projected/61bf72d1-ea9f-4ad5-9dd1-7fdb4a65e8e4-kube-api-access-grm5s\") pod \"horizon-6d4f97b8d9-fqgm7\" (UID: \"61bf72d1-ea9f-4ad5-9dd1-7fdb4a65e8e4\") " pod="openstack/horizon-6d4f97b8d9-fqgm7" Nov 22 08:15:51 crc kubenswrapper[4789]: I1122 08:15:51.682343 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/61bf72d1-ea9f-4ad5-9dd1-7fdb4a65e8e4-logs\") pod \"horizon-6d4f97b8d9-fqgm7\" (UID: \"61bf72d1-ea9f-4ad5-9dd1-7fdb4a65e8e4\") " pod="openstack/horizon-6d4f97b8d9-fqgm7" Nov 22 08:15:51 crc kubenswrapper[4789]: I1122 08:15:51.682536 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/61bf72d1-ea9f-4ad5-9dd1-7fdb4a65e8e4-scripts\") pod \"horizon-6d4f97b8d9-fqgm7\" (UID: \"61bf72d1-ea9f-4ad5-9dd1-7fdb4a65e8e4\") " pod="openstack/horizon-6d4f97b8d9-fqgm7" Nov 22 08:15:51 crc kubenswrapper[4789]: I1122 08:15:51.683931 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/61bf72d1-ea9f-4ad5-9dd1-7fdb4a65e8e4-config-data\") pod \"horizon-6d4f97b8d9-fqgm7\" (UID: \"61bf72d1-ea9f-4ad5-9dd1-7fdb4a65e8e4\") " pod="openstack/horizon-6d4f97b8d9-fqgm7" Nov 22 08:15:51 crc kubenswrapper[4789]: I1122 08:15:51.696929 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/61bf72d1-ea9f-4ad5-9dd1-7fdb4a65e8e4-horizon-secret-key\") pod \"horizon-6d4f97b8d9-fqgm7\" (UID: \"61bf72d1-ea9f-4ad5-9dd1-7fdb4a65e8e4\") " pod="openstack/horizon-6d4f97b8d9-fqgm7" Nov 22 08:15:51 crc kubenswrapper[4789]: I1122 08:15:51.704954 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-n4s7g" event={"ID":"733e94b9-65c2-406e-9ebe-244585e4e15d","Type":"ContainerStarted","Data":"af51430767cb5d1d2c0d8f8f42df0fa11f0930283020526b5643ab6192006ea8"} Nov 22 08:15:51 crc kubenswrapper[4789]: I1122 08:15:51.713500 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-grm5s\" (UniqueName: \"kubernetes.io/projected/61bf72d1-ea9f-4ad5-9dd1-7fdb4a65e8e4-kube-api-access-grm5s\") pod \"horizon-6d4f97b8d9-fqgm7\" (UID: \"61bf72d1-ea9f-4ad5-9dd1-7fdb4a65e8e4\") " pod="openstack/horizon-6d4f97b8d9-fqgm7" Nov 22 08:15:51 crc kubenswrapper[4789]: I1122 08:15:51.749338 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-r6nst" event={"ID":"f3ee5cf5-1275-448f-8713-e2b5a0d24439","Type":"ContainerStarted","Data":"09be797b1e863f926c538f47c129fcdbb99348630904565dc9c8607260421481"} Nov 22 08:15:51 crc kubenswrapper[4789]: I1122 08:15:51.750365 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"94d7bb01-64a7-442a-aee4-e958d8511b76","Type":"ContainerStarted","Data":"414dc84d92b52f295bade29141d15ec375784cdc75fb3cbb8ce807e50fe81343"} Nov 22 08:15:51 crc kubenswrapper[4789]: I1122 08:15:51.759179 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-66f4bdbdb7-tzwtz" event={"ID":"17e7e1f9-63dc-4917-b15c-c25cfca9ee85","Type":"ContainerStarted","Data":"166a96122b7fd3d01ff1eaedc0e61753faac0db24d9c0ccb069993333588ad5f"} Nov 22 08:15:51 crc kubenswrapper[4789]: I1122 08:15:51.762359 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-2ss46" event={"ID":"2eb187f8-1892-4d72-b442-43592e75d489","Type":"ContainerStarted","Data":"d7c6edc5fd50c8ead7f54244e5c18905bace2602ed1c684bbfd4be939bcd4f07"} Nov 22 08:15:51 crc kubenswrapper[4789]: I1122 08:15:51.764039 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6ddbf95fdc-64jxv" event={"ID":"eea1fd9b-942e-4d60-80ae-1fca761f5a32","Type":"ContainerStarted","Data":"6d889b3a7d37cdfcf2fe5cc4b099872c69d7353f7064764a959103e8f920d9cc"} Nov 22 08:15:51 crc kubenswrapper[4789]: I1122 08:15:51.787882 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6d4f97b8d9-fqgm7" Nov 22 08:15:51 crc kubenswrapper[4789]: I1122 08:15:51.975584 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b48e8695-47fa-4a3b-af9f-847184c147fd" path="/var/lib/kubelet/pods/b48e8695-47fa-4a3b-af9f-847184c147fd/volumes" Nov 22 08:15:52 crc kubenswrapper[4789]: I1122 08:15:52.250721 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6d4f97b8d9-fqgm7"] Nov 22 08:15:52 crc kubenswrapper[4789]: I1122 08:15:52.775914 4789 generic.go:334] "Generic (PLEG): container finished" podID="1a8be067-2aaa-4459-8dfb-2ebc7f21b258" containerID="8be0ef67e7a4fe1d91efb9e116889124eed58a4b1faeb11342c053063877796e" exitCode=0 Nov 22 08:15:52 crc kubenswrapper[4789]: I1122 08:15:52.775994 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b76c757b7-z85gc" event={"ID":"1a8be067-2aaa-4459-8dfb-2ebc7f21b258","Type":"ContainerDied","Data":"8be0ef67e7a4fe1d91efb9e116889124eed58a4b1faeb11342c053063877796e"} Nov 22 08:15:52 crc kubenswrapper[4789]: I1122 08:15:52.778215 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-9ts56" event={"ID":"09e8392f-f97d-4922-be5a-fb6229549344","Type":"ContainerStarted","Data":"83b2ef9fe1addd84e58855d1a0866dfb7dd425eeeeff2289d91aa552b31f47eb"} Nov 22 08:15:52 crc kubenswrapper[4789]: I1122 08:15:52.782451 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6d4f97b8d9-fqgm7" event={"ID":"61bf72d1-ea9f-4ad5-9dd1-7fdb4a65e8e4","Type":"ContainerStarted","Data":"1454f79ee94e414a3a4d93406614952af4ebf9b7d7bdc182754538a0728d40c1"} Nov 22 08:15:52 crc kubenswrapper[4789]: I1122 08:15:52.785242 4789 generic.go:334] "Generic (PLEG): container finished" podID="17e7e1f9-63dc-4917-b15c-c25cfca9ee85" containerID="849eea141aca67d8500974a142cb8c034c08a66d073fdef2c2686355a607fda1" exitCode=0 Nov 22 08:15:52 crc kubenswrapper[4789]: I1122 08:15:52.785320 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-66f4bdbdb7-tzwtz" event={"ID":"17e7e1f9-63dc-4917-b15c-c25cfca9ee85","Type":"ContainerDied","Data":"849eea141aca67d8500974a142cb8c034c08a66d073fdef2c2686355a607fda1"} Nov 22 08:15:52 crc kubenswrapper[4789]: I1122 08:15:52.792279 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-n4s7g" event={"ID":"733e94b9-65c2-406e-9ebe-244585e4e15d","Type":"ContainerStarted","Data":"aa73593836103e7d3c745c9dd3a425820bddd9b5607f5bdc6e6dfaa4e6e0506f"} Nov 22 08:15:52 crc kubenswrapper[4789]: I1122 08:15:52.826182 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-n4s7g" podStartSLOduration=4.826166207 podStartE2EDuration="4.826166207s" podCreationTimestamp="2025-11-22 08:15:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:15:52.817271515 +0000 UTC m=+1167.051671788" watchObservedRunningTime="2025-11-22 08:15:52.826166207 +0000 UTC m=+1167.060566480" Nov 22 08:15:52 crc kubenswrapper[4789]: I1122 08:15:52.839017 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-9ts56" podStartSLOduration=4.838961065 podStartE2EDuration="4.838961065s" podCreationTimestamp="2025-11-22 08:15:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:15:52.834685209 +0000 UTC m=+1167.069085492" watchObservedRunningTime="2025-11-22 08:15:52.838961065 +0000 UTC m=+1167.073361338" Nov 22 08:15:53 crc kubenswrapper[4789]: I1122 08:15:53.210686 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b76c757b7-z85gc" Nov 22 08:15:53 crc kubenswrapper[4789]: I1122 08:15:53.317871 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1a8be067-2aaa-4459-8dfb-2ebc7f21b258-config\") pod \"1a8be067-2aaa-4459-8dfb-2ebc7f21b258\" (UID: \"1a8be067-2aaa-4459-8dfb-2ebc7f21b258\") " Nov 22 08:15:53 crc kubenswrapper[4789]: I1122 08:15:53.317929 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1a8be067-2aaa-4459-8dfb-2ebc7f21b258-dns-svc\") pod \"1a8be067-2aaa-4459-8dfb-2ebc7f21b258\" (UID: \"1a8be067-2aaa-4459-8dfb-2ebc7f21b258\") " Nov 22 08:15:53 crc kubenswrapper[4789]: I1122 08:15:53.318053 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1a8be067-2aaa-4459-8dfb-2ebc7f21b258-ovsdbserver-nb\") pod \"1a8be067-2aaa-4459-8dfb-2ebc7f21b258\" (UID: \"1a8be067-2aaa-4459-8dfb-2ebc7f21b258\") " Nov 22 08:15:53 crc kubenswrapper[4789]: I1122 08:15:53.318103 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1a8be067-2aaa-4459-8dfb-2ebc7f21b258-ovsdbserver-sb\") pod \"1a8be067-2aaa-4459-8dfb-2ebc7f21b258\" (UID: \"1a8be067-2aaa-4459-8dfb-2ebc7f21b258\") " Nov 22 08:15:53 crc kubenswrapper[4789]: I1122 08:15:53.318160 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h4brh\" (UniqueName: \"kubernetes.io/projected/1a8be067-2aaa-4459-8dfb-2ebc7f21b258-kube-api-access-h4brh\") pod \"1a8be067-2aaa-4459-8dfb-2ebc7f21b258\" (UID: \"1a8be067-2aaa-4459-8dfb-2ebc7f21b258\") " Nov 22 08:15:53 crc kubenswrapper[4789]: I1122 08:15:53.358057 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a8be067-2aaa-4459-8dfb-2ebc7f21b258-kube-api-access-h4brh" (OuterVolumeSpecName: "kube-api-access-h4brh") pod "1a8be067-2aaa-4459-8dfb-2ebc7f21b258" (UID: "1a8be067-2aaa-4459-8dfb-2ebc7f21b258"). InnerVolumeSpecName "kube-api-access-h4brh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:15:53 crc kubenswrapper[4789]: I1122 08:15:53.359737 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1a8be067-2aaa-4459-8dfb-2ebc7f21b258-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "1a8be067-2aaa-4459-8dfb-2ebc7f21b258" (UID: "1a8be067-2aaa-4459-8dfb-2ebc7f21b258"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:15:53 crc kubenswrapper[4789]: I1122 08:15:53.359835 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1a8be067-2aaa-4459-8dfb-2ebc7f21b258-config" (OuterVolumeSpecName: "config") pod "1a8be067-2aaa-4459-8dfb-2ebc7f21b258" (UID: "1a8be067-2aaa-4459-8dfb-2ebc7f21b258"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:15:53 crc kubenswrapper[4789]: I1122 08:15:53.359849 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1a8be067-2aaa-4459-8dfb-2ebc7f21b258-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "1a8be067-2aaa-4459-8dfb-2ebc7f21b258" (UID: "1a8be067-2aaa-4459-8dfb-2ebc7f21b258"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:15:53 crc kubenswrapper[4789]: I1122 08:15:53.364413 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1a8be067-2aaa-4459-8dfb-2ebc7f21b258-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "1a8be067-2aaa-4459-8dfb-2ebc7f21b258" (UID: "1a8be067-2aaa-4459-8dfb-2ebc7f21b258"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:15:53 crc kubenswrapper[4789]: I1122 08:15:53.420265 4789 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1a8be067-2aaa-4459-8dfb-2ebc7f21b258-config\") on node \"crc\" DevicePath \"\"" Nov 22 08:15:53 crc kubenswrapper[4789]: I1122 08:15:53.420307 4789 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1a8be067-2aaa-4459-8dfb-2ebc7f21b258-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 22 08:15:53 crc kubenswrapper[4789]: I1122 08:15:53.420319 4789 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1a8be067-2aaa-4459-8dfb-2ebc7f21b258-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 22 08:15:53 crc kubenswrapper[4789]: I1122 08:15:53.420333 4789 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1a8be067-2aaa-4459-8dfb-2ebc7f21b258-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 22 08:15:53 crc kubenswrapper[4789]: I1122 08:15:53.420345 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h4brh\" (UniqueName: \"kubernetes.io/projected/1a8be067-2aaa-4459-8dfb-2ebc7f21b258-kube-api-access-h4brh\") on node \"crc\" DevicePath \"\"" Nov 22 08:15:53 crc kubenswrapper[4789]: I1122 08:15:53.806571 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b76c757b7-z85gc" event={"ID":"1a8be067-2aaa-4459-8dfb-2ebc7f21b258","Type":"ContainerDied","Data":"abc41336de0da5305e37a1f3b3f0bfb48c521b8dc0a92a46179742a15a7826a5"} Nov 22 08:15:53 crc kubenswrapper[4789]: I1122 08:15:53.806641 4789 scope.go:117] "RemoveContainer" containerID="8be0ef67e7a4fe1d91efb9e116889124eed58a4b1faeb11342c053063877796e" Nov 22 08:15:53 crc kubenswrapper[4789]: I1122 08:15:53.806786 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b76c757b7-z85gc" Nov 22 08:15:53 crc kubenswrapper[4789]: I1122 08:15:53.835116 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-66f4bdbdb7-tzwtz" event={"ID":"17e7e1f9-63dc-4917-b15c-c25cfca9ee85","Type":"ContainerStarted","Data":"da986e372351234f811d9c57b91629f3b44dd049dff52d99223a0d6584a78bfc"} Nov 22 08:15:53 crc kubenswrapper[4789]: I1122 08:15:53.835592 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-66f4bdbdb7-tzwtz" Nov 22 08:15:53 crc kubenswrapper[4789]: I1122 08:15:53.891278 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-66f4bdbdb7-tzwtz" podStartSLOduration=4.891258754 podStartE2EDuration="4.891258754s" podCreationTimestamp="2025-11-22 08:15:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:15:53.861085342 +0000 UTC m=+1168.095485615" watchObservedRunningTime="2025-11-22 08:15:53.891258754 +0000 UTC m=+1168.125659027" Nov 22 08:15:53 crc kubenswrapper[4789]: I1122 08:15:53.939952 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b76c757b7-z85gc"] Nov 22 08:15:53 crc kubenswrapper[4789]: I1122 08:15:53.948503 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-b76c757b7-z85gc"] Nov 22 08:15:53 crc kubenswrapper[4789]: I1122 08:15:53.985997 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1a8be067-2aaa-4459-8dfb-2ebc7f21b258" path="/var/lib/kubelet/pods/1a8be067-2aaa-4459-8dfb-2ebc7f21b258/volumes" Nov 22 08:15:54 crc kubenswrapper[4789]: E1122 08:15:54.085286 4789 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1a8be067_2aaa_4459_8dfb_2ebc7f21b258.slice\": RecentStats: unable to find data in memory cache]" Nov 22 08:15:56 crc kubenswrapper[4789]: I1122 08:15:56.869849 4789 generic.go:334] "Generic (PLEG): container finished" podID="09e8392f-f97d-4922-be5a-fb6229549344" containerID="83b2ef9fe1addd84e58855d1a0866dfb7dd425eeeeff2289d91aa552b31f47eb" exitCode=0 Nov 22 08:15:56 crc kubenswrapper[4789]: I1122 08:15:56.869895 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-9ts56" event={"ID":"09e8392f-f97d-4922-be5a-fb6229549344","Type":"ContainerDied","Data":"83b2ef9fe1addd84e58855d1a0866dfb7dd425eeeeff2289d91aa552b31f47eb"} Nov 22 08:15:58 crc kubenswrapper[4789]: I1122 08:15:58.076659 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-6ddbf95fdc-64jxv"] Nov 22 08:15:58 crc kubenswrapper[4789]: I1122 08:15:58.109902 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-5b56df9788-wsmhq"] Nov 22 08:15:58 crc kubenswrapper[4789]: E1122 08:15:58.110378 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a8be067-2aaa-4459-8dfb-2ebc7f21b258" containerName="init" Nov 22 08:15:58 crc kubenswrapper[4789]: I1122 08:15:58.110396 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a8be067-2aaa-4459-8dfb-2ebc7f21b258" containerName="init" Nov 22 08:15:58 crc kubenswrapper[4789]: I1122 08:15:58.110648 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a8be067-2aaa-4459-8dfb-2ebc7f21b258" containerName="init" Nov 22 08:15:58 crc kubenswrapper[4789]: I1122 08:15:58.111567 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5b56df9788-wsmhq" Nov 22 08:15:58 crc kubenswrapper[4789]: I1122 08:15:58.115057 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Nov 22 08:15:58 crc kubenswrapper[4789]: I1122 08:15:58.157274 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5b56df9788-wsmhq"] Nov 22 08:15:58 crc kubenswrapper[4789]: I1122 08:15:58.223357 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-6d4f97b8d9-fqgm7"] Nov 22 08:15:58 crc kubenswrapper[4789]: I1122 08:15:58.225111 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v9rvp\" (UniqueName: \"kubernetes.io/projected/56fd3074-a15c-4024-baf9-f784a50e14a5-kube-api-access-v9rvp\") pod \"horizon-5b56df9788-wsmhq\" (UID: \"56fd3074-a15c-4024-baf9-f784a50e14a5\") " pod="openstack/horizon-5b56df9788-wsmhq" Nov 22 08:15:58 crc kubenswrapper[4789]: I1122 08:15:58.225157 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/56fd3074-a15c-4024-baf9-f784a50e14a5-config-data\") pod \"horizon-5b56df9788-wsmhq\" (UID: \"56fd3074-a15c-4024-baf9-f784a50e14a5\") " pod="openstack/horizon-5b56df9788-wsmhq" Nov 22 08:15:58 crc kubenswrapper[4789]: I1122 08:15:58.225205 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/56fd3074-a15c-4024-baf9-f784a50e14a5-scripts\") pod \"horizon-5b56df9788-wsmhq\" (UID: \"56fd3074-a15c-4024-baf9-f784a50e14a5\") " pod="openstack/horizon-5b56df9788-wsmhq" Nov 22 08:15:58 crc kubenswrapper[4789]: I1122 08:15:58.225220 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/56fd3074-a15c-4024-baf9-f784a50e14a5-horizon-tls-certs\") pod \"horizon-5b56df9788-wsmhq\" (UID: \"56fd3074-a15c-4024-baf9-f784a50e14a5\") " pod="openstack/horizon-5b56df9788-wsmhq" Nov 22 08:15:58 crc kubenswrapper[4789]: I1122 08:15:58.225272 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/56fd3074-a15c-4024-baf9-f784a50e14a5-logs\") pod \"horizon-5b56df9788-wsmhq\" (UID: \"56fd3074-a15c-4024-baf9-f784a50e14a5\") " pod="openstack/horizon-5b56df9788-wsmhq" Nov 22 08:15:58 crc kubenswrapper[4789]: I1122 08:15:58.225309 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/56fd3074-a15c-4024-baf9-f784a50e14a5-horizon-secret-key\") pod \"horizon-5b56df9788-wsmhq\" (UID: \"56fd3074-a15c-4024-baf9-f784a50e14a5\") " pod="openstack/horizon-5b56df9788-wsmhq" Nov 22 08:15:58 crc kubenswrapper[4789]: I1122 08:15:58.225334 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56fd3074-a15c-4024-baf9-f784a50e14a5-combined-ca-bundle\") pod \"horizon-5b56df9788-wsmhq\" (UID: \"56fd3074-a15c-4024-baf9-f784a50e14a5\") " pod="openstack/horizon-5b56df9788-wsmhq" Nov 22 08:15:58 crc kubenswrapper[4789]: I1122 08:15:58.258132 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7b778f8c8-6rl4f"] Nov 22 08:15:58 crc kubenswrapper[4789]: I1122 08:15:58.260709 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7b778f8c8-6rl4f" Nov 22 08:15:58 crc kubenswrapper[4789]: I1122 08:15:58.272347 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7b778f8c8-6rl4f"] Nov 22 08:15:58 crc kubenswrapper[4789]: I1122 08:15:58.327596 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/56fd3074-a15c-4024-baf9-f784a50e14a5-logs\") pod \"horizon-5b56df9788-wsmhq\" (UID: \"56fd3074-a15c-4024-baf9-f784a50e14a5\") " pod="openstack/horizon-5b56df9788-wsmhq" Nov 22 08:15:58 crc kubenswrapper[4789]: I1122 08:15:58.327667 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/56fd3074-a15c-4024-baf9-f784a50e14a5-horizon-secret-key\") pod \"horizon-5b56df9788-wsmhq\" (UID: \"56fd3074-a15c-4024-baf9-f784a50e14a5\") " pod="openstack/horizon-5b56df9788-wsmhq" Nov 22 08:15:58 crc kubenswrapper[4789]: I1122 08:15:58.327700 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56fd3074-a15c-4024-baf9-f784a50e14a5-combined-ca-bundle\") pod \"horizon-5b56df9788-wsmhq\" (UID: \"56fd3074-a15c-4024-baf9-f784a50e14a5\") " pod="openstack/horizon-5b56df9788-wsmhq" Nov 22 08:15:58 crc kubenswrapper[4789]: I1122 08:15:58.327735 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v9rvp\" (UniqueName: \"kubernetes.io/projected/56fd3074-a15c-4024-baf9-f784a50e14a5-kube-api-access-v9rvp\") pod \"horizon-5b56df9788-wsmhq\" (UID: \"56fd3074-a15c-4024-baf9-f784a50e14a5\") " pod="openstack/horizon-5b56df9788-wsmhq" Nov 22 08:15:58 crc kubenswrapper[4789]: I1122 08:15:58.327778 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/56fd3074-a15c-4024-baf9-f784a50e14a5-config-data\") pod \"horizon-5b56df9788-wsmhq\" (UID: \"56fd3074-a15c-4024-baf9-f784a50e14a5\") " pod="openstack/horizon-5b56df9788-wsmhq" Nov 22 08:15:58 crc kubenswrapper[4789]: I1122 08:15:58.327822 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/56fd3074-a15c-4024-baf9-f784a50e14a5-scripts\") pod \"horizon-5b56df9788-wsmhq\" (UID: \"56fd3074-a15c-4024-baf9-f784a50e14a5\") " pod="openstack/horizon-5b56df9788-wsmhq" Nov 22 08:15:58 crc kubenswrapper[4789]: I1122 08:15:58.327837 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/56fd3074-a15c-4024-baf9-f784a50e14a5-horizon-tls-certs\") pod \"horizon-5b56df9788-wsmhq\" (UID: \"56fd3074-a15c-4024-baf9-f784a50e14a5\") " pod="openstack/horizon-5b56df9788-wsmhq" Nov 22 08:15:58 crc kubenswrapper[4789]: I1122 08:15:58.329302 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/56fd3074-a15c-4024-baf9-f784a50e14a5-logs\") pod \"horizon-5b56df9788-wsmhq\" (UID: \"56fd3074-a15c-4024-baf9-f784a50e14a5\") " pod="openstack/horizon-5b56df9788-wsmhq" Nov 22 08:15:58 crc kubenswrapper[4789]: I1122 08:15:58.329879 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/56fd3074-a15c-4024-baf9-f784a50e14a5-config-data\") pod \"horizon-5b56df9788-wsmhq\" (UID: \"56fd3074-a15c-4024-baf9-f784a50e14a5\") " pod="openstack/horizon-5b56df9788-wsmhq" Nov 22 08:15:58 crc kubenswrapper[4789]: I1122 08:15:58.329887 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/56fd3074-a15c-4024-baf9-f784a50e14a5-scripts\") pod \"horizon-5b56df9788-wsmhq\" (UID: \"56fd3074-a15c-4024-baf9-f784a50e14a5\") " pod="openstack/horizon-5b56df9788-wsmhq" Nov 22 08:15:58 crc kubenswrapper[4789]: I1122 08:15:58.336081 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/56fd3074-a15c-4024-baf9-f784a50e14a5-horizon-secret-key\") pod \"horizon-5b56df9788-wsmhq\" (UID: \"56fd3074-a15c-4024-baf9-f784a50e14a5\") " pod="openstack/horizon-5b56df9788-wsmhq" Nov 22 08:15:58 crc kubenswrapper[4789]: I1122 08:15:58.336088 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/56fd3074-a15c-4024-baf9-f784a50e14a5-horizon-tls-certs\") pod \"horizon-5b56df9788-wsmhq\" (UID: \"56fd3074-a15c-4024-baf9-f784a50e14a5\") " pod="openstack/horizon-5b56df9788-wsmhq" Nov 22 08:15:58 crc kubenswrapper[4789]: I1122 08:15:58.336685 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56fd3074-a15c-4024-baf9-f784a50e14a5-combined-ca-bundle\") pod \"horizon-5b56df9788-wsmhq\" (UID: \"56fd3074-a15c-4024-baf9-f784a50e14a5\") " pod="openstack/horizon-5b56df9788-wsmhq" Nov 22 08:15:58 crc kubenswrapper[4789]: I1122 08:15:58.349503 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v9rvp\" (UniqueName: \"kubernetes.io/projected/56fd3074-a15c-4024-baf9-f784a50e14a5-kube-api-access-v9rvp\") pod \"horizon-5b56df9788-wsmhq\" (UID: \"56fd3074-a15c-4024-baf9-f784a50e14a5\") " pod="openstack/horizon-5b56df9788-wsmhq" Nov 22 08:15:58 crc kubenswrapper[4789]: I1122 08:15:58.429521 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d9fb66f7-ee5b-453a-a69b-56fb9337d898-scripts\") pod \"horizon-7b778f8c8-6rl4f\" (UID: \"d9fb66f7-ee5b-453a-a69b-56fb9337d898\") " pod="openstack/horizon-7b778f8c8-6rl4f" Nov 22 08:15:58 crc kubenswrapper[4789]: I1122 08:15:58.429575 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d9fb66f7-ee5b-453a-a69b-56fb9337d898-logs\") pod \"horizon-7b778f8c8-6rl4f\" (UID: \"d9fb66f7-ee5b-453a-a69b-56fb9337d898\") " pod="openstack/horizon-7b778f8c8-6rl4f" Nov 22 08:15:58 crc kubenswrapper[4789]: I1122 08:15:58.429642 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d9fb66f7-ee5b-453a-a69b-56fb9337d898-config-data\") pod \"horizon-7b778f8c8-6rl4f\" (UID: \"d9fb66f7-ee5b-453a-a69b-56fb9337d898\") " pod="openstack/horizon-7b778f8c8-6rl4f" Nov 22 08:15:58 crc kubenswrapper[4789]: I1122 08:15:58.429679 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9fb66f7-ee5b-453a-a69b-56fb9337d898-combined-ca-bundle\") pod \"horizon-7b778f8c8-6rl4f\" (UID: \"d9fb66f7-ee5b-453a-a69b-56fb9337d898\") " pod="openstack/horizon-7b778f8c8-6rl4f" Nov 22 08:15:58 crc kubenswrapper[4789]: I1122 08:15:58.429697 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d9fb66f7-ee5b-453a-a69b-56fb9337d898-horizon-secret-key\") pod \"horizon-7b778f8c8-6rl4f\" (UID: \"d9fb66f7-ee5b-453a-a69b-56fb9337d898\") " pod="openstack/horizon-7b778f8c8-6rl4f" Nov 22 08:15:58 crc kubenswrapper[4789]: I1122 08:15:58.429859 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9fb66f7-ee5b-453a-a69b-56fb9337d898-horizon-tls-certs\") pod \"horizon-7b778f8c8-6rl4f\" (UID: \"d9fb66f7-ee5b-453a-a69b-56fb9337d898\") " pod="openstack/horizon-7b778f8c8-6rl4f" Nov 22 08:15:58 crc kubenswrapper[4789]: I1122 08:15:58.429947 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sqvpt\" (UniqueName: \"kubernetes.io/projected/d9fb66f7-ee5b-453a-a69b-56fb9337d898-kube-api-access-sqvpt\") pod \"horizon-7b778f8c8-6rl4f\" (UID: \"d9fb66f7-ee5b-453a-a69b-56fb9337d898\") " pod="openstack/horizon-7b778f8c8-6rl4f" Nov 22 08:15:58 crc kubenswrapper[4789]: I1122 08:15:58.451407 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5b56df9788-wsmhq" Nov 22 08:15:58 crc kubenswrapper[4789]: I1122 08:15:58.531186 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9fb66f7-ee5b-453a-a69b-56fb9337d898-horizon-tls-certs\") pod \"horizon-7b778f8c8-6rl4f\" (UID: \"d9fb66f7-ee5b-453a-a69b-56fb9337d898\") " pod="openstack/horizon-7b778f8c8-6rl4f" Nov 22 08:15:58 crc kubenswrapper[4789]: I1122 08:15:58.531261 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sqvpt\" (UniqueName: \"kubernetes.io/projected/d9fb66f7-ee5b-453a-a69b-56fb9337d898-kube-api-access-sqvpt\") pod \"horizon-7b778f8c8-6rl4f\" (UID: \"d9fb66f7-ee5b-453a-a69b-56fb9337d898\") " pod="openstack/horizon-7b778f8c8-6rl4f" Nov 22 08:15:58 crc kubenswrapper[4789]: I1122 08:15:58.531307 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d9fb66f7-ee5b-453a-a69b-56fb9337d898-scripts\") pod \"horizon-7b778f8c8-6rl4f\" (UID: \"d9fb66f7-ee5b-453a-a69b-56fb9337d898\") " pod="openstack/horizon-7b778f8c8-6rl4f" Nov 22 08:15:58 crc kubenswrapper[4789]: I1122 08:15:58.531338 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d9fb66f7-ee5b-453a-a69b-56fb9337d898-logs\") pod \"horizon-7b778f8c8-6rl4f\" (UID: \"d9fb66f7-ee5b-453a-a69b-56fb9337d898\") " pod="openstack/horizon-7b778f8c8-6rl4f" Nov 22 08:15:58 crc kubenswrapper[4789]: I1122 08:15:58.531385 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d9fb66f7-ee5b-453a-a69b-56fb9337d898-config-data\") pod \"horizon-7b778f8c8-6rl4f\" (UID: \"d9fb66f7-ee5b-453a-a69b-56fb9337d898\") " pod="openstack/horizon-7b778f8c8-6rl4f" Nov 22 08:15:58 crc kubenswrapper[4789]: I1122 08:15:58.531425 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9fb66f7-ee5b-453a-a69b-56fb9337d898-combined-ca-bundle\") pod \"horizon-7b778f8c8-6rl4f\" (UID: \"d9fb66f7-ee5b-453a-a69b-56fb9337d898\") " pod="openstack/horizon-7b778f8c8-6rl4f" Nov 22 08:15:58 crc kubenswrapper[4789]: I1122 08:15:58.531446 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d9fb66f7-ee5b-453a-a69b-56fb9337d898-horizon-secret-key\") pod \"horizon-7b778f8c8-6rl4f\" (UID: \"d9fb66f7-ee5b-453a-a69b-56fb9337d898\") " pod="openstack/horizon-7b778f8c8-6rl4f" Nov 22 08:15:58 crc kubenswrapper[4789]: I1122 08:15:58.531810 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d9fb66f7-ee5b-453a-a69b-56fb9337d898-logs\") pod \"horizon-7b778f8c8-6rl4f\" (UID: \"d9fb66f7-ee5b-453a-a69b-56fb9337d898\") " pod="openstack/horizon-7b778f8c8-6rl4f" Nov 22 08:15:58 crc kubenswrapper[4789]: I1122 08:15:58.532027 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d9fb66f7-ee5b-453a-a69b-56fb9337d898-scripts\") pod \"horizon-7b778f8c8-6rl4f\" (UID: \"d9fb66f7-ee5b-453a-a69b-56fb9337d898\") " pod="openstack/horizon-7b778f8c8-6rl4f" Nov 22 08:15:58 crc kubenswrapper[4789]: I1122 08:15:58.533053 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d9fb66f7-ee5b-453a-a69b-56fb9337d898-config-data\") pod \"horizon-7b778f8c8-6rl4f\" (UID: \"d9fb66f7-ee5b-453a-a69b-56fb9337d898\") " pod="openstack/horizon-7b778f8c8-6rl4f" Nov 22 08:15:58 crc kubenswrapper[4789]: I1122 08:15:58.535007 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9fb66f7-ee5b-453a-a69b-56fb9337d898-horizon-tls-certs\") pod \"horizon-7b778f8c8-6rl4f\" (UID: \"d9fb66f7-ee5b-453a-a69b-56fb9337d898\") " pod="openstack/horizon-7b778f8c8-6rl4f" Nov 22 08:15:58 crc kubenswrapper[4789]: I1122 08:15:58.540390 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d9fb66f7-ee5b-453a-a69b-56fb9337d898-horizon-secret-key\") pod \"horizon-7b778f8c8-6rl4f\" (UID: \"d9fb66f7-ee5b-453a-a69b-56fb9337d898\") " pod="openstack/horizon-7b778f8c8-6rl4f" Nov 22 08:15:58 crc kubenswrapper[4789]: I1122 08:15:58.543072 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9fb66f7-ee5b-453a-a69b-56fb9337d898-combined-ca-bundle\") pod \"horizon-7b778f8c8-6rl4f\" (UID: \"d9fb66f7-ee5b-453a-a69b-56fb9337d898\") " pod="openstack/horizon-7b778f8c8-6rl4f" Nov 22 08:15:58 crc kubenswrapper[4789]: I1122 08:15:58.558824 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sqvpt\" (UniqueName: \"kubernetes.io/projected/d9fb66f7-ee5b-453a-a69b-56fb9337d898-kube-api-access-sqvpt\") pod \"horizon-7b778f8c8-6rl4f\" (UID: \"d9fb66f7-ee5b-453a-a69b-56fb9337d898\") " pod="openstack/horizon-7b778f8c8-6rl4f" Nov 22 08:15:58 crc kubenswrapper[4789]: I1122 08:15:58.579357 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7b778f8c8-6rl4f" Nov 22 08:16:00 crc kubenswrapper[4789]: I1122 08:16:00.063182 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-66f4bdbdb7-tzwtz" Nov 22 08:16:00 crc kubenswrapper[4789]: I1122 08:16:00.114217 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c476d78c5-djb2g"] Nov 22 08:16:00 crc kubenswrapper[4789]: I1122 08:16:00.114447 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5c476d78c5-djb2g" podUID="0cf0fb02-e586-4daa-88c1-6556b376c02a" containerName="dnsmasq-dns" containerID="cri-o://3bba78e22b91a98d5a6fd8c4d50cedda778719b31f2900449fafb4e224b7a2bb" gracePeriod=10 Nov 22 08:16:01 crc kubenswrapper[4789]: I1122 08:16:01.914265 4789 generic.go:334] "Generic (PLEG): container finished" podID="0cf0fb02-e586-4daa-88c1-6556b376c02a" containerID="3bba78e22b91a98d5a6fd8c4d50cedda778719b31f2900449fafb4e224b7a2bb" exitCode=0 Nov 22 08:16:01 crc kubenswrapper[4789]: I1122 08:16:01.914434 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c476d78c5-djb2g" event={"ID":"0cf0fb02-e586-4daa-88c1-6556b376c02a","Type":"ContainerDied","Data":"3bba78e22b91a98d5a6fd8c4d50cedda778719b31f2900449fafb4e224b7a2bb"} Nov 22 08:16:03 crc kubenswrapper[4789]: I1122 08:16:03.063210 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5c476d78c5-djb2g" podUID="0cf0fb02-e586-4daa-88c1-6556b376c02a" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.110:5353: connect: connection refused" Nov 22 08:16:06 crc kubenswrapper[4789]: I1122 08:16:05.372445 4789 patch_prober.go:28] interesting pod/machine-config-daemon-pxvrm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 08:16:06 crc kubenswrapper[4789]: I1122 08:16:05.372779 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 08:16:06 crc kubenswrapper[4789]: I1122 08:16:05.372823 4789 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" Nov 22 08:16:06 crc kubenswrapper[4789]: I1122 08:16:05.373652 4789 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"fe65d4df369fc1d7c708e571722629956fbd341c1d6b8fc8c7f64c8dfbfa797f"} pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 22 08:16:06 crc kubenswrapper[4789]: I1122 08:16:05.373739 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" containerID="cri-o://fe65d4df369fc1d7c708e571722629956fbd341c1d6b8fc8c7f64c8dfbfa797f" gracePeriod=600 Nov 22 08:16:06 crc kubenswrapper[4789]: I1122 08:16:06.962979 4789 generic.go:334] "Generic (PLEG): container finished" podID="58aa8071-7ecc-4692-8789-2db04ec70647" containerID="fe65d4df369fc1d7c708e571722629956fbd341c1d6b8fc8c7f64c8dfbfa797f" exitCode=0 Nov 22 08:16:06 crc kubenswrapper[4789]: I1122 08:16:06.963072 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" event={"ID":"58aa8071-7ecc-4692-8789-2db04ec70647","Type":"ContainerDied","Data":"fe65d4df369fc1d7c708e571722629956fbd341c1d6b8fc8c7f64c8dfbfa797f"} Nov 22 08:16:06 crc kubenswrapper[4789]: I1122 08:16:06.963366 4789 scope.go:117] "RemoveContainer" containerID="5067bbca7c5576cc068f8534a87953cf9b5fc5cb8f8cfddaaed99f53192975c1" Nov 22 08:16:08 crc kubenswrapper[4789]: I1122 08:16:08.062491 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5c476d78c5-djb2g" podUID="0cf0fb02-e586-4daa-88c1-6556b376c02a" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.110:5353: connect: connection refused" Nov 22 08:16:09 crc kubenswrapper[4789]: E1122 08:16:09.070043 4789 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon@sha256:65b94ff9fcd486845fb0544583bf2a973246a61a0ad32340fb92d632285f1057" Nov 22 08:16:09 crc kubenswrapper[4789]: E1122 08:16:09.070240 4789 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon@sha256:65b94ff9fcd486845fb0544583bf2a973246a61a0ad32340fb92d632285f1057,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nddh578h64dhb9h555h599h65fh5fbh694hbfh65dh5f5h5c8h64dh567hc8h64fh58dh698h54dh5fbh5bh66fh6fhbdh56ch7h655h55fhfh8dhccq,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-grm5s,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-6d4f97b8d9-fqgm7_openstack(61bf72d1-ea9f-4ad5-9dd1-7fdb4a65e8e4): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 22 08:16:09 crc kubenswrapper[4789]: E1122 08:16:09.274023 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon@sha256:65b94ff9fcd486845fb0544583bf2a973246a61a0ad32340fb92d632285f1057\\\"\"]" pod="openstack/horizon-6d4f97b8d9-fqgm7" podUID="61bf72d1-ea9f-4ad5-9dd1-7fdb4a65e8e4" Nov 22 08:16:11 crc kubenswrapper[4789]: E1122 08:16:11.822685 4789 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-placement-api@sha256:7dd2e0dbb6bb5a6cecd1763e43479ca8cb6a0c502534e83c8795c0da2b50e099" Nov 22 08:16:11 crc kubenswrapper[4789]: E1122 08:16:11.823418 4789 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:placement-db-sync,Image:quay.io/podified-antelope-centos9/openstack-placement-api@sha256:7dd2e0dbb6bb5a6cecd1763e43479ca8cb6a0c502534e83c8795c0da2b50e099,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/placement,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:placement-dbsync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-h867w,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42482,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-db-sync-2ss46_openstack(2eb187f8-1892-4d72-b442-43592e75d489): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 22 08:16:11 crc kubenswrapper[4789]: E1122 08:16:11.824608 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/placement-db-sync-2ss46" podUID="2eb187f8-1892-4d72-b442-43592e75d489" Nov 22 08:16:12 crc kubenswrapper[4789]: E1122 08:16:12.011195 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-placement-api@sha256:7dd2e0dbb6bb5a6cecd1763e43479ca8cb6a0c502534e83c8795c0da2b50e099\\\"\"" pod="openstack/placement-db-sync-2ss46" podUID="2eb187f8-1892-4d72-b442-43592e75d489" Nov 22 08:16:13 crc kubenswrapper[4789]: I1122 08:16:13.063230 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5c476d78c5-djb2g" podUID="0cf0fb02-e586-4daa-88c1-6556b376c02a" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.110:5353: connect: connection refused" Nov 22 08:16:13 crc kubenswrapper[4789]: I1122 08:16:13.063669 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c476d78c5-djb2g" Nov 22 08:16:13 crc kubenswrapper[4789]: E1122 08:16:13.206416 4789 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon@sha256:65b94ff9fcd486845fb0544583bf2a973246a61a0ad32340fb92d632285f1057" Nov 22 08:16:13 crc kubenswrapper[4789]: E1122 08:16:13.206568 4789 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon@sha256:65b94ff9fcd486845fb0544583bf2a973246a61a0ad32340fb92d632285f1057,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n7chdh55hc5h649h74h59dh99h588h66dh66dh68h67dhc5h58fhcfh546h584h578h66dh5ch9bh54fh5ffh596h85hf5h6bhbbh55dhbdh64q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rbbt8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-6ddbf95fdc-64jxv_openstack(eea1fd9b-942e-4d60-80ae-1fca761f5a32): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 22 08:16:13 crc kubenswrapper[4789]: E1122 08:16:13.208448 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon@sha256:65b94ff9fcd486845fb0544583bf2a973246a61a0ad32340fb92d632285f1057\\\"\"]" pod="openstack/horizon-6ddbf95fdc-64jxv" podUID="eea1fd9b-942e-4d60-80ae-1fca761f5a32" Nov 22 08:16:13 crc kubenswrapper[4789]: E1122 08:16:13.218874 4789 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon@sha256:65b94ff9fcd486845fb0544583bf2a973246a61a0ad32340fb92d632285f1057" Nov 22 08:16:13 crc kubenswrapper[4789]: E1122 08:16:13.219077 4789 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon@sha256:65b94ff9fcd486845fb0544583bf2a973246a61a0ad32340fb92d632285f1057,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n54bh677h5h5d6h64dh66dh5cfhb6h58dhd6hf4h94h8fh64fh548h566h5f8h64bh96h649h5c9h87hb4h58fh549hdh657h5b7h56ch649h588h668q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-cswpl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-64b577d9f-n8w7t_openstack(e3b7f50f-961d-472f-9423-18ab69fcdaef): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 22 08:16:13 crc kubenswrapper[4789]: E1122 08:16:13.221910 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon@sha256:65b94ff9fcd486845fb0544583bf2a973246a61a0ad32340fb92d632285f1057\\\"\"]" pod="openstack/horizon-64b577d9f-n8w7t" podUID="e3b7f50f-961d-472f-9423-18ab69fcdaef" Nov 22 08:16:13 crc kubenswrapper[4789]: I1122 08:16:13.280539 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-9ts56" Nov 22 08:16:13 crc kubenswrapper[4789]: I1122 08:16:13.390600 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09e8392f-f97d-4922-be5a-fb6229549344-combined-ca-bundle\") pod \"09e8392f-f97d-4922-be5a-fb6229549344\" (UID: \"09e8392f-f97d-4922-be5a-fb6229549344\") " Nov 22 08:16:13 crc kubenswrapper[4789]: I1122 08:16:13.390985 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/09e8392f-f97d-4922-be5a-fb6229549344-fernet-keys\") pod \"09e8392f-f97d-4922-be5a-fb6229549344\" (UID: \"09e8392f-f97d-4922-be5a-fb6229549344\") " Nov 22 08:16:13 crc kubenswrapper[4789]: I1122 08:16:13.391026 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/09e8392f-f97d-4922-be5a-fb6229549344-scripts\") pod \"09e8392f-f97d-4922-be5a-fb6229549344\" (UID: \"09e8392f-f97d-4922-be5a-fb6229549344\") " Nov 22 08:16:13 crc kubenswrapper[4789]: I1122 08:16:13.391057 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/09e8392f-f97d-4922-be5a-fb6229549344-credential-keys\") pod \"09e8392f-f97d-4922-be5a-fb6229549344\" (UID: \"09e8392f-f97d-4922-be5a-fb6229549344\") " Nov 22 08:16:13 crc kubenswrapper[4789]: I1122 08:16:13.391110 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/09e8392f-f97d-4922-be5a-fb6229549344-config-data\") pod \"09e8392f-f97d-4922-be5a-fb6229549344\" (UID: \"09e8392f-f97d-4922-be5a-fb6229549344\") " Nov 22 08:16:13 crc kubenswrapper[4789]: I1122 08:16:13.391223 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gs4k9\" (UniqueName: \"kubernetes.io/projected/09e8392f-f97d-4922-be5a-fb6229549344-kube-api-access-gs4k9\") pod \"09e8392f-f97d-4922-be5a-fb6229549344\" (UID: \"09e8392f-f97d-4922-be5a-fb6229549344\") " Nov 22 08:16:13 crc kubenswrapper[4789]: I1122 08:16:13.395668 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09e8392f-f97d-4922-be5a-fb6229549344-kube-api-access-gs4k9" (OuterVolumeSpecName: "kube-api-access-gs4k9") pod "09e8392f-f97d-4922-be5a-fb6229549344" (UID: "09e8392f-f97d-4922-be5a-fb6229549344"). InnerVolumeSpecName "kube-api-access-gs4k9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:16:13 crc kubenswrapper[4789]: I1122 08:16:13.397291 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09e8392f-f97d-4922-be5a-fb6229549344-scripts" (OuterVolumeSpecName: "scripts") pod "09e8392f-f97d-4922-be5a-fb6229549344" (UID: "09e8392f-f97d-4922-be5a-fb6229549344"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:16:13 crc kubenswrapper[4789]: I1122 08:16:13.398701 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09e8392f-f97d-4922-be5a-fb6229549344-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "09e8392f-f97d-4922-be5a-fb6229549344" (UID: "09e8392f-f97d-4922-be5a-fb6229549344"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:16:13 crc kubenswrapper[4789]: I1122 08:16:13.402858 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09e8392f-f97d-4922-be5a-fb6229549344-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "09e8392f-f97d-4922-be5a-fb6229549344" (UID: "09e8392f-f97d-4922-be5a-fb6229549344"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:16:13 crc kubenswrapper[4789]: I1122 08:16:13.414784 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09e8392f-f97d-4922-be5a-fb6229549344-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "09e8392f-f97d-4922-be5a-fb6229549344" (UID: "09e8392f-f97d-4922-be5a-fb6229549344"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:16:13 crc kubenswrapper[4789]: I1122 08:16:13.418965 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09e8392f-f97d-4922-be5a-fb6229549344-config-data" (OuterVolumeSpecName: "config-data") pod "09e8392f-f97d-4922-be5a-fb6229549344" (UID: "09e8392f-f97d-4922-be5a-fb6229549344"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:16:13 crc kubenswrapper[4789]: I1122 08:16:13.496163 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gs4k9\" (UniqueName: \"kubernetes.io/projected/09e8392f-f97d-4922-be5a-fb6229549344-kube-api-access-gs4k9\") on node \"crc\" DevicePath \"\"" Nov 22 08:16:13 crc kubenswrapper[4789]: I1122 08:16:13.496192 4789 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09e8392f-f97d-4922-be5a-fb6229549344-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:16:13 crc kubenswrapper[4789]: I1122 08:16:13.496201 4789 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/09e8392f-f97d-4922-be5a-fb6229549344-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 22 08:16:13 crc kubenswrapper[4789]: I1122 08:16:13.496211 4789 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/09e8392f-f97d-4922-be5a-fb6229549344-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:16:13 crc kubenswrapper[4789]: I1122 08:16:13.496220 4789 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/09e8392f-f97d-4922-be5a-fb6229549344-credential-keys\") on node \"crc\" DevicePath \"\"" Nov 22 08:16:13 crc kubenswrapper[4789]: I1122 08:16:13.496227 4789 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/09e8392f-f97d-4922-be5a-fb6229549344-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:16:14 crc kubenswrapper[4789]: I1122 08:16:14.027425 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-9ts56" event={"ID":"09e8392f-f97d-4922-be5a-fb6229549344","Type":"ContainerDied","Data":"7ebc87cea54f0af771db08c43b687c409d8174a725e89ad7b8e62fffbae9a118"} Nov 22 08:16:14 crc kubenswrapper[4789]: I1122 08:16:14.027489 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7ebc87cea54f0af771db08c43b687c409d8174a725e89ad7b8e62fffbae9a118" Nov 22 08:16:14 crc kubenswrapper[4789]: I1122 08:16:14.027460 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-9ts56" Nov 22 08:16:14 crc kubenswrapper[4789]: I1122 08:16:14.357365 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-9ts56"] Nov 22 08:16:14 crc kubenswrapper[4789]: I1122 08:16:14.363720 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-9ts56"] Nov 22 08:16:14 crc kubenswrapper[4789]: I1122 08:16:14.467393 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-rnzhp"] Nov 22 08:16:14 crc kubenswrapper[4789]: E1122 08:16:14.467838 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09e8392f-f97d-4922-be5a-fb6229549344" containerName="keystone-bootstrap" Nov 22 08:16:14 crc kubenswrapper[4789]: I1122 08:16:14.467857 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="09e8392f-f97d-4922-be5a-fb6229549344" containerName="keystone-bootstrap" Nov 22 08:16:14 crc kubenswrapper[4789]: I1122 08:16:14.468028 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="09e8392f-f97d-4922-be5a-fb6229549344" containerName="keystone-bootstrap" Nov 22 08:16:14 crc kubenswrapper[4789]: I1122 08:16:14.468640 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-rnzhp" Nov 22 08:16:14 crc kubenswrapper[4789]: I1122 08:16:14.470854 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 22 08:16:14 crc kubenswrapper[4789]: I1122 08:16:14.472638 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 22 08:16:14 crc kubenswrapper[4789]: I1122 08:16:14.472732 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 22 08:16:14 crc kubenswrapper[4789]: I1122 08:16:14.472899 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-wnnts" Nov 22 08:16:14 crc kubenswrapper[4789]: I1122 08:16:14.473219 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 22 08:16:14 crc kubenswrapper[4789]: I1122 08:16:14.479287 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-rnzhp"] Nov 22 08:16:14 crc kubenswrapper[4789]: I1122 08:16:14.623285 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1f1fc498-45f3-4bec-a4a5-b9722d2445ea-fernet-keys\") pod \"keystone-bootstrap-rnzhp\" (UID: \"1f1fc498-45f3-4bec-a4a5-b9722d2445ea\") " pod="openstack/keystone-bootstrap-rnzhp" Nov 22 08:16:14 crc kubenswrapper[4789]: I1122 08:16:14.623326 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f1fc498-45f3-4bec-a4a5-b9722d2445ea-combined-ca-bundle\") pod \"keystone-bootstrap-rnzhp\" (UID: \"1f1fc498-45f3-4bec-a4a5-b9722d2445ea\") " pod="openstack/keystone-bootstrap-rnzhp" Nov 22 08:16:14 crc kubenswrapper[4789]: I1122 08:16:14.623345 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1f1fc498-45f3-4bec-a4a5-b9722d2445ea-credential-keys\") pod \"keystone-bootstrap-rnzhp\" (UID: \"1f1fc498-45f3-4bec-a4a5-b9722d2445ea\") " pod="openstack/keystone-bootstrap-rnzhp" Nov 22 08:16:14 crc kubenswrapper[4789]: I1122 08:16:14.623373 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1f1fc498-45f3-4bec-a4a5-b9722d2445ea-scripts\") pod \"keystone-bootstrap-rnzhp\" (UID: \"1f1fc498-45f3-4bec-a4a5-b9722d2445ea\") " pod="openstack/keystone-bootstrap-rnzhp" Nov 22 08:16:14 crc kubenswrapper[4789]: I1122 08:16:14.623475 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f1fc498-45f3-4bec-a4a5-b9722d2445ea-config-data\") pod \"keystone-bootstrap-rnzhp\" (UID: \"1f1fc498-45f3-4bec-a4a5-b9722d2445ea\") " pod="openstack/keystone-bootstrap-rnzhp" Nov 22 08:16:14 crc kubenswrapper[4789]: I1122 08:16:14.623542 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7m7kw\" (UniqueName: \"kubernetes.io/projected/1f1fc498-45f3-4bec-a4a5-b9722d2445ea-kube-api-access-7m7kw\") pod \"keystone-bootstrap-rnzhp\" (UID: \"1f1fc498-45f3-4bec-a4a5-b9722d2445ea\") " pod="openstack/keystone-bootstrap-rnzhp" Nov 22 08:16:14 crc kubenswrapper[4789]: I1122 08:16:14.725799 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1f1fc498-45f3-4bec-a4a5-b9722d2445ea-fernet-keys\") pod \"keystone-bootstrap-rnzhp\" (UID: \"1f1fc498-45f3-4bec-a4a5-b9722d2445ea\") " pod="openstack/keystone-bootstrap-rnzhp" Nov 22 08:16:14 crc kubenswrapper[4789]: I1122 08:16:14.725871 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f1fc498-45f3-4bec-a4a5-b9722d2445ea-combined-ca-bundle\") pod \"keystone-bootstrap-rnzhp\" (UID: \"1f1fc498-45f3-4bec-a4a5-b9722d2445ea\") " pod="openstack/keystone-bootstrap-rnzhp" Nov 22 08:16:14 crc kubenswrapper[4789]: I1122 08:16:14.725905 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1f1fc498-45f3-4bec-a4a5-b9722d2445ea-credential-keys\") pod \"keystone-bootstrap-rnzhp\" (UID: \"1f1fc498-45f3-4bec-a4a5-b9722d2445ea\") " pod="openstack/keystone-bootstrap-rnzhp" Nov 22 08:16:14 crc kubenswrapper[4789]: I1122 08:16:14.725964 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1f1fc498-45f3-4bec-a4a5-b9722d2445ea-scripts\") pod \"keystone-bootstrap-rnzhp\" (UID: \"1f1fc498-45f3-4bec-a4a5-b9722d2445ea\") " pod="openstack/keystone-bootstrap-rnzhp" Nov 22 08:16:14 crc kubenswrapper[4789]: I1122 08:16:14.726122 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f1fc498-45f3-4bec-a4a5-b9722d2445ea-config-data\") pod \"keystone-bootstrap-rnzhp\" (UID: \"1f1fc498-45f3-4bec-a4a5-b9722d2445ea\") " pod="openstack/keystone-bootstrap-rnzhp" Nov 22 08:16:14 crc kubenswrapper[4789]: I1122 08:16:14.726347 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7m7kw\" (UniqueName: \"kubernetes.io/projected/1f1fc498-45f3-4bec-a4a5-b9722d2445ea-kube-api-access-7m7kw\") pod \"keystone-bootstrap-rnzhp\" (UID: \"1f1fc498-45f3-4bec-a4a5-b9722d2445ea\") " pod="openstack/keystone-bootstrap-rnzhp" Nov 22 08:16:14 crc kubenswrapper[4789]: I1122 08:16:14.732230 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1f1fc498-45f3-4bec-a4a5-b9722d2445ea-scripts\") pod \"keystone-bootstrap-rnzhp\" (UID: \"1f1fc498-45f3-4bec-a4a5-b9722d2445ea\") " pod="openstack/keystone-bootstrap-rnzhp" Nov 22 08:16:14 crc kubenswrapper[4789]: I1122 08:16:14.732408 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f1fc498-45f3-4bec-a4a5-b9722d2445ea-combined-ca-bundle\") pod \"keystone-bootstrap-rnzhp\" (UID: \"1f1fc498-45f3-4bec-a4a5-b9722d2445ea\") " pod="openstack/keystone-bootstrap-rnzhp" Nov 22 08:16:14 crc kubenswrapper[4789]: I1122 08:16:14.732942 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1f1fc498-45f3-4bec-a4a5-b9722d2445ea-credential-keys\") pod \"keystone-bootstrap-rnzhp\" (UID: \"1f1fc498-45f3-4bec-a4a5-b9722d2445ea\") " pod="openstack/keystone-bootstrap-rnzhp" Nov 22 08:16:14 crc kubenswrapper[4789]: I1122 08:16:14.733526 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1f1fc498-45f3-4bec-a4a5-b9722d2445ea-fernet-keys\") pod \"keystone-bootstrap-rnzhp\" (UID: \"1f1fc498-45f3-4bec-a4a5-b9722d2445ea\") " pod="openstack/keystone-bootstrap-rnzhp" Nov 22 08:16:14 crc kubenswrapper[4789]: I1122 08:16:14.733592 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f1fc498-45f3-4bec-a4a5-b9722d2445ea-config-data\") pod \"keystone-bootstrap-rnzhp\" (UID: \"1f1fc498-45f3-4bec-a4a5-b9722d2445ea\") " pod="openstack/keystone-bootstrap-rnzhp" Nov 22 08:16:14 crc kubenswrapper[4789]: I1122 08:16:14.745231 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7m7kw\" (UniqueName: \"kubernetes.io/projected/1f1fc498-45f3-4bec-a4a5-b9722d2445ea-kube-api-access-7m7kw\") pod \"keystone-bootstrap-rnzhp\" (UID: \"1f1fc498-45f3-4bec-a4a5-b9722d2445ea\") " pod="openstack/keystone-bootstrap-rnzhp" Nov 22 08:16:14 crc kubenswrapper[4789]: I1122 08:16:14.790949 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-rnzhp" Nov 22 08:16:15 crc kubenswrapper[4789]: I1122 08:16:15.976256 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09e8392f-f97d-4922-be5a-fb6229549344" path="/var/lib/kubelet/pods/09e8392f-f97d-4922-be5a-fb6229549344/volumes" Nov 22 08:16:19 crc kubenswrapper[4789]: I1122 08:16:19.468766 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6d4f97b8d9-fqgm7" Nov 22 08:16:19 crc kubenswrapper[4789]: I1122 08:16:19.607360 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/61bf72d1-ea9f-4ad5-9dd1-7fdb4a65e8e4-config-data\") pod \"61bf72d1-ea9f-4ad5-9dd1-7fdb4a65e8e4\" (UID: \"61bf72d1-ea9f-4ad5-9dd1-7fdb4a65e8e4\") " Nov 22 08:16:19 crc kubenswrapper[4789]: I1122 08:16:19.607699 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/61bf72d1-ea9f-4ad5-9dd1-7fdb4a65e8e4-scripts\") pod \"61bf72d1-ea9f-4ad5-9dd1-7fdb4a65e8e4\" (UID: \"61bf72d1-ea9f-4ad5-9dd1-7fdb4a65e8e4\") " Nov 22 08:16:19 crc kubenswrapper[4789]: I1122 08:16:19.607799 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/61bf72d1-ea9f-4ad5-9dd1-7fdb4a65e8e4-logs\") pod \"61bf72d1-ea9f-4ad5-9dd1-7fdb4a65e8e4\" (UID: \"61bf72d1-ea9f-4ad5-9dd1-7fdb4a65e8e4\") " Nov 22 08:16:19 crc kubenswrapper[4789]: I1122 08:16:19.607872 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-grm5s\" (UniqueName: \"kubernetes.io/projected/61bf72d1-ea9f-4ad5-9dd1-7fdb4a65e8e4-kube-api-access-grm5s\") pod \"61bf72d1-ea9f-4ad5-9dd1-7fdb4a65e8e4\" (UID: \"61bf72d1-ea9f-4ad5-9dd1-7fdb4a65e8e4\") " Nov 22 08:16:19 crc kubenswrapper[4789]: I1122 08:16:19.608023 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/61bf72d1-ea9f-4ad5-9dd1-7fdb4a65e8e4-horizon-secret-key\") pod \"61bf72d1-ea9f-4ad5-9dd1-7fdb4a65e8e4\" (UID: \"61bf72d1-ea9f-4ad5-9dd1-7fdb4a65e8e4\") " Nov 22 08:16:19 crc kubenswrapper[4789]: I1122 08:16:19.608117 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/61bf72d1-ea9f-4ad5-9dd1-7fdb4a65e8e4-scripts" (OuterVolumeSpecName: "scripts") pod "61bf72d1-ea9f-4ad5-9dd1-7fdb4a65e8e4" (UID: "61bf72d1-ea9f-4ad5-9dd1-7fdb4a65e8e4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:16:19 crc kubenswrapper[4789]: I1122 08:16:19.608346 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/61bf72d1-ea9f-4ad5-9dd1-7fdb4a65e8e4-config-data" (OuterVolumeSpecName: "config-data") pod "61bf72d1-ea9f-4ad5-9dd1-7fdb4a65e8e4" (UID: "61bf72d1-ea9f-4ad5-9dd1-7fdb4a65e8e4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:16:19 crc kubenswrapper[4789]: I1122 08:16:19.609106 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/61bf72d1-ea9f-4ad5-9dd1-7fdb4a65e8e4-logs" (OuterVolumeSpecName: "logs") pod "61bf72d1-ea9f-4ad5-9dd1-7fdb4a65e8e4" (UID: "61bf72d1-ea9f-4ad5-9dd1-7fdb4a65e8e4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:16:19 crc kubenswrapper[4789]: I1122 08:16:19.609386 4789 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/61bf72d1-ea9f-4ad5-9dd1-7fdb4a65e8e4-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:16:19 crc kubenswrapper[4789]: I1122 08:16:19.609414 4789 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/61bf72d1-ea9f-4ad5-9dd1-7fdb4a65e8e4-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:16:19 crc kubenswrapper[4789]: I1122 08:16:19.609428 4789 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/61bf72d1-ea9f-4ad5-9dd1-7fdb4a65e8e4-logs\") on node \"crc\" DevicePath \"\"" Nov 22 08:16:19 crc kubenswrapper[4789]: I1122 08:16:19.612905 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/61bf72d1-ea9f-4ad5-9dd1-7fdb4a65e8e4-kube-api-access-grm5s" (OuterVolumeSpecName: "kube-api-access-grm5s") pod "61bf72d1-ea9f-4ad5-9dd1-7fdb4a65e8e4" (UID: "61bf72d1-ea9f-4ad5-9dd1-7fdb4a65e8e4"). InnerVolumeSpecName "kube-api-access-grm5s". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:16:19 crc kubenswrapper[4789]: I1122 08:16:19.613343 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61bf72d1-ea9f-4ad5-9dd1-7fdb4a65e8e4-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "61bf72d1-ea9f-4ad5-9dd1-7fdb4a65e8e4" (UID: "61bf72d1-ea9f-4ad5-9dd1-7fdb4a65e8e4"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:16:19 crc kubenswrapper[4789]: I1122 08:16:19.711471 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-grm5s\" (UniqueName: \"kubernetes.io/projected/61bf72d1-ea9f-4ad5-9dd1-7fdb4a65e8e4-kube-api-access-grm5s\") on node \"crc\" DevicePath \"\"" Nov 22 08:16:19 crc kubenswrapper[4789]: I1122 08:16:19.711506 4789 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/61bf72d1-ea9f-4ad5-9dd1-7fdb4a65e8e4-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Nov 22 08:16:19 crc kubenswrapper[4789]: E1122 08:16:19.837128 4789 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api@sha256:4c93a5cccb9971e24f05daf93b3aa11ba71752bc3469a1a1a2c4906f92f69645" Nov 22 08:16:19 crc kubenswrapper[4789]: E1122 08:16:19.837367 4789 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api@sha256:4c93a5cccb9971e24f05daf93b3aa11ba71752bc3469a1a1a2c4906f92f69645,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9slr4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-r6nst_openstack(f3ee5cf5-1275-448f-8713-e2b5a0d24439): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 22 08:16:19 crc kubenswrapper[4789]: E1122 08:16:19.838561 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-r6nst" podUID="f3ee5cf5-1275-448f-8713-e2b5a0d24439" Nov 22 08:16:19 crc kubenswrapper[4789]: I1122 08:16:19.903792 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c476d78c5-djb2g" Nov 22 08:16:19 crc kubenswrapper[4789]: I1122 08:16:19.910125 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6ddbf95fdc-64jxv" Nov 22 08:16:19 crc kubenswrapper[4789]: I1122 08:16:19.915710 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-64b577d9f-n8w7t" Nov 22 08:16:20 crc kubenswrapper[4789]: I1122 08:16:20.015083 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0cf0fb02-e586-4daa-88c1-6556b376c02a-config\") pod \"0cf0fb02-e586-4daa-88c1-6556b376c02a\" (UID: \"0cf0fb02-e586-4daa-88c1-6556b376c02a\") " Nov 22 08:16:20 crc kubenswrapper[4789]: I1122 08:16:20.015143 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0cf0fb02-e586-4daa-88c1-6556b376c02a-ovsdbserver-nb\") pod \"0cf0fb02-e586-4daa-88c1-6556b376c02a\" (UID: \"0cf0fb02-e586-4daa-88c1-6556b376c02a\") " Nov 22 08:16:20 crc kubenswrapper[4789]: I1122 08:16:20.015190 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xhkvg\" (UniqueName: \"kubernetes.io/projected/0cf0fb02-e586-4daa-88c1-6556b376c02a-kube-api-access-xhkvg\") pod \"0cf0fb02-e586-4daa-88c1-6556b376c02a\" (UID: \"0cf0fb02-e586-4daa-88c1-6556b376c02a\") " Nov 22 08:16:20 crc kubenswrapper[4789]: I1122 08:16:20.015271 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/e3b7f50f-961d-472f-9423-18ab69fcdaef-horizon-secret-key\") pod \"e3b7f50f-961d-472f-9423-18ab69fcdaef\" (UID: \"e3b7f50f-961d-472f-9423-18ab69fcdaef\") " Nov 22 08:16:20 crc kubenswrapper[4789]: I1122 08:16:20.015299 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/eea1fd9b-942e-4d60-80ae-1fca761f5a32-horizon-secret-key\") pod \"eea1fd9b-942e-4d60-80ae-1fca761f5a32\" (UID: \"eea1fd9b-942e-4d60-80ae-1fca761f5a32\") " Nov 22 08:16:20 crc kubenswrapper[4789]: I1122 08:16:20.015321 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eea1fd9b-942e-4d60-80ae-1fca761f5a32-logs\") pod \"eea1fd9b-942e-4d60-80ae-1fca761f5a32\" (UID: \"eea1fd9b-942e-4d60-80ae-1fca761f5a32\") " Nov 22 08:16:20 crc kubenswrapper[4789]: I1122 08:16:20.015380 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0cf0fb02-e586-4daa-88c1-6556b376c02a-dns-svc\") pod \"0cf0fb02-e586-4daa-88c1-6556b376c02a\" (UID: \"0cf0fb02-e586-4daa-88c1-6556b376c02a\") " Nov 22 08:16:20 crc kubenswrapper[4789]: I1122 08:16:20.015411 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rbbt8\" (UniqueName: \"kubernetes.io/projected/eea1fd9b-942e-4d60-80ae-1fca761f5a32-kube-api-access-rbbt8\") pod \"eea1fd9b-942e-4d60-80ae-1fca761f5a32\" (UID: \"eea1fd9b-942e-4d60-80ae-1fca761f5a32\") " Nov 22 08:16:20 crc kubenswrapper[4789]: I1122 08:16:20.015430 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e3b7f50f-961d-472f-9423-18ab69fcdaef-scripts\") pod \"e3b7f50f-961d-472f-9423-18ab69fcdaef\" (UID: \"e3b7f50f-961d-472f-9423-18ab69fcdaef\") " Nov 22 08:16:20 crc kubenswrapper[4789]: I1122 08:16:20.015474 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e3b7f50f-961d-472f-9423-18ab69fcdaef-config-data\") pod \"e3b7f50f-961d-472f-9423-18ab69fcdaef\" (UID: \"e3b7f50f-961d-472f-9423-18ab69fcdaef\") " Nov 22 08:16:20 crc kubenswrapper[4789]: I1122 08:16:20.015490 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cswpl\" (UniqueName: \"kubernetes.io/projected/e3b7f50f-961d-472f-9423-18ab69fcdaef-kube-api-access-cswpl\") pod \"e3b7f50f-961d-472f-9423-18ab69fcdaef\" (UID: \"e3b7f50f-961d-472f-9423-18ab69fcdaef\") " Nov 22 08:16:20 crc kubenswrapper[4789]: I1122 08:16:20.015519 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/eea1fd9b-942e-4d60-80ae-1fca761f5a32-scripts\") pod \"eea1fd9b-942e-4d60-80ae-1fca761f5a32\" (UID: \"eea1fd9b-942e-4d60-80ae-1fca761f5a32\") " Nov 22 08:16:20 crc kubenswrapper[4789]: I1122 08:16:20.015545 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0cf0fb02-e586-4daa-88c1-6556b376c02a-ovsdbserver-sb\") pod \"0cf0fb02-e586-4daa-88c1-6556b376c02a\" (UID: \"0cf0fb02-e586-4daa-88c1-6556b376c02a\") " Nov 22 08:16:20 crc kubenswrapper[4789]: I1122 08:16:20.015578 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e3b7f50f-961d-472f-9423-18ab69fcdaef-logs\") pod \"e3b7f50f-961d-472f-9423-18ab69fcdaef\" (UID: \"e3b7f50f-961d-472f-9423-18ab69fcdaef\") " Nov 22 08:16:20 crc kubenswrapper[4789]: I1122 08:16:20.015614 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/eea1fd9b-942e-4d60-80ae-1fca761f5a32-config-data\") pod \"eea1fd9b-942e-4d60-80ae-1fca761f5a32\" (UID: \"eea1fd9b-942e-4d60-80ae-1fca761f5a32\") " Nov 22 08:16:20 crc kubenswrapper[4789]: I1122 08:16:20.018255 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e3b7f50f-961d-472f-9423-18ab69fcdaef-logs" (OuterVolumeSpecName: "logs") pod "e3b7f50f-961d-472f-9423-18ab69fcdaef" (UID: "e3b7f50f-961d-472f-9423-18ab69fcdaef"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:16:20 crc kubenswrapper[4789]: I1122 08:16:20.018324 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eea1fd9b-942e-4d60-80ae-1fca761f5a32-scripts" (OuterVolumeSpecName: "scripts") pod "eea1fd9b-942e-4d60-80ae-1fca761f5a32" (UID: "eea1fd9b-942e-4d60-80ae-1fca761f5a32"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:16:20 crc kubenswrapper[4789]: I1122 08:16:20.018415 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eea1fd9b-942e-4d60-80ae-1fca761f5a32-config-data" (OuterVolumeSpecName: "config-data") pod "eea1fd9b-942e-4d60-80ae-1fca761f5a32" (UID: "eea1fd9b-942e-4d60-80ae-1fca761f5a32"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:16:20 crc kubenswrapper[4789]: I1122 08:16:20.019262 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eea1fd9b-942e-4d60-80ae-1fca761f5a32-logs" (OuterVolumeSpecName: "logs") pod "eea1fd9b-942e-4d60-80ae-1fca761f5a32" (UID: "eea1fd9b-942e-4d60-80ae-1fca761f5a32"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:16:20 crc kubenswrapper[4789]: I1122 08:16:20.020477 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eea1fd9b-942e-4d60-80ae-1fca761f5a32-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "eea1fd9b-942e-4d60-80ae-1fca761f5a32" (UID: "eea1fd9b-942e-4d60-80ae-1fca761f5a32"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:16:20 crc kubenswrapper[4789]: I1122 08:16:20.020582 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3b7f50f-961d-472f-9423-18ab69fcdaef-kube-api-access-cswpl" (OuterVolumeSpecName: "kube-api-access-cswpl") pod "e3b7f50f-961d-472f-9423-18ab69fcdaef" (UID: "e3b7f50f-961d-472f-9423-18ab69fcdaef"). InnerVolumeSpecName "kube-api-access-cswpl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:16:20 crc kubenswrapper[4789]: I1122 08:16:20.024123 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0cf0fb02-e586-4daa-88c1-6556b376c02a-kube-api-access-xhkvg" (OuterVolumeSpecName: "kube-api-access-xhkvg") pod "0cf0fb02-e586-4daa-88c1-6556b376c02a" (UID: "0cf0fb02-e586-4daa-88c1-6556b376c02a"). InnerVolumeSpecName "kube-api-access-xhkvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:16:20 crc kubenswrapper[4789]: I1122 08:16:20.025391 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3b7f50f-961d-472f-9423-18ab69fcdaef-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "e3b7f50f-961d-472f-9423-18ab69fcdaef" (UID: "e3b7f50f-961d-472f-9423-18ab69fcdaef"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:16:20 crc kubenswrapper[4789]: I1122 08:16:20.025605 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eea1fd9b-942e-4d60-80ae-1fca761f5a32-kube-api-access-rbbt8" (OuterVolumeSpecName: "kube-api-access-rbbt8") pod "eea1fd9b-942e-4d60-80ae-1fca761f5a32" (UID: "eea1fd9b-942e-4d60-80ae-1fca761f5a32"). InnerVolumeSpecName "kube-api-access-rbbt8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:16:20 crc kubenswrapper[4789]: I1122 08:16:20.053941 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e3b7f50f-961d-472f-9423-18ab69fcdaef-scripts" (OuterVolumeSpecName: "scripts") pod "e3b7f50f-961d-472f-9423-18ab69fcdaef" (UID: "e3b7f50f-961d-472f-9423-18ab69fcdaef"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:16:20 crc kubenswrapper[4789]: I1122 08:16:20.054364 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e3b7f50f-961d-472f-9423-18ab69fcdaef-config-data" (OuterVolumeSpecName: "config-data") pod "e3b7f50f-961d-472f-9423-18ab69fcdaef" (UID: "e3b7f50f-961d-472f-9423-18ab69fcdaef"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:16:20 crc kubenswrapper[4789]: I1122 08:16:20.057140 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0cf0fb02-e586-4daa-88c1-6556b376c02a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "0cf0fb02-e586-4daa-88c1-6556b376c02a" (UID: "0cf0fb02-e586-4daa-88c1-6556b376c02a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:16:20 crc kubenswrapper[4789]: I1122 08:16:20.060672 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0cf0fb02-e586-4daa-88c1-6556b376c02a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "0cf0fb02-e586-4daa-88c1-6556b376c02a" (UID: "0cf0fb02-e586-4daa-88c1-6556b376c02a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:16:20 crc kubenswrapper[4789]: I1122 08:16:20.061789 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0cf0fb02-e586-4daa-88c1-6556b376c02a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0cf0fb02-e586-4daa-88c1-6556b376c02a" (UID: "0cf0fb02-e586-4daa-88c1-6556b376c02a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:16:20 crc kubenswrapper[4789]: I1122 08:16:20.067148 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0cf0fb02-e586-4daa-88c1-6556b376c02a-config" (OuterVolumeSpecName: "config") pod "0cf0fb02-e586-4daa-88c1-6556b376c02a" (UID: "0cf0fb02-e586-4daa-88c1-6556b376c02a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:16:20 crc kubenswrapper[4789]: I1122 08:16:20.081553 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6d4f97b8d9-fqgm7" event={"ID":"61bf72d1-ea9f-4ad5-9dd1-7fdb4a65e8e4","Type":"ContainerDied","Data":"1454f79ee94e414a3a4d93406614952af4ebf9b7d7bdc182754538a0728d40c1"} Nov 22 08:16:20 crc kubenswrapper[4789]: I1122 08:16:20.081587 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6d4f97b8d9-fqgm7" Nov 22 08:16:20 crc kubenswrapper[4789]: I1122 08:16:20.084616 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c476d78c5-djb2g" Nov 22 08:16:20 crc kubenswrapper[4789]: I1122 08:16:20.084730 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c476d78c5-djb2g" event={"ID":"0cf0fb02-e586-4daa-88c1-6556b376c02a","Type":"ContainerDied","Data":"7721d00be02c91565b524c85dff133a1bb86958e34fd9092e9d73d985f6ce81b"} Nov 22 08:16:20 crc kubenswrapper[4789]: I1122 08:16:20.090282 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6ddbf95fdc-64jxv" event={"ID":"eea1fd9b-942e-4d60-80ae-1fca761f5a32","Type":"ContainerDied","Data":"6d889b3a7d37cdfcf2fe5cc4b099872c69d7353f7064764a959103e8f920d9cc"} Nov 22 08:16:20 crc kubenswrapper[4789]: I1122 08:16:20.090297 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6ddbf95fdc-64jxv" Nov 22 08:16:20 crc kubenswrapper[4789]: I1122 08:16:20.092243 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-64b577d9f-n8w7t" Nov 22 08:16:20 crc kubenswrapper[4789]: I1122 08:16:20.093624 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-64b577d9f-n8w7t" event={"ID":"e3b7f50f-961d-472f-9423-18ab69fcdaef","Type":"ContainerDied","Data":"43b316398a79a1aad6168c1db602fd3d3dc6f7a04a903fb03d3ed9efdd69fcfd"} Nov 22 08:16:20 crc kubenswrapper[4789]: E1122 08:16:20.094094 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api@sha256:4c93a5cccb9971e24f05daf93b3aa11ba71752bc3469a1a1a2c4906f92f69645\\\"\"" pod="openstack/barbican-db-sync-r6nst" podUID="f3ee5cf5-1275-448f-8713-e2b5a0d24439" Nov 22 08:16:20 crc kubenswrapper[4789]: I1122 08:16:20.118575 4789 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/eea1fd9b-942e-4d60-80ae-1fca761f5a32-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:16:20 crc kubenswrapper[4789]: I1122 08:16:20.118613 4789 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0cf0fb02-e586-4daa-88c1-6556b376c02a-config\") on node \"crc\" DevicePath \"\"" Nov 22 08:16:20 crc kubenswrapper[4789]: I1122 08:16:20.118625 4789 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0cf0fb02-e586-4daa-88c1-6556b376c02a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 22 08:16:20 crc kubenswrapper[4789]: I1122 08:16:20.118639 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xhkvg\" (UniqueName: \"kubernetes.io/projected/0cf0fb02-e586-4daa-88c1-6556b376c02a-kube-api-access-xhkvg\") on node \"crc\" DevicePath \"\"" Nov 22 08:16:20 crc kubenswrapper[4789]: I1122 08:16:20.118652 4789 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/e3b7f50f-961d-472f-9423-18ab69fcdaef-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Nov 22 08:16:20 crc kubenswrapper[4789]: I1122 08:16:20.118664 4789 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/eea1fd9b-942e-4d60-80ae-1fca761f5a32-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Nov 22 08:16:20 crc kubenswrapper[4789]: I1122 08:16:20.118675 4789 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eea1fd9b-942e-4d60-80ae-1fca761f5a32-logs\") on node \"crc\" DevicePath \"\"" Nov 22 08:16:20 crc kubenswrapper[4789]: I1122 08:16:20.118685 4789 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0cf0fb02-e586-4daa-88c1-6556b376c02a-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 22 08:16:20 crc kubenswrapper[4789]: I1122 08:16:20.118696 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rbbt8\" (UniqueName: \"kubernetes.io/projected/eea1fd9b-942e-4d60-80ae-1fca761f5a32-kube-api-access-rbbt8\") on node \"crc\" DevicePath \"\"" Nov 22 08:16:20 crc kubenswrapper[4789]: I1122 08:16:20.118708 4789 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e3b7f50f-961d-472f-9423-18ab69fcdaef-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:16:20 crc kubenswrapper[4789]: I1122 08:16:20.118717 4789 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e3b7f50f-961d-472f-9423-18ab69fcdaef-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:16:20 crc kubenswrapper[4789]: I1122 08:16:20.118731 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cswpl\" (UniqueName: \"kubernetes.io/projected/e3b7f50f-961d-472f-9423-18ab69fcdaef-kube-api-access-cswpl\") on node \"crc\" DevicePath \"\"" Nov 22 08:16:20 crc kubenswrapper[4789]: I1122 08:16:20.118740 4789 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/eea1fd9b-942e-4d60-80ae-1fca761f5a32-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:16:20 crc kubenswrapper[4789]: I1122 08:16:20.118770 4789 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0cf0fb02-e586-4daa-88c1-6556b376c02a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 22 08:16:20 crc kubenswrapper[4789]: I1122 08:16:20.118780 4789 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e3b7f50f-961d-472f-9423-18ab69fcdaef-logs\") on node \"crc\" DevicePath \"\"" Nov 22 08:16:20 crc kubenswrapper[4789]: I1122 08:16:20.126668 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-6d4f97b8d9-fqgm7"] Nov 22 08:16:20 crc kubenswrapper[4789]: I1122 08:16:20.134663 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-6d4f97b8d9-fqgm7"] Nov 22 08:16:20 crc kubenswrapper[4789]: I1122 08:16:20.140992 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c476d78c5-djb2g"] Nov 22 08:16:20 crc kubenswrapper[4789]: I1122 08:16:20.146692 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c476d78c5-djb2g"] Nov 22 08:16:20 crc kubenswrapper[4789]: I1122 08:16:20.176312 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-64b577d9f-n8w7t"] Nov 22 08:16:20 crc kubenswrapper[4789]: I1122 08:16:20.182550 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-64b577d9f-n8w7t"] Nov 22 08:16:20 crc kubenswrapper[4789]: I1122 08:16:20.214772 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-6ddbf95fdc-64jxv"] Nov 22 08:16:20 crc kubenswrapper[4789]: I1122 08:16:20.235348 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-6ddbf95fdc-64jxv"] Nov 22 08:16:21 crc kubenswrapper[4789]: I1122 08:16:21.975695 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0cf0fb02-e586-4daa-88c1-6556b376c02a" path="/var/lib/kubelet/pods/0cf0fb02-e586-4daa-88c1-6556b376c02a/volumes" Nov 22 08:16:21 crc kubenswrapper[4789]: I1122 08:16:21.977165 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="61bf72d1-ea9f-4ad5-9dd1-7fdb4a65e8e4" path="/var/lib/kubelet/pods/61bf72d1-ea9f-4ad5-9dd1-7fdb4a65e8e4/volumes" Nov 22 08:16:21 crc kubenswrapper[4789]: I1122 08:16:21.977774 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e3b7f50f-961d-472f-9423-18ab69fcdaef" path="/var/lib/kubelet/pods/e3b7f50f-961d-472f-9423-18ab69fcdaef/volumes" Nov 22 08:16:21 crc kubenswrapper[4789]: I1122 08:16:21.978444 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eea1fd9b-942e-4d60-80ae-1fca761f5a32" path="/var/lib/kubelet/pods/eea1fd9b-942e-4d60-80ae-1fca761f5a32/volumes" Nov 22 08:16:22 crc kubenswrapper[4789]: I1122 08:16:22.385058 4789 scope.go:117] "RemoveContainer" containerID="3bba78e22b91a98d5a6fd8c4d50cedda778719b31f2900449fafb4e224b7a2bb" Nov 22 08:16:22 crc kubenswrapper[4789]: I1122 08:16:22.561356 4789 scope.go:117] "RemoveContainer" containerID="98e61e9ab6d1be90a45fd2a32392caf89eebaff16982d2f7347fb57dd8948bdb" Nov 22 08:16:22 crc kubenswrapper[4789]: W1122 08:16:22.817027 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd9fb66f7_ee5b_453a_a69b_56fb9337d898.slice/crio-6ae72ef113b6dcabf5db9821cd9c9005003528ea2c2bc0b3b8b316df4a822d82 WatchSource:0}: Error finding container 6ae72ef113b6dcabf5db9821cd9c9005003528ea2c2bc0b3b8b316df4a822d82: Status 404 returned error can't find the container with id 6ae72ef113b6dcabf5db9821cd9c9005003528ea2c2bc0b3b8b316df4a822d82 Nov 22 08:16:22 crc kubenswrapper[4789]: I1122 08:16:22.818153 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7b778f8c8-6rl4f"] Nov 22 08:16:22 crc kubenswrapper[4789]: I1122 08:16:22.880854 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-rnzhp"] Nov 22 08:16:22 crc kubenswrapper[4789]: W1122 08:16:22.887601 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod56fd3074_a15c_4024_baf9_f784a50e14a5.slice/crio-3b3a905c1c5482e887bf3be488fe68c9df53059d564d5f9051d737b756870978 WatchSource:0}: Error finding container 3b3a905c1c5482e887bf3be488fe68c9df53059d564d5f9051d737b756870978: Status 404 returned error can't find the container with id 3b3a905c1c5482e887bf3be488fe68c9df53059d564d5f9051d737b756870978 Nov 22 08:16:22 crc kubenswrapper[4789]: I1122 08:16:22.889451 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5b56df9788-wsmhq"] Nov 22 08:16:23 crc kubenswrapper[4789]: I1122 08:16:23.062244 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5c476d78c5-djb2g" podUID="0cf0fb02-e586-4daa-88c1-6556b376c02a" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.110:5353: i/o timeout" Nov 22 08:16:23 crc kubenswrapper[4789]: I1122 08:16:23.117389 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-rnzhp" event={"ID":"1f1fc498-45f3-4bec-a4a5-b9722d2445ea","Type":"ContainerStarted","Data":"7e773a345db883e25881939a3967c9cdb9cd1884e7961e162cd4706407a1b070"} Nov 22 08:16:23 crc kubenswrapper[4789]: I1122 08:16:23.118544 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7b778f8c8-6rl4f" event={"ID":"d9fb66f7-ee5b-453a-a69b-56fb9337d898","Type":"ContainerStarted","Data":"6ae72ef113b6dcabf5db9821cd9c9005003528ea2c2bc0b3b8b316df4a822d82"} Nov 22 08:16:23 crc kubenswrapper[4789]: I1122 08:16:23.120293 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5b56df9788-wsmhq" event={"ID":"56fd3074-a15c-4024-baf9-f784a50e14a5","Type":"ContainerStarted","Data":"3b3a905c1c5482e887bf3be488fe68c9df53059d564d5f9051d737b756870978"} Nov 22 08:16:24 crc kubenswrapper[4789]: E1122 08:16:24.073232 4789 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:37d64e0a00c54e71a4c1fcbbbf7e832f6886ffd03c9a02b6ee3ca48fabc30879" Nov 22 08:16:24 crc kubenswrapper[4789]: E1122 08:16:24.073739 4789 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:37d64e0a00c54e71a4c1fcbbbf7e832f6886ffd03c9a02b6ee3ca48fabc30879,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-t9qz8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-lg59m_openstack(6bf8aa3e-a7e7-4f0c-a542-d668a5623707): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 22 08:16:24 crc kubenswrapper[4789]: E1122 08:16:24.074924 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-lg59m" podUID="6bf8aa3e-a7e7-4f0c-a542-d668a5623707" Nov 22 08:16:24 crc kubenswrapper[4789]: I1122 08:16:24.135286 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" event={"ID":"58aa8071-7ecc-4692-8789-2db04ec70647","Type":"ContainerStarted","Data":"3d1ad86d18575ad837bca3f627df48477a6e890e1d7c69736133f6389315bc07"} Nov 22 08:16:24 crc kubenswrapper[4789]: I1122 08:16:24.142348 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-rnzhp" event={"ID":"1f1fc498-45f3-4bec-a4a5-b9722d2445ea","Type":"ContainerStarted","Data":"e7c9611deeb5f070f55bebba5b8617f188d07f83303c25a7a62ae83a003facef"} Nov 22 08:16:24 crc kubenswrapper[4789]: I1122 08:16:24.145449 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"94d7bb01-64a7-442a-aee4-e958d8511b76","Type":"ContainerStarted","Data":"5c1bd328b27a7fd1cdca6ff567666e218a70a3b926831af09bf3a542194a3a1b"} Nov 22 08:16:24 crc kubenswrapper[4789]: I1122 08:16:24.202128 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-rnzhp" podStartSLOduration=10.202105075 podStartE2EDuration="10.202105075s" podCreationTimestamp="2025-11-22 08:16:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:16:24.195034602 +0000 UTC m=+1198.429434875" watchObservedRunningTime="2025-11-22 08:16:24.202105075 +0000 UTC m=+1198.436505348" Nov 22 08:16:24 crc kubenswrapper[4789]: E1122 08:16:24.288979 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:37d64e0a00c54e71a4c1fcbbbf7e832f6886ffd03c9a02b6ee3ca48fabc30879\\\"\"" pod="openstack/cinder-db-sync-lg59m" podUID="6bf8aa3e-a7e7-4f0c-a542-d668a5623707" Nov 22 08:16:25 crc kubenswrapper[4789]: I1122 08:16:25.161355 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7b778f8c8-6rl4f" event={"ID":"d9fb66f7-ee5b-453a-a69b-56fb9337d898","Type":"ContainerStarted","Data":"6da4d7faf130e5fcd2e9848abcff026ce044791da6d6b31909ba5a27698cadea"} Nov 22 08:16:25 crc kubenswrapper[4789]: I1122 08:16:25.162016 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7b778f8c8-6rl4f" event={"ID":"d9fb66f7-ee5b-453a-a69b-56fb9337d898","Type":"ContainerStarted","Data":"3a35e0ecc68b27350fd4527c752399a0afd99df5c00bb3d0049bcc9c734cb729"} Nov 22 08:16:25 crc kubenswrapper[4789]: I1122 08:16:25.165184 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-2ss46" event={"ID":"2eb187f8-1892-4d72-b442-43592e75d489","Type":"ContainerStarted","Data":"848a353afae49b131c8b3082febbe24a3efdd0d7749888b4a1aa1c94bf99fb13"} Nov 22 08:16:25 crc kubenswrapper[4789]: I1122 08:16:25.174224 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5b56df9788-wsmhq" event={"ID":"56fd3074-a15c-4024-baf9-f784a50e14a5","Type":"ContainerStarted","Data":"2e99dc56ab31ddc560842a39626551056b3b1db94ece09fb54c781cd0d891832"} Nov 22 08:16:25 crc kubenswrapper[4789]: I1122 08:16:25.174272 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5b56df9788-wsmhq" event={"ID":"56fd3074-a15c-4024-baf9-f784a50e14a5","Type":"ContainerStarted","Data":"93ff7a096adcf0bb8be87de88e1df3ddfb208ee02cbae6c682c5da91f25f03f9"} Nov 22 08:16:25 crc kubenswrapper[4789]: I1122 08:16:25.191851 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-7b778f8c8-6rl4f" podStartSLOduration=25.680657208 podStartE2EDuration="27.191833389s" podCreationTimestamp="2025-11-22 08:15:58 +0000 UTC" firstStartedPulling="2025-11-22 08:16:22.818807688 +0000 UTC m=+1197.053207961" lastFinishedPulling="2025-11-22 08:16:24.329983859 +0000 UTC m=+1198.564384142" observedRunningTime="2025-11-22 08:16:25.185166327 +0000 UTC m=+1199.419566600" watchObservedRunningTime="2025-11-22 08:16:25.191833389 +0000 UTC m=+1199.426233662" Nov 22 08:16:25 crc kubenswrapper[4789]: I1122 08:16:25.213651 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-2ss46" podStartSLOduration=2.647980909 podStartE2EDuration="36.213628453s" podCreationTimestamp="2025-11-22 08:15:49 +0000 UTC" firstStartedPulling="2025-11-22 08:15:50.763892212 +0000 UTC m=+1164.998292485" lastFinishedPulling="2025-11-22 08:16:24.329539736 +0000 UTC m=+1198.563940029" observedRunningTime="2025-11-22 08:16:25.205047129 +0000 UTC m=+1199.439447402" watchObservedRunningTime="2025-11-22 08:16:25.213628453 +0000 UTC m=+1199.448028726" Nov 22 08:16:26 crc kubenswrapper[4789]: I1122 08:16:26.190120 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"94d7bb01-64a7-442a-aee4-e958d8511b76","Type":"ContainerStarted","Data":"b8f37fba2551f2fc622fc288b287229e43574f1d9231f7967747f9645db2d0e9"} Nov 22 08:16:28 crc kubenswrapper[4789]: I1122 08:16:28.452141 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-5b56df9788-wsmhq" Nov 22 08:16:28 crc kubenswrapper[4789]: I1122 08:16:28.454250 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-5b56df9788-wsmhq" Nov 22 08:16:28 crc kubenswrapper[4789]: I1122 08:16:28.580057 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7b778f8c8-6rl4f" Nov 22 08:16:28 crc kubenswrapper[4789]: I1122 08:16:28.580542 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7b778f8c8-6rl4f" Nov 22 08:16:33 crc kubenswrapper[4789]: I1122 08:16:33.247546 4789 generic.go:334] "Generic (PLEG): container finished" podID="1f1fc498-45f3-4bec-a4a5-b9722d2445ea" containerID="e7c9611deeb5f070f55bebba5b8617f188d07f83303c25a7a62ae83a003facef" exitCode=0 Nov 22 08:16:33 crc kubenswrapper[4789]: I1122 08:16:33.247620 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-rnzhp" event={"ID":"1f1fc498-45f3-4bec-a4a5-b9722d2445ea","Type":"ContainerDied","Data":"e7c9611deeb5f070f55bebba5b8617f188d07f83303c25a7a62ae83a003facef"} Nov 22 08:16:33 crc kubenswrapper[4789]: I1122 08:16:33.251628 4789 generic.go:334] "Generic (PLEG): container finished" podID="2eb187f8-1892-4d72-b442-43592e75d489" containerID="848a353afae49b131c8b3082febbe24a3efdd0d7749888b4a1aa1c94bf99fb13" exitCode=0 Nov 22 08:16:33 crc kubenswrapper[4789]: I1122 08:16:33.251664 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-2ss46" event={"ID":"2eb187f8-1892-4d72-b442-43592e75d489","Type":"ContainerDied","Data":"848a353afae49b131c8b3082febbe24a3efdd0d7749888b4a1aa1c94bf99fb13"} Nov 22 08:16:33 crc kubenswrapper[4789]: I1122 08:16:33.272221 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-5b56df9788-wsmhq" podStartSLOduration=33.832752074 podStartE2EDuration="35.27219736s" podCreationTimestamp="2025-11-22 08:15:58 +0000 UTC" firstStartedPulling="2025-11-22 08:16:22.890193093 +0000 UTC m=+1197.124593366" lastFinishedPulling="2025-11-22 08:16:24.329638379 +0000 UTC m=+1198.564038652" observedRunningTime="2025-11-22 08:16:25.233488684 +0000 UTC m=+1199.467888977" watchObservedRunningTime="2025-11-22 08:16:33.27219736 +0000 UTC m=+1207.506597633" Nov 22 08:16:36 crc kubenswrapper[4789]: I1122 08:16:36.290913 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-rnzhp" event={"ID":"1f1fc498-45f3-4bec-a4a5-b9722d2445ea","Type":"ContainerDied","Data":"7e773a345db883e25881939a3967c9cdb9cd1884e7961e162cd4706407a1b070"} Nov 22 08:16:36 crc kubenswrapper[4789]: I1122 08:16:36.291658 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7e773a345db883e25881939a3967c9cdb9cd1884e7961e162cd4706407a1b070" Nov 22 08:16:36 crc kubenswrapper[4789]: I1122 08:16:36.293951 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-2ss46" event={"ID":"2eb187f8-1892-4d72-b442-43592e75d489","Type":"ContainerDied","Data":"d7c6edc5fd50c8ead7f54244e5c18905bace2602ed1c684bbfd4be939bcd4f07"} Nov 22 08:16:36 crc kubenswrapper[4789]: I1122 08:16:36.294006 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d7c6edc5fd50c8ead7f54244e5c18905bace2602ed1c684bbfd4be939bcd4f07" Nov 22 08:16:36 crc kubenswrapper[4789]: I1122 08:16:36.322877 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-2ss46" Nov 22 08:16:36 crc kubenswrapper[4789]: I1122 08:16:36.332670 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-rnzhp" Nov 22 08:16:36 crc kubenswrapper[4789]: I1122 08:16:36.429815 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2eb187f8-1892-4d72-b442-43592e75d489-scripts\") pod \"2eb187f8-1892-4d72-b442-43592e75d489\" (UID: \"2eb187f8-1892-4d72-b442-43592e75d489\") " Nov 22 08:16:36 crc kubenswrapper[4789]: I1122 08:16:36.429871 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2eb187f8-1892-4d72-b442-43592e75d489-config-data\") pod \"2eb187f8-1892-4d72-b442-43592e75d489\" (UID: \"2eb187f8-1892-4d72-b442-43592e75d489\") " Nov 22 08:16:36 crc kubenswrapper[4789]: I1122 08:16:36.429954 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1f1fc498-45f3-4bec-a4a5-b9722d2445ea-fernet-keys\") pod \"1f1fc498-45f3-4bec-a4a5-b9722d2445ea\" (UID: \"1f1fc498-45f3-4bec-a4a5-b9722d2445ea\") " Nov 22 08:16:36 crc kubenswrapper[4789]: I1122 08:16:36.429984 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h867w\" (UniqueName: \"kubernetes.io/projected/2eb187f8-1892-4d72-b442-43592e75d489-kube-api-access-h867w\") pod \"2eb187f8-1892-4d72-b442-43592e75d489\" (UID: \"2eb187f8-1892-4d72-b442-43592e75d489\") " Nov 22 08:16:36 crc kubenswrapper[4789]: I1122 08:16:36.430016 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7m7kw\" (UniqueName: \"kubernetes.io/projected/1f1fc498-45f3-4bec-a4a5-b9722d2445ea-kube-api-access-7m7kw\") pod \"1f1fc498-45f3-4bec-a4a5-b9722d2445ea\" (UID: \"1f1fc498-45f3-4bec-a4a5-b9722d2445ea\") " Nov 22 08:16:36 crc kubenswrapper[4789]: I1122 08:16:36.430047 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2eb187f8-1892-4d72-b442-43592e75d489-logs\") pod \"2eb187f8-1892-4d72-b442-43592e75d489\" (UID: \"2eb187f8-1892-4d72-b442-43592e75d489\") " Nov 22 08:16:36 crc kubenswrapper[4789]: I1122 08:16:36.430424 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1f1fc498-45f3-4bec-a4a5-b9722d2445ea-credential-keys\") pod \"1f1fc498-45f3-4bec-a4a5-b9722d2445ea\" (UID: \"1f1fc498-45f3-4bec-a4a5-b9722d2445ea\") " Nov 22 08:16:36 crc kubenswrapper[4789]: I1122 08:16:36.430760 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2eb187f8-1892-4d72-b442-43592e75d489-logs" (OuterVolumeSpecName: "logs") pod "2eb187f8-1892-4d72-b442-43592e75d489" (UID: "2eb187f8-1892-4d72-b442-43592e75d489"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:16:36 crc kubenswrapper[4789]: I1122 08:16:36.430831 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1f1fc498-45f3-4bec-a4a5-b9722d2445ea-scripts\") pod \"1f1fc498-45f3-4bec-a4a5-b9722d2445ea\" (UID: \"1f1fc498-45f3-4bec-a4a5-b9722d2445ea\") " Nov 22 08:16:36 crc kubenswrapper[4789]: I1122 08:16:36.431143 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2eb187f8-1892-4d72-b442-43592e75d489-combined-ca-bundle\") pod \"2eb187f8-1892-4d72-b442-43592e75d489\" (UID: \"2eb187f8-1892-4d72-b442-43592e75d489\") " Nov 22 08:16:36 crc kubenswrapper[4789]: I1122 08:16:36.431169 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f1fc498-45f3-4bec-a4a5-b9722d2445ea-config-data\") pod \"1f1fc498-45f3-4bec-a4a5-b9722d2445ea\" (UID: \"1f1fc498-45f3-4bec-a4a5-b9722d2445ea\") " Nov 22 08:16:36 crc kubenswrapper[4789]: I1122 08:16:36.431212 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f1fc498-45f3-4bec-a4a5-b9722d2445ea-combined-ca-bundle\") pod \"1f1fc498-45f3-4bec-a4a5-b9722d2445ea\" (UID: \"1f1fc498-45f3-4bec-a4a5-b9722d2445ea\") " Nov 22 08:16:36 crc kubenswrapper[4789]: I1122 08:16:36.431869 4789 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2eb187f8-1892-4d72-b442-43592e75d489-logs\") on node \"crc\" DevicePath \"\"" Nov 22 08:16:36 crc kubenswrapper[4789]: I1122 08:16:36.434240 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1f1fc498-45f3-4bec-a4a5-b9722d2445ea-kube-api-access-7m7kw" (OuterVolumeSpecName: "kube-api-access-7m7kw") pod "1f1fc498-45f3-4bec-a4a5-b9722d2445ea" (UID: "1f1fc498-45f3-4bec-a4a5-b9722d2445ea"). InnerVolumeSpecName "kube-api-access-7m7kw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:16:36 crc kubenswrapper[4789]: I1122 08:16:36.435048 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2eb187f8-1892-4d72-b442-43592e75d489-kube-api-access-h867w" (OuterVolumeSpecName: "kube-api-access-h867w") pod "2eb187f8-1892-4d72-b442-43592e75d489" (UID: "2eb187f8-1892-4d72-b442-43592e75d489"). InnerVolumeSpecName "kube-api-access-h867w". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:16:36 crc kubenswrapper[4789]: I1122 08:16:36.435197 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f1fc498-45f3-4bec-a4a5-b9722d2445ea-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "1f1fc498-45f3-4bec-a4a5-b9722d2445ea" (UID: "1f1fc498-45f3-4bec-a4a5-b9722d2445ea"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:16:36 crc kubenswrapper[4789]: I1122 08:16:36.438250 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f1fc498-45f3-4bec-a4a5-b9722d2445ea-scripts" (OuterVolumeSpecName: "scripts") pod "1f1fc498-45f3-4bec-a4a5-b9722d2445ea" (UID: "1f1fc498-45f3-4bec-a4a5-b9722d2445ea"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:16:36 crc kubenswrapper[4789]: I1122 08:16:36.438563 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2eb187f8-1892-4d72-b442-43592e75d489-scripts" (OuterVolumeSpecName: "scripts") pod "2eb187f8-1892-4d72-b442-43592e75d489" (UID: "2eb187f8-1892-4d72-b442-43592e75d489"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:16:36 crc kubenswrapper[4789]: I1122 08:16:36.438982 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f1fc498-45f3-4bec-a4a5-b9722d2445ea-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "1f1fc498-45f3-4bec-a4a5-b9722d2445ea" (UID: "1f1fc498-45f3-4bec-a4a5-b9722d2445ea"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:16:36 crc kubenswrapper[4789]: I1122 08:16:36.458972 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2eb187f8-1892-4d72-b442-43592e75d489-config-data" (OuterVolumeSpecName: "config-data") pod "2eb187f8-1892-4d72-b442-43592e75d489" (UID: "2eb187f8-1892-4d72-b442-43592e75d489"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:16:36 crc kubenswrapper[4789]: I1122 08:16:36.459557 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2eb187f8-1892-4d72-b442-43592e75d489-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2eb187f8-1892-4d72-b442-43592e75d489" (UID: "2eb187f8-1892-4d72-b442-43592e75d489"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:16:36 crc kubenswrapper[4789]: I1122 08:16:36.462999 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f1fc498-45f3-4bec-a4a5-b9722d2445ea-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1f1fc498-45f3-4bec-a4a5-b9722d2445ea" (UID: "1f1fc498-45f3-4bec-a4a5-b9722d2445ea"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:16:36 crc kubenswrapper[4789]: I1122 08:16:36.464232 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f1fc498-45f3-4bec-a4a5-b9722d2445ea-config-data" (OuterVolumeSpecName: "config-data") pod "1f1fc498-45f3-4bec-a4a5-b9722d2445ea" (UID: "1f1fc498-45f3-4bec-a4a5-b9722d2445ea"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:16:36 crc kubenswrapper[4789]: I1122 08:16:36.533950 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h867w\" (UniqueName: \"kubernetes.io/projected/2eb187f8-1892-4d72-b442-43592e75d489-kube-api-access-h867w\") on node \"crc\" DevicePath \"\"" Nov 22 08:16:36 crc kubenswrapper[4789]: I1122 08:16:36.533982 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7m7kw\" (UniqueName: \"kubernetes.io/projected/1f1fc498-45f3-4bec-a4a5-b9722d2445ea-kube-api-access-7m7kw\") on node \"crc\" DevicePath \"\"" Nov 22 08:16:36 crc kubenswrapper[4789]: I1122 08:16:36.533992 4789 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1f1fc498-45f3-4bec-a4a5-b9722d2445ea-credential-keys\") on node \"crc\" DevicePath \"\"" Nov 22 08:16:36 crc kubenswrapper[4789]: I1122 08:16:36.534000 4789 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1f1fc498-45f3-4bec-a4a5-b9722d2445ea-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:16:36 crc kubenswrapper[4789]: I1122 08:16:36.534009 4789 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2eb187f8-1892-4d72-b442-43592e75d489-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:16:36 crc kubenswrapper[4789]: I1122 08:16:36.534017 4789 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f1fc498-45f3-4bec-a4a5-b9722d2445ea-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:16:36 crc kubenswrapper[4789]: I1122 08:16:36.534025 4789 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f1fc498-45f3-4bec-a4a5-b9722d2445ea-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:16:36 crc kubenswrapper[4789]: I1122 08:16:36.534033 4789 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2eb187f8-1892-4d72-b442-43592e75d489-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:16:36 crc kubenswrapper[4789]: I1122 08:16:36.534040 4789 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2eb187f8-1892-4d72-b442-43592e75d489-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:16:36 crc kubenswrapper[4789]: I1122 08:16:36.534048 4789 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1f1fc498-45f3-4bec-a4a5-b9722d2445ea-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 22 08:16:37 crc kubenswrapper[4789]: I1122 08:16:37.303011 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"94d7bb01-64a7-442a-aee4-e958d8511b76","Type":"ContainerStarted","Data":"8463b06d53f122d2dd4bda6d1f5b32b440b64a9e858e15291e72b4e4f847355f"} Nov 22 08:16:37 crc kubenswrapper[4789]: I1122 08:16:37.308450 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-r6nst" event={"ID":"f3ee5cf5-1275-448f-8713-e2b5a0d24439","Type":"ContainerStarted","Data":"43622e76a3aa6a505c8df9823f7041e41a83be4429150db9572330e047631f10"} Nov 22 08:16:37 crc kubenswrapper[4789]: I1122 08:16:37.313934 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-rnzhp" Nov 22 08:16:37 crc kubenswrapper[4789]: I1122 08:16:37.314016 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-2ss46" Nov 22 08:16:37 crc kubenswrapper[4789]: I1122 08:16:37.313931 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-lg59m" event={"ID":"6bf8aa3e-a7e7-4f0c-a542-d668a5623707","Type":"ContainerStarted","Data":"50273be882e06c5bebf9373f13c41c1860a0308e06ec2d3632e22d25ca316617"} Nov 22 08:16:37 crc kubenswrapper[4789]: I1122 08:16:37.334146 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-r6nst" podStartSLOduration=2.742092503 podStartE2EDuration="48.334122964s" podCreationTimestamp="2025-11-22 08:15:49 +0000 UTC" firstStartedPulling="2025-11-22 08:15:50.887263904 +0000 UTC m=+1165.121664177" lastFinishedPulling="2025-11-22 08:16:36.479294365 +0000 UTC m=+1210.713694638" observedRunningTime="2025-11-22 08:16:37.327072912 +0000 UTC m=+1211.561473195" watchObservedRunningTime="2025-11-22 08:16:37.334122964 +0000 UTC m=+1211.568523247" Nov 22 08:16:37 crc kubenswrapper[4789]: I1122 08:16:37.347572 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-lg59m" podStartSLOduration=2.197981329 podStartE2EDuration="48.34754825s" podCreationTimestamp="2025-11-22 08:15:49 +0000 UTC" firstStartedPulling="2025-11-22 08:15:50.472633447 +0000 UTC m=+1164.707033720" lastFinishedPulling="2025-11-22 08:16:36.622200368 +0000 UTC m=+1210.856600641" observedRunningTime="2025-11-22 08:16:37.343377326 +0000 UTC m=+1211.577777599" watchObservedRunningTime="2025-11-22 08:16:37.34754825 +0000 UTC m=+1211.581948523" Nov 22 08:16:37 crc kubenswrapper[4789]: I1122 08:16:37.436320 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-7846c8b766-kj5xm"] Nov 22 08:16:37 crc kubenswrapper[4789]: E1122 08:16:37.436657 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cf0fb02-e586-4daa-88c1-6556b376c02a" containerName="init" Nov 22 08:16:37 crc kubenswrapper[4789]: I1122 08:16:37.436672 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cf0fb02-e586-4daa-88c1-6556b376c02a" containerName="init" Nov 22 08:16:37 crc kubenswrapper[4789]: E1122 08:16:37.436698 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f1fc498-45f3-4bec-a4a5-b9722d2445ea" containerName="keystone-bootstrap" Nov 22 08:16:37 crc kubenswrapper[4789]: I1122 08:16:37.436706 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f1fc498-45f3-4bec-a4a5-b9722d2445ea" containerName="keystone-bootstrap" Nov 22 08:16:37 crc kubenswrapper[4789]: E1122 08:16:37.436720 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cf0fb02-e586-4daa-88c1-6556b376c02a" containerName="dnsmasq-dns" Nov 22 08:16:37 crc kubenswrapper[4789]: I1122 08:16:37.436726 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cf0fb02-e586-4daa-88c1-6556b376c02a" containerName="dnsmasq-dns" Nov 22 08:16:37 crc kubenswrapper[4789]: E1122 08:16:37.436739 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2eb187f8-1892-4d72-b442-43592e75d489" containerName="placement-db-sync" Nov 22 08:16:37 crc kubenswrapper[4789]: I1122 08:16:37.436758 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="2eb187f8-1892-4d72-b442-43592e75d489" containerName="placement-db-sync" Nov 22 08:16:37 crc kubenswrapper[4789]: I1122 08:16:37.436915 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="2eb187f8-1892-4d72-b442-43592e75d489" containerName="placement-db-sync" Nov 22 08:16:37 crc kubenswrapper[4789]: I1122 08:16:37.436928 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f1fc498-45f3-4bec-a4a5-b9722d2445ea" containerName="keystone-bootstrap" Nov 22 08:16:37 crc kubenswrapper[4789]: I1122 08:16:37.436944 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="0cf0fb02-e586-4daa-88c1-6556b376c02a" containerName="dnsmasq-dns" Nov 22 08:16:37 crc kubenswrapper[4789]: I1122 08:16:37.437786 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-7846c8b766-kj5xm" Nov 22 08:16:37 crc kubenswrapper[4789]: I1122 08:16:37.440322 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Nov 22 08:16:37 crc kubenswrapper[4789]: I1122 08:16:37.440893 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Nov 22 08:16:37 crc kubenswrapper[4789]: I1122 08:16:37.441074 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Nov 22 08:16:37 crc kubenswrapper[4789]: I1122 08:16:37.443641 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-vrlbv" Nov 22 08:16:37 crc kubenswrapper[4789]: I1122 08:16:37.449803 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Nov 22 08:16:37 crc kubenswrapper[4789]: I1122 08:16:37.461055 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-7846c8b766-kj5xm"] Nov 22 08:16:37 crc kubenswrapper[4789]: I1122 08:16:37.540167 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-5c64d54575-bxqxv"] Nov 22 08:16:37 crc kubenswrapper[4789]: I1122 08:16:37.541689 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5c64d54575-bxqxv" Nov 22 08:16:37 crc kubenswrapper[4789]: I1122 08:16:37.545739 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Nov 22 08:16:37 crc kubenswrapper[4789]: I1122 08:16:37.546045 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-wnnts" Nov 22 08:16:37 crc kubenswrapper[4789]: I1122 08:16:37.546074 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Nov 22 08:16:37 crc kubenswrapper[4789]: I1122 08:16:37.546079 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 22 08:16:37 crc kubenswrapper[4789]: I1122 08:16:37.546197 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 22 08:16:37 crc kubenswrapper[4789]: I1122 08:16:37.550307 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 22 08:16:37 crc kubenswrapper[4789]: I1122 08:16:37.563052 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-5c64d54575-bxqxv"] Nov 22 08:16:37 crc kubenswrapper[4789]: I1122 08:16:37.575639 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87e6a5f6-fbad-4cdf-bc22-1b6be733812a-combined-ca-bundle\") pod \"placement-7846c8b766-kj5xm\" (UID: \"87e6a5f6-fbad-4cdf-bc22-1b6be733812a\") " pod="openstack/placement-7846c8b766-kj5xm" Nov 22 08:16:37 crc kubenswrapper[4789]: I1122 08:16:37.575736 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/87e6a5f6-fbad-4cdf-bc22-1b6be733812a-scripts\") pod \"placement-7846c8b766-kj5xm\" (UID: \"87e6a5f6-fbad-4cdf-bc22-1b6be733812a\") " pod="openstack/placement-7846c8b766-kj5xm" Nov 22 08:16:37 crc kubenswrapper[4789]: I1122 08:16:37.575892 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/87e6a5f6-fbad-4cdf-bc22-1b6be733812a-internal-tls-certs\") pod \"placement-7846c8b766-kj5xm\" (UID: \"87e6a5f6-fbad-4cdf-bc22-1b6be733812a\") " pod="openstack/placement-7846c8b766-kj5xm" Nov 22 08:16:37 crc kubenswrapper[4789]: I1122 08:16:37.576691 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/87e6a5f6-fbad-4cdf-bc22-1b6be733812a-public-tls-certs\") pod \"placement-7846c8b766-kj5xm\" (UID: \"87e6a5f6-fbad-4cdf-bc22-1b6be733812a\") " pod="openstack/placement-7846c8b766-kj5xm" Nov 22 08:16:37 crc kubenswrapper[4789]: I1122 08:16:37.576870 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/87e6a5f6-fbad-4cdf-bc22-1b6be733812a-logs\") pod \"placement-7846c8b766-kj5xm\" (UID: \"87e6a5f6-fbad-4cdf-bc22-1b6be733812a\") " pod="openstack/placement-7846c8b766-kj5xm" Nov 22 08:16:37 crc kubenswrapper[4789]: I1122 08:16:37.577204 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mh4z2\" (UniqueName: \"kubernetes.io/projected/87e6a5f6-fbad-4cdf-bc22-1b6be733812a-kube-api-access-mh4z2\") pod \"placement-7846c8b766-kj5xm\" (UID: \"87e6a5f6-fbad-4cdf-bc22-1b6be733812a\") " pod="openstack/placement-7846c8b766-kj5xm" Nov 22 08:16:37 crc kubenswrapper[4789]: I1122 08:16:37.577354 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87e6a5f6-fbad-4cdf-bc22-1b6be733812a-config-data\") pod \"placement-7846c8b766-kj5xm\" (UID: \"87e6a5f6-fbad-4cdf-bc22-1b6be733812a\") " pod="openstack/placement-7846c8b766-kj5xm" Nov 22 08:16:37 crc kubenswrapper[4789]: I1122 08:16:37.678448 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/87e6a5f6-fbad-4cdf-bc22-1b6be733812a-logs\") pod \"placement-7846c8b766-kj5xm\" (UID: \"87e6a5f6-fbad-4cdf-bc22-1b6be733812a\") " pod="openstack/placement-7846c8b766-kj5xm" Nov 22 08:16:37 crc kubenswrapper[4789]: I1122 08:16:37.678522 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60140344-7346-4115-adcb-45e3f7f366a0-combined-ca-bundle\") pod \"keystone-5c64d54575-bxqxv\" (UID: \"60140344-7346-4115-adcb-45e3f7f366a0\") " pod="openstack/keystone-5c64d54575-bxqxv" Nov 22 08:16:37 crc kubenswrapper[4789]: I1122 08:16:37.678557 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/60140344-7346-4115-adcb-45e3f7f366a0-fernet-keys\") pod \"keystone-5c64d54575-bxqxv\" (UID: \"60140344-7346-4115-adcb-45e3f7f366a0\") " pod="openstack/keystone-5c64d54575-bxqxv" Nov 22 08:16:37 crc kubenswrapper[4789]: I1122 08:16:37.678606 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/60140344-7346-4115-adcb-45e3f7f366a0-scripts\") pod \"keystone-5c64d54575-bxqxv\" (UID: \"60140344-7346-4115-adcb-45e3f7f366a0\") " pod="openstack/keystone-5c64d54575-bxqxv" Nov 22 08:16:37 crc kubenswrapper[4789]: I1122 08:16:37.678658 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mh4z2\" (UniqueName: \"kubernetes.io/projected/87e6a5f6-fbad-4cdf-bc22-1b6be733812a-kube-api-access-mh4z2\") pod \"placement-7846c8b766-kj5xm\" (UID: \"87e6a5f6-fbad-4cdf-bc22-1b6be733812a\") " pod="openstack/placement-7846c8b766-kj5xm" Nov 22 08:16:37 crc kubenswrapper[4789]: I1122 08:16:37.678694 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/60140344-7346-4115-adcb-45e3f7f366a0-public-tls-certs\") pod \"keystone-5c64d54575-bxqxv\" (UID: \"60140344-7346-4115-adcb-45e3f7f366a0\") " pod="openstack/keystone-5c64d54575-bxqxv" Nov 22 08:16:37 crc kubenswrapper[4789]: I1122 08:16:37.678721 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87e6a5f6-fbad-4cdf-bc22-1b6be733812a-config-data\") pod \"placement-7846c8b766-kj5xm\" (UID: \"87e6a5f6-fbad-4cdf-bc22-1b6be733812a\") " pod="openstack/placement-7846c8b766-kj5xm" Nov 22 08:16:37 crc kubenswrapper[4789]: I1122 08:16:37.678788 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/60140344-7346-4115-adcb-45e3f7f366a0-internal-tls-certs\") pod \"keystone-5c64d54575-bxqxv\" (UID: \"60140344-7346-4115-adcb-45e3f7f366a0\") " pod="openstack/keystone-5c64d54575-bxqxv" Nov 22 08:16:37 crc kubenswrapper[4789]: I1122 08:16:37.678814 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60140344-7346-4115-adcb-45e3f7f366a0-config-data\") pod \"keystone-5c64d54575-bxqxv\" (UID: \"60140344-7346-4115-adcb-45e3f7f366a0\") " pod="openstack/keystone-5c64d54575-bxqxv" Nov 22 08:16:37 crc kubenswrapper[4789]: I1122 08:16:37.678847 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87e6a5f6-fbad-4cdf-bc22-1b6be733812a-combined-ca-bundle\") pod \"placement-7846c8b766-kj5xm\" (UID: \"87e6a5f6-fbad-4cdf-bc22-1b6be733812a\") " pod="openstack/placement-7846c8b766-kj5xm" Nov 22 08:16:37 crc kubenswrapper[4789]: I1122 08:16:37.678875 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/87e6a5f6-fbad-4cdf-bc22-1b6be733812a-scripts\") pod \"placement-7846c8b766-kj5xm\" (UID: \"87e6a5f6-fbad-4cdf-bc22-1b6be733812a\") " pod="openstack/placement-7846c8b766-kj5xm" Nov 22 08:16:37 crc kubenswrapper[4789]: I1122 08:16:37.678908 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9sdlv\" (UniqueName: \"kubernetes.io/projected/60140344-7346-4115-adcb-45e3f7f366a0-kube-api-access-9sdlv\") pod \"keystone-5c64d54575-bxqxv\" (UID: \"60140344-7346-4115-adcb-45e3f7f366a0\") " pod="openstack/keystone-5c64d54575-bxqxv" Nov 22 08:16:37 crc kubenswrapper[4789]: I1122 08:16:37.678933 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/87e6a5f6-fbad-4cdf-bc22-1b6be733812a-internal-tls-certs\") pod \"placement-7846c8b766-kj5xm\" (UID: \"87e6a5f6-fbad-4cdf-bc22-1b6be733812a\") " pod="openstack/placement-7846c8b766-kj5xm" Nov 22 08:16:37 crc kubenswrapper[4789]: I1122 08:16:37.678958 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/87e6a5f6-fbad-4cdf-bc22-1b6be733812a-public-tls-certs\") pod \"placement-7846c8b766-kj5xm\" (UID: \"87e6a5f6-fbad-4cdf-bc22-1b6be733812a\") " pod="openstack/placement-7846c8b766-kj5xm" Nov 22 08:16:37 crc kubenswrapper[4789]: I1122 08:16:37.678996 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/60140344-7346-4115-adcb-45e3f7f366a0-credential-keys\") pod \"keystone-5c64d54575-bxqxv\" (UID: \"60140344-7346-4115-adcb-45e3f7f366a0\") " pod="openstack/keystone-5c64d54575-bxqxv" Nov 22 08:16:37 crc kubenswrapper[4789]: I1122 08:16:37.679464 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/87e6a5f6-fbad-4cdf-bc22-1b6be733812a-logs\") pod \"placement-7846c8b766-kj5xm\" (UID: \"87e6a5f6-fbad-4cdf-bc22-1b6be733812a\") " pod="openstack/placement-7846c8b766-kj5xm" Nov 22 08:16:37 crc kubenswrapper[4789]: I1122 08:16:37.685006 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/87e6a5f6-fbad-4cdf-bc22-1b6be733812a-scripts\") pod \"placement-7846c8b766-kj5xm\" (UID: \"87e6a5f6-fbad-4cdf-bc22-1b6be733812a\") " pod="openstack/placement-7846c8b766-kj5xm" Nov 22 08:16:37 crc kubenswrapper[4789]: I1122 08:16:37.685635 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87e6a5f6-fbad-4cdf-bc22-1b6be733812a-combined-ca-bundle\") pod \"placement-7846c8b766-kj5xm\" (UID: \"87e6a5f6-fbad-4cdf-bc22-1b6be733812a\") " pod="openstack/placement-7846c8b766-kj5xm" Nov 22 08:16:37 crc kubenswrapper[4789]: I1122 08:16:37.687032 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/87e6a5f6-fbad-4cdf-bc22-1b6be733812a-internal-tls-certs\") pod \"placement-7846c8b766-kj5xm\" (UID: \"87e6a5f6-fbad-4cdf-bc22-1b6be733812a\") " pod="openstack/placement-7846c8b766-kj5xm" Nov 22 08:16:37 crc kubenswrapper[4789]: I1122 08:16:37.687587 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87e6a5f6-fbad-4cdf-bc22-1b6be733812a-config-data\") pod \"placement-7846c8b766-kj5xm\" (UID: \"87e6a5f6-fbad-4cdf-bc22-1b6be733812a\") " pod="openstack/placement-7846c8b766-kj5xm" Nov 22 08:16:37 crc kubenswrapper[4789]: I1122 08:16:37.687627 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/87e6a5f6-fbad-4cdf-bc22-1b6be733812a-public-tls-certs\") pod \"placement-7846c8b766-kj5xm\" (UID: \"87e6a5f6-fbad-4cdf-bc22-1b6be733812a\") " pod="openstack/placement-7846c8b766-kj5xm" Nov 22 08:16:37 crc kubenswrapper[4789]: I1122 08:16:37.699792 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mh4z2\" (UniqueName: \"kubernetes.io/projected/87e6a5f6-fbad-4cdf-bc22-1b6be733812a-kube-api-access-mh4z2\") pod \"placement-7846c8b766-kj5xm\" (UID: \"87e6a5f6-fbad-4cdf-bc22-1b6be733812a\") " pod="openstack/placement-7846c8b766-kj5xm" Nov 22 08:16:37 crc kubenswrapper[4789]: I1122 08:16:37.752195 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-7846c8b766-kj5xm" Nov 22 08:16:37 crc kubenswrapper[4789]: I1122 08:16:37.780290 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/60140344-7346-4115-adcb-45e3f7f366a0-internal-tls-certs\") pod \"keystone-5c64d54575-bxqxv\" (UID: \"60140344-7346-4115-adcb-45e3f7f366a0\") " pod="openstack/keystone-5c64d54575-bxqxv" Nov 22 08:16:37 crc kubenswrapper[4789]: I1122 08:16:37.780602 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60140344-7346-4115-adcb-45e3f7f366a0-config-data\") pod \"keystone-5c64d54575-bxqxv\" (UID: \"60140344-7346-4115-adcb-45e3f7f366a0\") " pod="openstack/keystone-5c64d54575-bxqxv" Nov 22 08:16:37 crc kubenswrapper[4789]: I1122 08:16:37.780743 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9sdlv\" (UniqueName: \"kubernetes.io/projected/60140344-7346-4115-adcb-45e3f7f366a0-kube-api-access-9sdlv\") pod \"keystone-5c64d54575-bxqxv\" (UID: \"60140344-7346-4115-adcb-45e3f7f366a0\") " pod="openstack/keystone-5c64d54575-bxqxv" Nov 22 08:16:37 crc kubenswrapper[4789]: I1122 08:16:37.780900 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/60140344-7346-4115-adcb-45e3f7f366a0-credential-keys\") pod \"keystone-5c64d54575-bxqxv\" (UID: \"60140344-7346-4115-adcb-45e3f7f366a0\") " pod="openstack/keystone-5c64d54575-bxqxv" Nov 22 08:16:37 crc kubenswrapper[4789]: I1122 08:16:37.781058 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60140344-7346-4115-adcb-45e3f7f366a0-combined-ca-bundle\") pod \"keystone-5c64d54575-bxqxv\" (UID: \"60140344-7346-4115-adcb-45e3f7f366a0\") " pod="openstack/keystone-5c64d54575-bxqxv" Nov 22 08:16:37 crc kubenswrapper[4789]: I1122 08:16:37.781896 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/60140344-7346-4115-adcb-45e3f7f366a0-fernet-keys\") pod \"keystone-5c64d54575-bxqxv\" (UID: \"60140344-7346-4115-adcb-45e3f7f366a0\") " pod="openstack/keystone-5c64d54575-bxqxv" Nov 22 08:16:37 crc kubenswrapper[4789]: I1122 08:16:37.782030 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/60140344-7346-4115-adcb-45e3f7f366a0-scripts\") pod \"keystone-5c64d54575-bxqxv\" (UID: \"60140344-7346-4115-adcb-45e3f7f366a0\") " pod="openstack/keystone-5c64d54575-bxqxv" Nov 22 08:16:37 crc kubenswrapper[4789]: I1122 08:16:37.782197 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/60140344-7346-4115-adcb-45e3f7f366a0-public-tls-certs\") pod \"keystone-5c64d54575-bxqxv\" (UID: \"60140344-7346-4115-adcb-45e3f7f366a0\") " pod="openstack/keystone-5c64d54575-bxqxv" Nov 22 08:16:37 crc kubenswrapper[4789]: I1122 08:16:37.785647 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/60140344-7346-4115-adcb-45e3f7f366a0-internal-tls-certs\") pod \"keystone-5c64d54575-bxqxv\" (UID: \"60140344-7346-4115-adcb-45e3f7f366a0\") " pod="openstack/keystone-5c64d54575-bxqxv" Nov 22 08:16:37 crc kubenswrapper[4789]: I1122 08:16:37.795858 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60140344-7346-4115-adcb-45e3f7f366a0-config-data\") pod \"keystone-5c64d54575-bxqxv\" (UID: \"60140344-7346-4115-adcb-45e3f7f366a0\") " pod="openstack/keystone-5c64d54575-bxqxv" Nov 22 08:16:37 crc kubenswrapper[4789]: I1122 08:16:37.800375 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60140344-7346-4115-adcb-45e3f7f366a0-combined-ca-bundle\") pod \"keystone-5c64d54575-bxqxv\" (UID: \"60140344-7346-4115-adcb-45e3f7f366a0\") " pod="openstack/keystone-5c64d54575-bxqxv" Nov 22 08:16:37 crc kubenswrapper[4789]: I1122 08:16:37.801420 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/60140344-7346-4115-adcb-45e3f7f366a0-credential-keys\") pod \"keystone-5c64d54575-bxqxv\" (UID: \"60140344-7346-4115-adcb-45e3f7f366a0\") " pod="openstack/keystone-5c64d54575-bxqxv" Nov 22 08:16:37 crc kubenswrapper[4789]: I1122 08:16:37.806092 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/60140344-7346-4115-adcb-45e3f7f366a0-scripts\") pod \"keystone-5c64d54575-bxqxv\" (UID: \"60140344-7346-4115-adcb-45e3f7f366a0\") " pod="openstack/keystone-5c64d54575-bxqxv" Nov 22 08:16:37 crc kubenswrapper[4789]: I1122 08:16:37.807276 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9sdlv\" (UniqueName: \"kubernetes.io/projected/60140344-7346-4115-adcb-45e3f7f366a0-kube-api-access-9sdlv\") pod \"keystone-5c64d54575-bxqxv\" (UID: \"60140344-7346-4115-adcb-45e3f7f366a0\") " pod="openstack/keystone-5c64d54575-bxqxv" Nov 22 08:16:37 crc kubenswrapper[4789]: I1122 08:16:37.817187 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/60140344-7346-4115-adcb-45e3f7f366a0-public-tls-certs\") pod \"keystone-5c64d54575-bxqxv\" (UID: \"60140344-7346-4115-adcb-45e3f7f366a0\") " pod="openstack/keystone-5c64d54575-bxqxv" Nov 22 08:16:37 crc kubenswrapper[4789]: I1122 08:16:37.821344 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/60140344-7346-4115-adcb-45e3f7f366a0-fernet-keys\") pod \"keystone-5c64d54575-bxqxv\" (UID: \"60140344-7346-4115-adcb-45e3f7f366a0\") " pod="openstack/keystone-5c64d54575-bxqxv" Nov 22 08:16:37 crc kubenswrapper[4789]: I1122 08:16:37.865277 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5c64d54575-bxqxv" Nov 22 08:16:38 crc kubenswrapper[4789]: I1122 08:16:38.296160 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-5c64d54575-bxqxv"] Nov 22 08:16:38 crc kubenswrapper[4789]: W1122 08:16:38.311627 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod60140344_7346_4115_adcb_45e3f7f366a0.slice/crio-de43cd07e50795d1a54217529940ea23fe9bd625f1ed1796db942bc7ae4c75bb WatchSource:0}: Error finding container de43cd07e50795d1a54217529940ea23fe9bd625f1ed1796db942bc7ae4c75bb: Status 404 returned error can't find the container with id de43cd07e50795d1a54217529940ea23fe9bd625f1ed1796db942bc7ae4c75bb Nov 22 08:16:38 crc kubenswrapper[4789]: I1122 08:16:38.325020 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5c64d54575-bxqxv" event={"ID":"60140344-7346-4115-adcb-45e3f7f366a0","Type":"ContainerStarted","Data":"de43cd07e50795d1a54217529940ea23fe9bd625f1ed1796db942bc7ae4c75bb"} Nov 22 08:16:38 crc kubenswrapper[4789]: I1122 08:16:38.326283 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-7846c8b766-kj5xm"] Nov 22 08:16:38 crc kubenswrapper[4789]: I1122 08:16:38.454315 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-5b56df9788-wsmhq" podUID="56fd3074-a15c-4024-baf9-f784a50e14a5" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.140:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.140:8443: connect: connection refused" Nov 22 08:16:38 crc kubenswrapper[4789]: I1122 08:16:38.584271 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7b778f8c8-6rl4f" podUID="d9fb66f7-ee5b-453a-a69b-56fb9337d898" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.141:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.141:8443: connect: connection refused" Nov 22 08:16:39 crc kubenswrapper[4789]: I1122 08:16:39.336551 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7846c8b766-kj5xm" event={"ID":"87e6a5f6-fbad-4cdf-bc22-1b6be733812a","Type":"ContainerStarted","Data":"ffca3ee2f2be2616f3c46775b0eae2ec6dfc3013e1f64361e0472bd0eb62d742"} Nov 22 08:16:39 crc kubenswrapper[4789]: I1122 08:16:39.336939 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-7846c8b766-kj5xm" Nov 22 08:16:39 crc kubenswrapper[4789]: I1122 08:16:39.336954 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7846c8b766-kj5xm" event={"ID":"87e6a5f6-fbad-4cdf-bc22-1b6be733812a","Type":"ContainerStarted","Data":"6b07251bdab8a6af09f675fd67e107c55b0be7f7618e959f7f937482a434db9d"} Nov 22 08:16:39 crc kubenswrapper[4789]: I1122 08:16:39.336968 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7846c8b766-kj5xm" event={"ID":"87e6a5f6-fbad-4cdf-bc22-1b6be733812a","Type":"ContainerStarted","Data":"6eb9a0b446bfb6ff5de7cd0a9c03bf1e56054815e8740c1eea17063b7eb88710"} Nov 22 08:16:39 crc kubenswrapper[4789]: I1122 08:16:39.336991 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-7846c8b766-kj5xm" Nov 22 08:16:39 crc kubenswrapper[4789]: I1122 08:16:39.338586 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5c64d54575-bxqxv" event={"ID":"60140344-7346-4115-adcb-45e3f7f366a0","Type":"ContainerStarted","Data":"65ac28c8450179102a14e94cb19dca5c5d9fe5f67c8c516519c1c275d04d91b9"} Nov 22 08:16:39 crc kubenswrapper[4789]: I1122 08:16:39.338736 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-5c64d54575-bxqxv" Nov 22 08:16:39 crc kubenswrapper[4789]: I1122 08:16:39.357141 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-7846c8b766-kj5xm" podStartSLOduration=2.357121948 podStartE2EDuration="2.357121948s" podCreationTimestamp="2025-11-22 08:16:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:16:39.353993143 +0000 UTC m=+1213.588393416" watchObservedRunningTime="2025-11-22 08:16:39.357121948 +0000 UTC m=+1213.591522221" Nov 22 08:16:39 crc kubenswrapper[4789]: I1122 08:16:39.391182 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-5c64d54575-bxqxv" podStartSLOduration=2.391162386 podStartE2EDuration="2.391162386s" podCreationTimestamp="2025-11-22 08:16:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:16:39.383351613 +0000 UTC m=+1213.617751876" watchObservedRunningTime="2025-11-22 08:16:39.391162386 +0000 UTC m=+1213.625562659" Nov 22 08:16:42 crc kubenswrapper[4789]: I1122 08:16:42.390851 4789 generic.go:334] "Generic (PLEG): container finished" podID="733e94b9-65c2-406e-9ebe-244585e4e15d" containerID="aa73593836103e7d3c745c9dd3a425820bddd9b5607f5bdc6e6dfaa4e6e0506f" exitCode=0 Nov 22 08:16:42 crc kubenswrapper[4789]: I1122 08:16:42.391381 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-n4s7g" event={"ID":"733e94b9-65c2-406e-9ebe-244585e4e15d","Type":"ContainerDied","Data":"aa73593836103e7d3c745c9dd3a425820bddd9b5607f5bdc6e6dfaa4e6e0506f"} Nov 22 08:16:44 crc kubenswrapper[4789]: I1122 08:16:44.412273 4789 generic.go:334] "Generic (PLEG): container finished" podID="f3ee5cf5-1275-448f-8713-e2b5a0d24439" containerID="43622e76a3aa6a505c8df9823f7041e41a83be4429150db9572330e047631f10" exitCode=0 Nov 22 08:16:44 crc kubenswrapper[4789]: I1122 08:16:44.412410 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-r6nst" event={"ID":"f3ee5cf5-1275-448f-8713-e2b5a0d24439","Type":"ContainerDied","Data":"43622e76a3aa6a505c8df9823f7041e41a83be4429150db9572330e047631f10"} Nov 22 08:16:46 crc kubenswrapper[4789]: I1122 08:16:46.125131 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-r6nst" Nov 22 08:16:46 crc kubenswrapper[4789]: I1122 08:16:46.237611 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-n4s7g" Nov 22 08:16:46 crc kubenswrapper[4789]: I1122 08:16:46.317413 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3ee5cf5-1275-448f-8713-e2b5a0d24439-combined-ca-bundle\") pod \"f3ee5cf5-1275-448f-8713-e2b5a0d24439\" (UID: \"f3ee5cf5-1275-448f-8713-e2b5a0d24439\") " Nov 22 08:16:46 crc kubenswrapper[4789]: I1122 08:16:46.317660 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f3ee5cf5-1275-448f-8713-e2b5a0d24439-db-sync-config-data\") pod \"f3ee5cf5-1275-448f-8713-e2b5a0d24439\" (UID: \"f3ee5cf5-1275-448f-8713-e2b5a0d24439\") " Nov 22 08:16:46 crc kubenswrapper[4789]: I1122 08:16:46.317728 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9slr4\" (UniqueName: \"kubernetes.io/projected/f3ee5cf5-1275-448f-8713-e2b5a0d24439-kube-api-access-9slr4\") pod \"f3ee5cf5-1275-448f-8713-e2b5a0d24439\" (UID: \"f3ee5cf5-1275-448f-8713-e2b5a0d24439\") " Nov 22 08:16:46 crc kubenswrapper[4789]: I1122 08:16:46.330481 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f3ee5cf5-1275-448f-8713-e2b5a0d24439-kube-api-access-9slr4" (OuterVolumeSpecName: "kube-api-access-9slr4") pod "f3ee5cf5-1275-448f-8713-e2b5a0d24439" (UID: "f3ee5cf5-1275-448f-8713-e2b5a0d24439"). InnerVolumeSpecName "kube-api-access-9slr4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:16:46 crc kubenswrapper[4789]: I1122 08:16:46.343873 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3ee5cf5-1275-448f-8713-e2b5a0d24439-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "f3ee5cf5-1275-448f-8713-e2b5a0d24439" (UID: "f3ee5cf5-1275-448f-8713-e2b5a0d24439"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:16:46 crc kubenswrapper[4789]: I1122 08:16:46.381911 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3ee5cf5-1275-448f-8713-e2b5a0d24439-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f3ee5cf5-1275-448f-8713-e2b5a0d24439" (UID: "f3ee5cf5-1275-448f-8713-e2b5a0d24439"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:16:46 crc kubenswrapper[4789]: I1122 08:16:46.419419 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/733e94b9-65c2-406e-9ebe-244585e4e15d-combined-ca-bundle\") pod \"733e94b9-65c2-406e-9ebe-244585e4e15d\" (UID: \"733e94b9-65c2-406e-9ebe-244585e4e15d\") " Nov 22 08:16:46 crc kubenswrapper[4789]: I1122 08:16:46.419488 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/733e94b9-65c2-406e-9ebe-244585e4e15d-config\") pod \"733e94b9-65c2-406e-9ebe-244585e4e15d\" (UID: \"733e94b9-65c2-406e-9ebe-244585e4e15d\") " Nov 22 08:16:46 crc kubenswrapper[4789]: I1122 08:16:46.419639 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9mw6d\" (UniqueName: \"kubernetes.io/projected/733e94b9-65c2-406e-9ebe-244585e4e15d-kube-api-access-9mw6d\") pod \"733e94b9-65c2-406e-9ebe-244585e4e15d\" (UID: \"733e94b9-65c2-406e-9ebe-244585e4e15d\") " Nov 22 08:16:46 crc kubenswrapper[4789]: I1122 08:16:46.422433 4789 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f3ee5cf5-1275-448f-8713-e2b5a0d24439-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:16:46 crc kubenswrapper[4789]: I1122 08:16:46.422830 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9slr4\" (UniqueName: \"kubernetes.io/projected/f3ee5cf5-1275-448f-8713-e2b5a0d24439-kube-api-access-9slr4\") on node \"crc\" DevicePath \"\"" Nov 22 08:16:46 crc kubenswrapper[4789]: I1122 08:16:46.422906 4789 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3ee5cf5-1275-448f-8713-e2b5a0d24439-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:16:46 crc kubenswrapper[4789]: I1122 08:16:46.433920 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/733e94b9-65c2-406e-9ebe-244585e4e15d-kube-api-access-9mw6d" (OuterVolumeSpecName: "kube-api-access-9mw6d") pod "733e94b9-65c2-406e-9ebe-244585e4e15d" (UID: "733e94b9-65c2-406e-9ebe-244585e4e15d"). InnerVolumeSpecName "kube-api-access-9mw6d". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:16:46 crc kubenswrapper[4789]: I1122 08:16:46.462634 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/733e94b9-65c2-406e-9ebe-244585e4e15d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "733e94b9-65c2-406e-9ebe-244585e4e15d" (UID: "733e94b9-65c2-406e-9ebe-244585e4e15d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:16:46 crc kubenswrapper[4789]: I1122 08:16:46.468648 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-r6nst" event={"ID":"f3ee5cf5-1275-448f-8713-e2b5a0d24439","Type":"ContainerDied","Data":"09be797b1e863f926c538f47c129fcdbb99348630904565dc9c8607260421481"} Nov 22 08:16:46 crc kubenswrapper[4789]: I1122 08:16:46.468891 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-r6nst" Nov 22 08:16:46 crc kubenswrapper[4789]: I1122 08:16:46.468696 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="09be797b1e863f926c538f47c129fcdbb99348630904565dc9c8607260421481" Nov 22 08:16:46 crc kubenswrapper[4789]: I1122 08:16:46.471385 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-n4s7g" event={"ID":"733e94b9-65c2-406e-9ebe-244585e4e15d","Type":"ContainerDied","Data":"af51430767cb5d1d2c0d8f8f42df0fa11f0930283020526b5643ab6192006ea8"} Nov 22 08:16:46 crc kubenswrapper[4789]: I1122 08:16:46.471516 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-n4s7g" Nov 22 08:16:46 crc kubenswrapper[4789]: I1122 08:16:46.472012 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="af51430767cb5d1d2c0d8f8f42df0fa11f0930283020526b5643ab6192006ea8" Nov 22 08:16:46 crc kubenswrapper[4789]: I1122 08:16:46.480690 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/733e94b9-65c2-406e-9ebe-244585e4e15d-config" (OuterVolumeSpecName: "config") pod "733e94b9-65c2-406e-9ebe-244585e4e15d" (UID: "733e94b9-65c2-406e-9ebe-244585e4e15d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:16:46 crc kubenswrapper[4789]: I1122 08:16:46.524732 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9mw6d\" (UniqueName: \"kubernetes.io/projected/733e94b9-65c2-406e-9ebe-244585e4e15d-kube-api-access-9mw6d\") on node \"crc\" DevicePath \"\"" Nov 22 08:16:46 crc kubenswrapper[4789]: I1122 08:16:46.524785 4789 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/733e94b9-65c2-406e-9ebe-244585e4e15d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:16:46 crc kubenswrapper[4789]: I1122 08:16:46.524801 4789 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/733e94b9-65c2-406e-9ebe-244585e4e15d-config\") on node \"crc\" DevicePath \"\"" Nov 22 08:16:46 crc kubenswrapper[4789]: I1122 08:16:46.693354 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-75d6db87b5-kbchw"] Nov 22 08:16:46 crc kubenswrapper[4789]: E1122 08:16:46.693818 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3ee5cf5-1275-448f-8713-e2b5a0d24439" containerName="barbican-db-sync" Nov 22 08:16:46 crc kubenswrapper[4789]: I1122 08:16:46.693868 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3ee5cf5-1275-448f-8713-e2b5a0d24439" containerName="barbican-db-sync" Nov 22 08:16:46 crc kubenswrapper[4789]: E1122 08:16:46.693890 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="733e94b9-65c2-406e-9ebe-244585e4e15d" containerName="neutron-db-sync" Nov 22 08:16:46 crc kubenswrapper[4789]: I1122 08:16:46.693899 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="733e94b9-65c2-406e-9ebe-244585e4e15d" containerName="neutron-db-sync" Nov 22 08:16:46 crc kubenswrapper[4789]: I1122 08:16:46.694095 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3ee5cf5-1275-448f-8713-e2b5a0d24439" containerName="barbican-db-sync" Nov 22 08:16:46 crc kubenswrapper[4789]: I1122 08:16:46.694123 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="733e94b9-65c2-406e-9ebe-244585e4e15d" containerName="neutron-db-sync" Nov 22 08:16:46 crc kubenswrapper[4789]: I1122 08:16:46.695276 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-75d6db87b5-kbchw" Nov 22 08:16:46 crc kubenswrapper[4789]: I1122 08:16:46.708186 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-hc559" Nov 22 08:16:46 crc kubenswrapper[4789]: I1122 08:16:46.708482 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Nov 22 08:16:46 crc kubenswrapper[4789]: I1122 08:16:46.708485 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Nov 22 08:16:46 crc kubenswrapper[4789]: I1122 08:16:46.716818 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-75d6db87b5-kbchw"] Nov 22 08:16:46 crc kubenswrapper[4789]: I1122 08:16:46.730333 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-7ccbbd9f98-jd5vz"] Nov 22 08:16:46 crc kubenswrapper[4789]: I1122 08:16:46.736039 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-7ccbbd9f98-jd5vz" Nov 22 08:16:46 crc kubenswrapper[4789]: I1122 08:16:46.737619 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Nov 22 08:16:46 crc kubenswrapper[4789]: I1122 08:16:46.764159 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-7ccbbd9f98-jd5vz"] Nov 22 08:16:46 crc kubenswrapper[4789]: I1122 08:16:46.806201 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-68cd7949f9-tqtjq"] Nov 22 08:16:46 crc kubenswrapper[4789]: I1122 08:16:46.823905 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-68cd7949f9-tqtjq"] Nov 22 08:16:46 crc kubenswrapper[4789]: I1122 08:16:46.824004 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-68cd7949f9-tqtjq" Nov 22 08:16:46 crc kubenswrapper[4789]: I1122 08:16:46.828404 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zr9fz\" (UniqueName: \"kubernetes.io/projected/affd3f11-c0d4-41a2-888b-f4ae0c7e3d41-kube-api-access-zr9fz\") pod \"barbican-keystone-listener-7ccbbd9f98-jd5vz\" (UID: \"affd3f11-c0d4-41a2-888b-f4ae0c7e3d41\") " pod="openstack/barbican-keystone-listener-7ccbbd9f98-jd5vz" Nov 22 08:16:46 crc kubenswrapper[4789]: I1122 08:16:46.828487 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4be1163c-8900-4db9-a041-826526ae8751-config-data\") pod \"barbican-worker-75d6db87b5-kbchw\" (UID: \"4be1163c-8900-4db9-a041-826526ae8751\") " pod="openstack/barbican-worker-75d6db87b5-kbchw" Nov 22 08:16:46 crc kubenswrapper[4789]: I1122 08:16:46.828519 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rmth9\" (UniqueName: \"kubernetes.io/projected/4be1163c-8900-4db9-a041-826526ae8751-kube-api-access-rmth9\") pod \"barbican-worker-75d6db87b5-kbchw\" (UID: \"4be1163c-8900-4db9-a041-826526ae8751\") " pod="openstack/barbican-worker-75d6db87b5-kbchw" Nov 22 08:16:46 crc kubenswrapper[4789]: I1122 08:16:46.828539 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4be1163c-8900-4db9-a041-826526ae8751-combined-ca-bundle\") pod \"barbican-worker-75d6db87b5-kbchw\" (UID: \"4be1163c-8900-4db9-a041-826526ae8751\") " pod="openstack/barbican-worker-75d6db87b5-kbchw" Nov 22 08:16:46 crc kubenswrapper[4789]: I1122 08:16:46.828601 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/affd3f11-c0d4-41a2-888b-f4ae0c7e3d41-combined-ca-bundle\") pod \"barbican-keystone-listener-7ccbbd9f98-jd5vz\" (UID: \"affd3f11-c0d4-41a2-888b-f4ae0c7e3d41\") " pod="openstack/barbican-keystone-listener-7ccbbd9f98-jd5vz" Nov 22 08:16:46 crc kubenswrapper[4789]: I1122 08:16:46.828629 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/affd3f11-c0d4-41a2-888b-f4ae0c7e3d41-logs\") pod \"barbican-keystone-listener-7ccbbd9f98-jd5vz\" (UID: \"affd3f11-c0d4-41a2-888b-f4ae0c7e3d41\") " pod="openstack/barbican-keystone-listener-7ccbbd9f98-jd5vz" Nov 22 08:16:46 crc kubenswrapper[4789]: I1122 08:16:46.828656 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/affd3f11-c0d4-41a2-888b-f4ae0c7e3d41-config-data\") pod \"barbican-keystone-listener-7ccbbd9f98-jd5vz\" (UID: \"affd3f11-c0d4-41a2-888b-f4ae0c7e3d41\") " pod="openstack/barbican-keystone-listener-7ccbbd9f98-jd5vz" Nov 22 08:16:46 crc kubenswrapper[4789]: I1122 08:16:46.828683 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4be1163c-8900-4db9-a041-826526ae8751-config-data-custom\") pod \"barbican-worker-75d6db87b5-kbchw\" (UID: \"4be1163c-8900-4db9-a041-826526ae8751\") " pod="openstack/barbican-worker-75d6db87b5-kbchw" Nov 22 08:16:46 crc kubenswrapper[4789]: I1122 08:16:46.828698 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4be1163c-8900-4db9-a041-826526ae8751-logs\") pod \"barbican-worker-75d6db87b5-kbchw\" (UID: \"4be1163c-8900-4db9-a041-826526ae8751\") " pod="openstack/barbican-worker-75d6db87b5-kbchw" Nov 22 08:16:46 crc kubenswrapper[4789]: I1122 08:16:46.828724 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/affd3f11-c0d4-41a2-888b-f4ae0c7e3d41-config-data-custom\") pod \"barbican-keystone-listener-7ccbbd9f98-jd5vz\" (UID: \"affd3f11-c0d4-41a2-888b-f4ae0c7e3d41\") " pod="openstack/barbican-keystone-listener-7ccbbd9f98-jd5vz" Nov 22 08:16:46 crc kubenswrapper[4789]: I1122 08:16:46.883576 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-6976474df8-jfq2h"] Nov 22 08:16:46 crc kubenswrapper[4789]: I1122 08:16:46.884854 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6976474df8-jfq2h" Nov 22 08:16:46 crc kubenswrapper[4789]: I1122 08:16:46.887404 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Nov 22 08:16:46 crc kubenswrapper[4789]: I1122 08:16:46.902827 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6976474df8-jfq2h"] Nov 22 08:16:46 crc kubenswrapper[4789]: I1122 08:16:46.930408 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/affd3f11-c0d4-41a2-888b-f4ae0c7e3d41-config-data-custom\") pod \"barbican-keystone-listener-7ccbbd9f98-jd5vz\" (UID: \"affd3f11-c0d4-41a2-888b-f4ae0c7e3d41\") " pod="openstack/barbican-keystone-listener-7ccbbd9f98-jd5vz" Nov 22 08:16:46 crc kubenswrapper[4789]: I1122 08:16:46.930470 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zr9fz\" (UniqueName: \"kubernetes.io/projected/affd3f11-c0d4-41a2-888b-f4ae0c7e3d41-kube-api-access-zr9fz\") pod \"barbican-keystone-listener-7ccbbd9f98-jd5vz\" (UID: \"affd3f11-c0d4-41a2-888b-f4ae0c7e3d41\") " pod="openstack/barbican-keystone-listener-7ccbbd9f98-jd5vz" Nov 22 08:16:46 crc kubenswrapper[4789]: I1122 08:16:46.930507 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2d8e424e-6fdf-4967-a24c-0a70af3724e3-ovsdbserver-nb\") pod \"dnsmasq-dns-68cd7949f9-tqtjq\" (UID: \"2d8e424e-6fdf-4967-a24c-0a70af3724e3\") " pod="openstack/dnsmasq-dns-68cd7949f9-tqtjq" Nov 22 08:16:46 crc kubenswrapper[4789]: I1122 08:16:46.930545 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2d8e424e-6fdf-4967-a24c-0a70af3724e3-ovsdbserver-sb\") pod \"dnsmasq-dns-68cd7949f9-tqtjq\" (UID: \"2d8e424e-6fdf-4967-a24c-0a70af3724e3\") " pod="openstack/dnsmasq-dns-68cd7949f9-tqtjq" Nov 22 08:16:46 crc kubenswrapper[4789]: I1122 08:16:46.930587 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4be1163c-8900-4db9-a041-826526ae8751-config-data\") pod \"barbican-worker-75d6db87b5-kbchw\" (UID: \"4be1163c-8900-4db9-a041-826526ae8751\") " pod="openstack/barbican-worker-75d6db87b5-kbchw" Nov 22 08:16:46 crc kubenswrapper[4789]: I1122 08:16:46.930614 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2d8e424e-6fdf-4967-a24c-0a70af3724e3-dns-svc\") pod \"dnsmasq-dns-68cd7949f9-tqtjq\" (UID: \"2d8e424e-6fdf-4967-a24c-0a70af3724e3\") " pod="openstack/dnsmasq-dns-68cd7949f9-tqtjq" Nov 22 08:16:46 crc kubenswrapper[4789]: I1122 08:16:46.930642 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d8e424e-6fdf-4967-a24c-0a70af3724e3-config\") pod \"dnsmasq-dns-68cd7949f9-tqtjq\" (UID: \"2d8e424e-6fdf-4967-a24c-0a70af3724e3\") " pod="openstack/dnsmasq-dns-68cd7949f9-tqtjq" Nov 22 08:16:46 crc kubenswrapper[4789]: I1122 08:16:46.930666 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rmth9\" (UniqueName: \"kubernetes.io/projected/4be1163c-8900-4db9-a041-826526ae8751-kube-api-access-rmth9\") pod \"barbican-worker-75d6db87b5-kbchw\" (UID: \"4be1163c-8900-4db9-a041-826526ae8751\") " pod="openstack/barbican-worker-75d6db87b5-kbchw" Nov 22 08:16:46 crc kubenswrapper[4789]: I1122 08:16:46.930689 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4be1163c-8900-4db9-a041-826526ae8751-combined-ca-bundle\") pod \"barbican-worker-75d6db87b5-kbchw\" (UID: \"4be1163c-8900-4db9-a041-826526ae8751\") " pod="openstack/barbican-worker-75d6db87b5-kbchw" Nov 22 08:16:46 crc kubenswrapper[4789]: I1122 08:16:46.930769 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/affd3f11-c0d4-41a2-888b-f4ae0c7e3d41-combined-ca-bundle\") pod \"barbican-keystone-listener-7ccbbd9f98-jd5vz\" (UID: \"affd3f11-c0d4-41a2-888b-f4ae0c7e3d41\") " pod="openstack/barbican-keystone-listener-7ccbbd9f98-jd5vz" Nov 22 08:16:46 crc kubenswrapper[4789]: I1122 08:16:46.930805 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/affd3f11-c0d4-41a2-888b-f4ae0c7e3d41-logs\") pod \"barbican-keystone-listener-7ccbbd9f98-jd5vz\" (UID: \"affd3f11-c0d4-41a2-888b-f4ae0c7e3d41\") " pod="openstack/barbican-keystone-listener-7ccbbd9f98-jd5vz" Nov 22 08:16:46 crc kubenswrapper[4789]: I1122 08:16:46.930842 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/affd3f11-c0d4-41a2-888b-f4ae0c7e3d41-config-data\") pod \"barbican-keystone-listener-7ccbbd9f98-jd5vz\" (UID: \"affd3f11-c0d4-41a2-888b-f4ae0c7e3d41\") " pod="openstack/barbican-keystone-listener-7ccbbd9f98-jd5vz" Nov 22 08:16:46 crc kubenswrapper[4789]: I1122 08:16:46.930869 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h6lm7\" (UniqueName: \"kubernetes.io/projected/2d8e424e-6fdf-4967-a24c-0a70af3724e3-kube-api-access-h6lm7\") pod \"dnsmasq-dns-68cd7949f9-tqtjq\" (UID: \"2d8e424e-6fdf-4967-a24c-0a70af3724e3\") " pod="openstack/dnsmasq-dns-68cd7949f9-tqtjq" Nov 22 08:16:46 crc kubenswrapper[4789]: I1122 08:16:46.930893 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4be1163c-8900-4db9-a041-826526ae8751-config-data-custom\") pod \"barbican-worker-75d6db87b5-kbchw\" (UID: \"4be1163c-8900-4db9-a041-826526ae8751\") " pod="openstack/barbican-worker-75d6db87b5-kbchw" Nov 22 08:16:46 crc kubenswrapper[4789]: I1122 08:16:46.930912 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4be1163c-8900-4db9-a041-826526ae8751-logs\") pod \"barbican-worker-75d6db87b5-kbchw\" (UID: \"4be1163c-8900-4db9-a041-826526ae8751\") " pod="openstack/barbican-worker-75d6db87b5-kbchw" Nov 22 08:16:46 crc kubenswrapper[4789]: I1122 08:16:46.931307 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4be1163c-8900-4db9-a041-826526ae8751-logs\") pod \"barbican-worker-75d6db87b5-kbchw\" (UID: \"4be1163c-8900-4db9-a041-826526ae8751\") " pod="openstack/barbican-worker-75d6db87b5-kbchw" Nov 22 08:16:46 crc kubenswrapper[4789]: I1122 08:16:46.931835 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/affd3f11-c0d4-41a2-888b-f4ae0c7e3d41-logs\") pod \"barbican-keystone-listener-7ccbbd9f98-jd5vz\" (UID: \"affd3f11-c0d4-41a2-888b-f4ae0c7e3d41\") " pod="openstack/barbican-keystone-listener-7ccbbd9f98-jd5vz" Nov 22 08:16:46 crc kubenswrapper[4789]: I1122 08:16:46.937716 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/affd3f11-c0d4-41a2-888b-f4ae0c7e3d41-config-data-custom\") pod \"barbican-keystone-listener-7ccbbd9f98-jd5vz\" (UID: \"affd3f11-c0d4-41a2-888b-f4ae0c7e3d41\") " pod="openstack/barbican-keystone-listener-7ccbbd9f98-jd5vz" Nov 22 08:16:46 crc kubenswrapper[4789]: I1122 08:16:46.937976 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4be1163c-8900-4db9-a041-826526ae8751-combined-ca-bundle\") pod \"barbican-worker-75d6db87b5-kbchw\" (UID: \"4be1163c-8900-4db9-a041-826526ae8751\") " pod="openstack/barbican-worker-75d6db87b5-kbchw" Nov 22 08:16:46 crc kubenswrapper[4789]: I1122 08:16:46.938111 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/affd3f11-c0d4-41a2-888b-f4ae0c7e3d41-combined-ca-bundle\") pod \"barbican-keystone-listener-7ccbbd9f98-jd5vz\" (UID: \"affd3f11-c0d4-41a2-888b-f4ae0c7e3d41\") " pod="openstack/barbican-keystone-listener-7ccbbd9f98-jd5vz" Nov 22 08:16:46 crc kubenswrapper[4789]: I1122 08:16:46.939309 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4be1163c-8900-4db9-a041-826526ae8751-config-data-custom\") pod \"barbican-worker-75d6db87b5-kbchw\" (UID: \"4be1163c-8900-4db9-a041-826526ae8751\") " pod="openstack/barbican-worker-75d6db87b5-kbchw" Nov 22 08:16:46 crc kubenswrapper[4789]: I1122 08:16:46.940838 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4be1163c-8900-4db9-a041-826526ae8751-config-data\") pod \"barbican-worker-75d6db87b5-kbchw\" (UID: \"4be1163c-8900-4db9-a041-826526ae8751\") " pod="openstack/barbican-worker-75d6db87b5-kbchw" Nov 22 08:16:46 crc kubenswrapper[4789]: I1122 08:16:46.941104 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/affd3f11-c0d4-41a2-888b-f4ae0c7e3d41-config-data\") pod \"barbican-keystone-listener-7ccbbd9f98-jd5vz\" (UID: \"affd3f11-c0d4-41a2-888b-f4ae0c7e3d41\") " pod="openstack/barbican-keystone-listener-7ccbbd9f98-jd5vz" Nov 22 08:16:46 crc kubenswrapper[4789]: I1122 08:16:46.946515 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rmth9\" (UniqueName: \"kubernetes.io/projected/4be1163c-8900-4db9-a041-826526ae8751-kube-api-access-rmth9\") pod \"barbican-worker-75d6db87b5-kbchw\" (UID: \"4be1163c-8900-4db9-a041-826526ae8751\") " pod="openstack/barbican-worker-75d6db87b5-kbchw" Nov 22 08:16:46 crc kubenswrapper[4789]: I1122 08:16:46.951247 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zr9fz\" (UniqueName: \"kubernetes.io/projected/affd3f11-c0d4-41a2-888b-f4ae0c7e3d41-kube-api-access-zr9fz\") pod \"barbican-keystone-listener-7ccbbd9f98-jd5vz\" (UID: \"affd3f11-c0d4-41a2-888b-f4ae0c7e3d41\") " pod="openstack/barbican-keystone-listener-7ccbbd9f98-jd5vz" Nov 22 08:16:47 crc kubenswrapper[4789]: I1122 08:16:47.012187 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-75d6db87b5-kbchw" Nov 22 08:16:47 crc kubenswrapper[4789]: I1122 08:16:47.032565 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e382d956-58e8-4c3b-abd9-0732d19bc35e-combined-ca-bundle\") pod \"barbican-api-6976474df8-jfq2h\" (UID: \"e382d956-58e8-4c3b-abd9-0732d19bc35e\") " pod="openstack/barbican-api-6976474df8-jfq2h" Nov 22 08:16:47 crc kubenswrapper[4789]: I1122 08:16:47.032651 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e382d956-58e8-4c3b-abd9-0732d19bc35e-logs\") pod \"barbican-api-6976474df8-jfq2h\" (UID: \"e382d956-58e8-4c3b-abd9-0732d19bc35e\") " pod="openstack/barbican-api-6976474df8-jfq2h" Nov 22 08:16:47 crc kubenswrapper[4789]: I1122 08:16:47.032696 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h6lm7\" (UniqueName: \"kubernetes.io/projected/2d8e424e-6fdf-4967-a24c-0a70af3724e3-kube-api-access-h6lm7\") pod \"dnsmasq-dns-68cd7949f9-tqtjq\" (UID: \"2d8e424e-6fdf-4967-a24c-0a70af3724e3\") " pod="openstack/dnsmasq-dns-68cd7949f9-tqtjq" Nov 22 08:16:47 crc kubenswrapper[4789]: I1122 08:16:47.032738 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e382d956-58e8-4c3b-abd9-0732d19bc35e-config-data\") pod \"barbican-api-6976474df8-jfq2h\" (UID: \"e382d956-58e8-4c3b-abd9-0732d19bc35e\") " pod="openstack/barbican-api-6976474df8-jfq2h" Nov 22 08:16:47 crc kubenswrapper[4789]: I1122 08:16:47.032791 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2d8e424e-6fdf-4967-a24c-0a70af3724e3-ovsdbserver-nb\") pod \"dnsmasq-dns-68cd7949f9-tqtjq\" (UID: \"2d8e424e-6fdf-4967-a24c-0a70af3724e3\") " pod="openstack/dnsmasq-dns-68cd7949f9-tqtjq" Nov 22 08:16:47 crc kubenswrapper[4789]: I1122 08:16:47.032811 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e382d956-58e8-4c3b-abd9-0732d19bc35e-config-data-custom\") pod \"barbican-api-6976474df8-jfq2h\" (UID: \"e382d956-58e8-4c3b-abd9-0732d19bc35e\") " pod="openstack/barbican-api-6976474df8-jfq2h" Nov 22 08:16:47 crc kubenswrapper[4789]: I1122 08:16:47.032844 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2d8e424e-6fdf-4967-a24c-0a70af3724e3-ovsdbserver-sb\") pod \"dnsmasq-dns-68cd7949f9-tqtjq\" (UID: \"2d8e424e-6fdf-4967-a24c-0a70af3724e3\") " pod="openstack/dnsmasq-dns-68cd7949f9-tqtjq" Nov 22 08:16:47 crc kubenswrapper[4789]: I1122 08:16:47.032910 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lw2d2\" (UniqueName: \"kubernetes.io/projected/e382d956-58e8-4c3b-abd9-0732d19bc35e-kube-api-access-lw2d2\") pod \"barbican-api-6976474df8-jfq2h\" (UID: \"e382d956-58e8-4c3b-abd9-0732d19bc35e\") " pod="openstack/barbican-api-6976474df8-jfq2h" Nov 22 08:16:47 crc kubenswrapper[4789]: I1122 08:16:47.032941 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2d8e424e-6fdf-4967-a24c-0a70af3724e3-dns-svc\") pod \"dnsmasq-dns-68cd7949f9-tqtjq\" (UID: \"2d8e424e-6fdf-4967-a24c-0a70af3724e3\") " pod="openstack/dnsmasq-dns-68cd7949f9-tqtjq" Nov 22 08:16:47 crc kubenswrapper[4789]: I1122 08:16:47.032970 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d8e424e-6fdf-4967-a24c-0a70af3724e3-config\") pod \"dnsmasq-dns-68cd7949f9-tqtjq\" (UID: \"2d8e424e-6fdf-4967-a24c-0a70af3724e3\") " pod="openstack/dnsmasq-dns-68cd7949f9-tqtjq" Nov 22 08:16:47 crc kubenswrapper[4789]: I1122 08:16:47.033994 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d8e424e-6fdf-4967-a24c-0a70af3724e3-config\") pod \"dnsmasq-dns-68cd7949f9-tqtjq\" (UID: \"2d8e424e-6fdf-4967-a24c-0a70af3724e3\") " pod="openstack/dnsmasq-dns-68cd7949f9-tqtjq" Nov 22 08:16:47 crc kubenswrapper[4789]: I1122 08:16:47.034111 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2d8e424e-6fdf-4967-a24c-0a70af3724e3-ovsdbserver-nb\") pod \"dnsmasq-dns-68cd7949f9-tqtjq\" (UID: \"2d8e424e-6fdf-4967-a24c-0a70af3724e3\") " pod="openstack/dnsmasq-dns-68cd7949f9-tqtjq" Nov 22 08:16:47 crc kubenswrapper[4789]: I1122 08:16:47.034323 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2d8e424e-6fdf-4967-a24c-0a70af3724e3-ovsdbserver-sb\") pod \"dnsmasq-dns-68cd7949f9-tqtjq\" (UID: \"2d8e424e-6fdf-4967-a24c-0a70af3724e3\") " pod="openstack/dnsmasq-dns-68cd7949f9-tqtjq" Nov 22 08:16:47 crc kubenswrapper[4789]: I1122 08:16:47.034346 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2d8e424e-6fdf-4967-a24c-0a70af3724e3-dns-svc\") pod \"dnsmasq-dns-68cd7949f9-tqtjq\" (UID: \"2d8e424e-6fdf-4967-a24c-0a70af3724e3\") " pod="openstack/dnsmasq-dns-68cd7949f9-tqtjq" Nov 22 08:16:47 crc kubenswrapper[4789]: I1122 08:16:47.052528 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h6lm7\" (UniqueName: \"kubernetes.io/projected/2d8e424e-6fdf-4967-a24c-0a70af3724e3-kube-api-access-h6lm7\") pod \"dnsmasq-dns-68cd7949f9-tqtjq\" (UID: \"2d8e424e-6fdf-4967-a24c-0a70af3724e3\") " pod="openstack/dnsmasq-dns-68cd7949f9-tqtjq" Nov 22 08:16:47 crc kubenswrapper[4789]: I1122 08:16:47.061566 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-7ccbbd9f98-jd5vz" Nov 22 08:16:47 crc kubenswrapper[4789]: I1122 08:16:47.135196 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e382d956-58e8-4c3b-abd9-0732d19bc35e-combined-ca-bundle\") pod \"barbican-api-6976474df8-jfq2h\" (UID: \"e382d956-58e8-4c3b-abd9-0732d19bc35e\") " pod="openstack/barbican-api-6976474df8-jfq2h" Nov 22 08:16:47 crc kubenswrapper[4789]: I1122 08:16:47.135547 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e382d956-58e8-4c3b-abd9-0732d19bc35e-logs\") pod \"barbican-api-6976474df8-jfq2h\" (UID: \"e382d956-58e8-4c3b-abd9-0732d19bc35e\") " pod="openstack/barbican-api-6976474df8-jfq2h" Nov 22 08:16:47 crc kubenswrapper[4789]: I1122 08:16:47.135604 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e382d956-58e8-4c3b-abd9-0732d19bc35e-config-data\") pod \"barbican-api-6976474df8-jfq2h\" (UID: \"e382d956-58e8-4c3b-abd9-0732d19bc35e\") " pod="openstack/barbican-api-6976474df8-jfq2h" Nov 22 08:16:47 crc kubenswrapper[4789]: I1122 08:16:47.135661 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e382d956-58e8-4c3b-abd9-0732d19bc35e-config-data-custom\") pod \"barbican-api-6976474df8-jfq2h\" (UID: \"e382d956-58e8-4c3b-abd9-0732d19bc35e\") " pod="openstack/barbican-api-6976474df8-jfq2h" Nov 22 08:16:47 crc kubenswrapper[4789]: I1122 08:16:47.135716 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lw2d2\" (UniqueName: \"kubernetes.io/projected/e382d956-58e8-4c3b-abd9-0732d19bc35e-kube-api-access-lw2d2\") pod \"barbican-api-6976474df8-jfq2h\" (UID: \"e382d956-58e8-4c3b-abd9-0732d19bc35e\") " pod="openstack/barbican-api-6976474df8-jfq2h" Nov 22 08:16:47 crc kubenswrapper[4789]: I1122 08:16:47.137799 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e382d956-58e8-4c3b-abd9-0732d19bc35e-logs\") pod \"barbican-api-6976474df8-jfq2h\" (UID: \"e382d956-58e8-4c3b-abd9-0732d19bc35e\") " pod="openstack/barbican-api-6976474df8-jfq2h" Nov 22 08:16:47 crc kubenswrapper[4789]: I1122 08:16:47.139871 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-68cd7949f9-tqtjq" Nov 22 08:16:47 crc kubenswrapper[4789]: I1122 08:16:47.142189 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e382d956-58e8-4c3b-abd9-0732d19bc35e-combined-ca-bundle\") pod \"barbican-api-6976474df8-jfq2h\" (UID: \"e382d956-58e8-4c3b-abd9-0732d19bc35e\") " pod="openstack/barbican-api-6976474df8-jfq2h" Nov 22 08:16:47 crc kubenswrapper[4789]: I1122 08:16:47.144371 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e382d956-58e8-4c3b-abd9-0732d19bc35e-config-data\") pod \"barbican-api-6976474df8-jfq2h\" (UID: \"e382d956-58e8-4c3b-abd9-0732d19bc35e\") " pod="openstack/barbican-api-6976474df8-jfq2h" Nov 22 08:16:47 crc kubenswrapper[4789]: I1122 08:16:47.145404 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e382d956-58e8-4c3b-abd9-0732d19bc35e-config-data-custom\") pod \"barbican-api-6976474df8-jfq2h\" (UID: \"e382d956-58e8-4c3b-abd9-0732d19bc35e\") " pod="openstack/barbican-api-6976474df8-jfq2h" Nov 22 08:16:47 crc kubenswrapper[4789]: I1122 08:16:47.158043 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lw2d2\" (UniqueName: \"kubernetes.io/projected/e382d956-58e8-4c3b-abd9-0732d19bc35e-kube-api-access-lw2d2\") pod \"barbican-api-6976474df8-jfq2h\" (UID: \"e382d956-58e8-4c3b-abd9-0732d19bc35e\") " pod="openstack/barbican-api-6976474df8-jfq2h" Nov 22 08:16:47 crc kubenswrapper[4789]: I1122 08:16:47.201801 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6976474df8-jfq2h" Nov 22 08:16:47 crc kubenswrapper[4789]: I1122 08:16:47.485624 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-75d6db87b5-kbchw"] Nov 22 08:16:47 crc kubenswrapper[4789]: I1122 08:16:47.487015 4789 generic.go:334] "Generic (PLEG): container finished" podID="6bf8aa3e-a7e7-4f0c-a542-d668a5623707" containerID="50273be882e06c5bebf9373f13c41c1860a0308e06ec2d3632e22d25ca316617" exitCode=0 Nov 22 08:16:47 crc kubenswrapper[4789]: I1122 08:16:47.487093 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-lg59m" event={"ID":"6bf8aa3e-a7e7-4f0c-a542-d668a5623707","Type":"ContainerDied","Data":"50273be882e06c5bebf9373f13c41c1860a0308e06ec2d3632e22d25ca316617"} Nov 22 08:16:47 crc kubenswrapper[4789]: I1122 08:16:47.494134 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"94d7bb01-64a7-442a-aee4-e958d8511b76","Type":"ContainerStarted","Data":"3381e3d40939b89f2f49a7f7a930e51358fcd14142eac9225f51202ec1a339a4"} Nov 22 08:16:47 crc kubenswrapper[4789]: I1122 08:16:47.494355 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="94d7bb01-64a7-442a-aee4-e958d8511b76" containerName="ceilometer-central-agent" containerID="cri-o://5c1bd328b27a7fd1cdca6ff567666e218a70a3b926831af09bf3a542194a3a1b" gracePeriod=30 Nov 22 08:16:47 crc kubenswrapper[4789]: I1122 08:16:47.494659 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 22 08:16:47 crc kubenswrapper[4789]: I1122 08:16:47.494715 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="94d7bb01-64a7-442a-aee4-e958d8511b76" containerName="proxy-httpd" containerID="cri-o://3381e3d40939b89f2f49a7f7a930e51358fcd14142eac9225f51202ec1a339a4" gracePeriod=30 Nov 22 08:16:47 crc kubenswrapper[4789]: I1122 08:16:47.494798 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="94d7bb01-64a7-442a-aee4-e958d8511b76" containerName="sg-core" containerID="cri-o://8463b06d53f122d2dd4bda6d1f5b32b440b64a9e858e15291e72b4e4f847355f" gracePeriod=30 Nov 22 08:16:47 crc kubenswrapper[4789]: I1122 08:16:47.494871 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="94d7bb01-64a7-442a-aee4-e958d8511b76" containerName="ceilometer-notification-agent" containerID="cri-o://b8f37fba2551f2fc622fc288b287229e43574f1d9231f7967747f9645db2d0e9" gracePeriod=30 Nov 22 08:16:47 crc kubenswrapper[4789]: I1122 08:16:47.501385 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-68cd7949f9-tqtjq"] Nov 22 08:16:47 crc kubenswrapper[4789]: I1122 08:16:47.573267 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-844b557b9c-lvrgl"] Nov 22 08:16:47 crc kubenswrapper[4789]: I1122 08:16:47.575055 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-844b557b9c-lvrgl" Nov 22 08:16:47 crc kubenswrapper[4789]: I1122 08:16:47.595946 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-844b557b9c-lvrgl"] Nov 22 08:16:47 crc kubenswrapper[4789]: I1122 08:16:47.615217 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.208472139 podStartE2EDuration="58.615192882s" podCreationTimestamp="2025-11-22 08:15:49 +0000 UTC" firstStartedPulling="2025-11-22 08:15:50.709715667 +0000 UTC m=+1164.944115940" lastFinishedPulling="2025-11-22 08:16:46.11643641 +0000 UTC m=+1220.350836683" observedRunningTime="2025-11-22 08:16:47.545435471 +0000 UTC m=+1221.779835744" watchObservedRunningTime="2025-11-22 08:16:47.615192882 +0000 UTC m=+1221.849593155" Nov 22 08:16:47 crc kubenswrapper[4789]: I1122 08:16:47.666204 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-7ccbbd9f98-jd5vz"] Nov 22 08:16:47 crc kubenswrapper[4789]: I1122 08:16:47.679855 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-568cd49f4d-ms9nh"] Nov 22 08:16:47 crc kubenswrapper[4789]: I1122 08:16:47.681214 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-568cd49f4d-ms9nh" Nov 22 08:16:47 crc kubenswrapper[4789]: I1122 08:16:47.684922 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-zhkq9" Nov 22 08:16:47 crc kubenswrapper[4789]: I1122 08:16:47.685230 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Nov 22 08:16:47 crc kubenswrapper[4789]: I1122 08:16:47.686967 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Nov 22 08:16:47 crc kubenswrapper[4789]: I1122 08:16:47.687723 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Nov 22 08:16:47 crc kubenswrapper[4789]: I1122 08:16:47.690725 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-568cd49f4d-ms9nh"] Nov 22 08:16:47 crc kubenswrapper[4789]: I1122 08:16:47.728113 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-68cd7949f9-tqtjq"] Nov 22 08:16:47 crc kubenswrapper[4789]: W1122 08:16:47.730155 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2d8e424e_6fdf_4967_a24c_0a70af3724e3.slice/crio-b4ffbb41d12659a706e25cde9d916c4388b3b507070551f1a3086a2530841672 WatchSource:0}: Error finding container b4ffbb41d12659a706e25cde9d916c4388b3b507070551f1a3086a2530841672: Status 404 returned error can't find the container with id b4ffbb41d12659a706e25cde9d916c4388b3b507070551f1a3086a2530841672 Nov 22 08:16:47 crc kubenswrapper[4789]: I1122 08:16:47.749396 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lk4xm\" (UniqueName: \"kubernetes.io/projected/f3130d2e-6671-4a4b-978b-eac10b62cdc8-kube-api-access-lk4xm\") pod \"dnsmasq-dns-844b557b9c-lvrgl\" (UID: \"f3130d2e-6671-4a4b-978b-eac10b62cdc8\") " pod="openstack/dnsmasq-dns-844b557b9c-lvrgl" Nov 22 08:16:47 crc kubenswrapper[4789]: I1122 08:16:47.749489 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f3130d2e-6671-4a4b-978b-eac10b62cdc8-config\") pod \"dnsmasq-dns-844b557b9c-lvrgl\" (UID: \"f3130d2e-6671-4a4b-978b-eac10b62cdc8\") " pod="openstack/dnsmasq-dns-844b557b9c-lvrgl" Nov 22 08:16:47 crc kubenswrapper[4789]: I1122 08:16:47.749518 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f3130d2e-6671-4a4b-978b-eac10b62cdc8-dns-svc\") pod \"dnsmasq-dns-844b557b9c-lvrgl\" (UID: \"f3130d2e-6671-4a4b-978b-eac10b62cdc8\") " pod="openstack/dnsmasq-dns-844b557b9c-lvrgl" Nov 22 08:16:47 crc kubenswrapper[4789]: I1122 08:16:47.749563 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f3130d2e-6671-4a4b-978b-eac10b62cdc8-ovsdbserver-nb\") pod \"dnsmasq-dns-844b557b9c-lvrgl\" (UID: \"f3130d2e-6671-4a4b-978b-eac10b62cdc8\") " pod="openstack/dnsmasq-dns-844b557b9c-lvrgl" Nov 22 08:16:47 crc kubenswrapper[4789]: I1122 08:16:47.749600 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f3130d2e-6671-4a4b-978b-eac10b62cdc8-ovsdbserver-sb\") pod \"dnsmasq-dns-844b557b9c-lvrgl\" (UID: \"f3130d2e-6671-4a4b-978b-eac10b62cdc8\") " pod="openstack/dnsmasq-dns-844b557b9c-lvrgl" Nov 22 08:16:47 crc kubenswrapper[4789]: I1122 08:16:47.851033 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4jjl7\" (UniqueName: \"kubernetes.io/projected/2949ebe1-7541-42a2-894c-2588c2bd3396-kube-api-access-4jjl7\") pod \"neutron-568cd49f4d-ms9nh\" (UID: \"2949ebe1-7541-42a2-894c-2588c2bd3396\") " pod="openstack/neutron-568cd49f4d-ms9nh" Nov 22 08:16:47 crc kubenswrapper[4789]: I1122 08:16:47.851357 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lk4xm\" (UniqueName: \"kubernetes.io/projected/f3130d2e-6671-4a4b-978b-eac10b62cdc8-kube-api-access-lk4xm\") pod \"dnsmasq-dns-844b557b9c-lvrgl\" (UID: \"f3130d2e-6671-4a4b-978b-eac10b62cdc8\") " pod="openstack/dnsmasq-dns-844b557b9c-lvrgl" Nov 22 08:16:47 crc kubenswrapper[4789]: I1122 08:16:47.851402 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/2949ebe1-7541-42a2-894c-2588c2bd3396-httpd-config\") pod \"neutron-568cd49f4d-ms9nh\" (UID: \"2949ebe1-7541-42a2-894c-2588c2bd3396\") " pod="openstack/neutron-568cd49f4d-ms9nh" Nov 22 08:16:47 crc kubenswrapper[4789]: I1122 08:16:47.851431 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f3130d2e-6671-4a4b-978b-eac10b62cdc8-config\") pod \"dnsmasq-dns-844b557b9c-lvrgl\" (UID: \"f3130d2e-6671-4a4b-978b-eac10b62cdc8\") " pod="openstack/dnsmasq-dns-844b557b9c-lvrgl" Nov 22 08:16:47 crc kubenswrapper[4789]: I1122 08:16:47.851449 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f3130d2e-6671-4a4b-978b-eac10b62cdc8-dns-svc\") pod \"dnsmasq-dns-844b557b9c-lvrgl\" (UID: \"f3130d2e-6671-4a4b-978b-eac10b62cdc8\") " pod="openstack/dnsmasq-dns-844b557b9c-lvrgl" Nov 22 08:16:47 crc kubenswrapper[4789]: I1122 08:16:47.851471 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2949ebe1-7541-42a2-894c-2588c2bd3396-combined-ca-bundle\") pod \"neutron-568cd49f4d-ms9nh\" (UID: \"2949ebe1-7541-42a2-894c-2588c2bd3396\") " pod="openstack/neutron-568cd49f4d-ms9nh" Nov 22 08:16:47 crc kubenswrapper[4789]: I1122 08:16:47.851496 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f3130d2e-6671-4a4b-978b-eac10b62cdc8-ovsdbserver-nb\") pod \"dnsmasq-dns-844b557b9c-lvrgl\" (UID: \"f3130d2e-6671-4a4b-978b-eac10b62cdc8\") " pod="openstack/dnsmasq-dns-844b557b9c-lvrgl" Nov 22 08:16:47 crc kubenswrapper[4789]: I1122 08:16:47.851522 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f3130d2e-6671-4a4b-978b-eac10b62cdc8-ovsdbserver-sb\") pod \"dnsmasq-dns-844b557b9c-lvrgl\" (UID: \"f3130d2e-6671-4a4b-978b-eac10b62cdc8\") " pod="openstack/dnsmasq-dns-844b557b9c-lvrgl" Nov 22 08:16:47 crc kubenswrapper[4789]: I1122 08:16:47.851543 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/2949ebe1-7541-42a2-894c-2588c2bd3396-config\") pod \"neutron-568cd49f4d-ms9nh\" (UID: \"2949ebe1-7541-42a2-894c-2588c2bd3396\") " pod="openstack/neutron-568cd49f4d-ms9nh" Nov 22 08:16:47 crc kubenswrapper[4789]: I1122 08:16:47.851564 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/2949ebe1-7541-42a2-894c-2588c2bd3396-ovndb-tls-certs\") pod \"neutron-568cd49f4d-ms9nh\" (UID: \"2949ebe1-7541-42a2-894c-2588c2bd3396\") " pod="openstack/neutron-568cd49f4d-ms9nh" Nov 22 08:16:47 crc kubenswrapper[4789]: I1122 08:16:47.852692 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f3130d2e-6671-4a4b-978b-eac10b62cdc8-config\") pod \"dnsmasq-dns-844b557b9c-lvrgl\" (UID: \"f3130d2e-6671-4a4b-978b-eac10b62cdc8\") " pod="openstack/dnsmasq-dns-844b557b9c-lvrgl" Nov 22 08:16:47 crc kubenswrapper[4789]: I1122 08:16:47.853401 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f3130d2e-6671-4a4b-978b-eac10b62cdc8-ovsdbserver-nb\") pod \"dnsmasq-dns-844b557b9c-lvrgl\" (UID: \"f3130d2e-6671-4a4b-978b-eac10b62cdc8\") " pod="openstack/dnsmasq-dns-844b557b9c-lvrgl" Nov 22 08:16:47 crc kubenswrapper[4789]: I1122 08:16:47.853458 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f3130d2e-6671-4a4b-978b-eac10b62cdc8-ovsdbserver-sb\") pod \"dnsmasq-dns-844b557b9c-lvrgl\" (UID: \"f3130d2e-6671-4a4b-978b-eac10b62cdc8\") " pod="openstack/dnsmasq-dns-844b557b9c-lvrgl" Nov 22 08:16:47 crc kubenswrapper[4789]: I1122 08:16:47.854050 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f3130d2e-6671-4a4b-978b-eac10b62cdc8-dns-svc\") pod \"dnsmasq-dns-844b557b9c-lvrgl\" (UID: \"f3130d2e-6671-4a4b-978b-eac10b62cdc8\") " pod="openstack/dnsmasq-dns-844b557b9c-lvrgl" Nov 22 08:16:47 crc kubenswrapper[4789]: I1122 08:16:47.872011 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lk4xm\" (UniqueName: \"kubernetes.io/projected/f3130d2e-6671-4a4b-978b-eac10b62cdc8-kube-api-access-lk4xm\") pod \"dnsmasq-dns-844b557b9c-lvrgl\" (UID: \"f3130d2e-6671-4a4b-978b-eac10b62cdc8\") " pod="openstack/dnsmasq-dns-844b557b9c-lvrgl" Nov 22 08:16:47 crc kubenswrapper[4789]: I1122 08:16:47.918392 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-844b557b9c-lvrgl" Nov 22 08:16:47 crc kubenswrapper[4789]: I1122 08:16:47.922266 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6976474df8-jfq2h"] Nov 22 08:16:47 crc kubenswrapper[4789]: I1122 08:16:47.952941 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/2949ebe1-7541-42a2-894c-2588c2bd3396-config\") pod \"neutron-568cd49f4d-ms9nh\" (UID: \"2949ebe1-7541-42a2-894c-2588c2bd3396\") " pod="openstack/neutron-568cd49f4d-ms9nh" Nov 22 08:16:47 crc kubenswrapper[4789]: I1122 08:16:47.952989 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/2949ebe1-7541-42a2-894c-2588c2bd3396-ovndb-tls-certs\") pod \"neutron-568cd49f4d-ms9nh\" (UID: \"2949ebe1-7541-42a2-894c-2588c2bd3396\") " pod="openstack/neutron-568cd49f4d-ms9nh" Nov 22 08:16:47 crc kubenswrapper[4789]: I1122 08:16:47.953049 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4jjl7\" (UniqueName: \"kubernetes.io/projected/2949ebe1-7541-42a2-894c-2588c2bd3396-kube-api-access-4jjl7\") pod \"neutron-568cd49f4d-ms9nh\" (UID: \"2949ebe1-7541-42a2-894c-2588c2bd3396\") " pod="openstack/neutron-568cd49f4d-ms9nh" Nov 22 08:16:47 crc kubenswrapper[4789]: I1122 08:16:47.953107 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/2949ebe1-7541-42a2-894c-2588c2bd3396-httpd-config\") pod \"neutron-568cd49f4d-ms9nh\" (UID: \"2949ebe1-7541-42a2-894c-2588c2bd3396\") " pod="openstack/neutron-568cd49f4d-ms9nh" Nov 22 08:16:47 crc kubenswrapper[4789]: I1122 08:16:47.953161 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2949ebe1-7541-42a2-894c-2588c2bd3396-combined-ca-bundle\") pod \"neutron-568cd49f4d-ms9nh\" (UID: \"2949ebe1-7541-42a2-894c-2588c2bd3396\") " pod="openstack/neutron-568cd49f4d-ms9nh" Nov 22 08:16:47 crc kubenswrapper[4789]: I1122 08:16:47.957712 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2949ebe1-7541-42a2-894c-2588c2bd3396-combined-ca-bundle\") pod \"neutron-568cd49f4d-ms9nh\" (UID: \"2949ebe1-7541-42a2-894c-2588c2bd3396\") " pod="openstack/neutron-568cd49f4d-ms9nh" Nov 22 08:16:47 crc kubenswrapper[4789]: I1122 08:16:47.957445 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/2949ebe1-7541-42a2-894c-2588c2bd3396-config\") pod \"neutron-568cd49f4d-ms9nh\" (UID: \"2949ebe1-7541-42a2-894c-2588c2bd3396\") " pod="openstack/neutron-568cd49f4d-ms9nh" Nov 22 08:16:47 crc kubenswrapper[4789]: I1122 08:16:47.960045 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/2949ebe1-7541-42a2-894c-2588c2bd3396-ovndb-tls-certs\") pod \"neutron-568cd49f4d-ms9nh\" (UID: \"2949ebe1-7541-42a2-894c-2588c2bd3396\") " pod="openstack/neutron-568cd49f4d-ms9nh" Nov 22 08:16:47 crc kubenswrapper[4789]: I1122 08:16:47.962399 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/2949ebe1-7541-42a2-894c-2588c2bd3396-httpd-config\") pod \"neutron-568cd49f4d-ms9nh\" (UID: \"2949ebe1-7541-42a2-894c-2588c2bd3396\") " pod="openstack/neutron-568cd49f4d-ms9nh" Nov 22 08:16:47 crc kubenswrapper[4789]: I1122 08:16:47.973483 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4jjl7\" (UniqueName: \"kubernetes.io/projected/2949ebe1-7541-42a2-894c-2588c2bd3396-kube-api-access-4jjl7\") pod \"neutron-568cd49f4d-ms9nh\" (UID: \"2949ebe1-7541-42a2-894c-2588c2bd3396\") " pod="openstack/neutron-568cd49f4d-ms9nh" Nov 22 08:16:48 crc kubenswrapper[4789]: I1122 08:16:48.019980 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-568cd49f4d-ms9nh" Nov 22 08:16:48 crc kubenswrapper[4789]: I1122 08:16:48.487006 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-844b557b9c-lvrgl"] Nov 22 08:16:48 crc kubenswrapper[4789]: I1122 08:16:48.512898 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7ccbbd9f98-jd5vz" event={"ID":"affd3f11-c0d4-41a2-888b-f4ae0c7e3d41","Type":"ContainerStarted","Data":"0d6774eac3e732727dfb1df9ea73d1a30332e12829b43a243d5b4fb136cc1139"} Nov 22 08:16:48 crc kubenswrapper[4789]: I1122 08:16:48.518245 4789 generic.go:334] "Generic (PLEG): container finished" podID="2d8e424e-6fdf-4967-a24c-0a70af3724e3" containerID="437b6c2ad5d376597aece00254567a75824d4490c507e5e042f2dd3a306bbf83" exitCode=0 Nov 22 08:16:48 crc kubenswrapper[4789]: I1122 08:16:48.518342 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68cd7949f9-tqtjq" event={"ID":"2d8e424e-6fdf-4967-a24c-0a70af3724e3","Type":"ContainerDied","Data":"437b6c2ad5d376597aece00254567a75824d4490c507e5e042f2dd3a306bbf83"} Nov 22 08:16:48 crc kubenswrapper[4789]: I1122 08:16:48.518378 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68cd7949f9-tqtjq" event={"ID":"2d8e424e-6fdf-4967-a24c-0a70af3724e3","Type":"ContainerStarted","Data":"b4ffbb41d12659a706e25cde9d916c4388b3b507070551f1a3086a2530841672"} Nov 22 08:16:48 crc kubenswrapper[4789]: I1122 08:16:48.525489 4789 generic.go:334] "Generic (PLEG): container finished" podID="94d7bb01-64a7-442a-aee4-e958d8511b76" containerID="3381e3d40939b89f2f49a7f7a930e51358fcd14142eac9225f51202ec1a339a4" exitCode=0 Nov 22 08:16:48 crc kubenswrapper[4789]: I1122 08:16:48.525527 4789 generic.go:334] "Generic (PLEG): container finished" podID="94d7bb01-64a7-442a-aee4-e958d8511b76" containerID="8463b06d53f122d2dd4bda6d1f5b32b440b64a9e858e15291e72b4e4f847355f" exitCode=2 Nov 22 08:16:48 crc kubenswrapper[4789]: I1122 08:16:48.525534 4789 generic.go:334] "Generic (PLEG): container finished" podID="94d7bb01-64a7-442a-aee4-e958d8511b76" containerID="5c1bd328b27a7fd1cdca6ff567666e218a70a3b926831af09bf3a542194a3a1b" exitCode=0 Nov 22 08:16:48 crc kubenswrapper[4789]: I1122 08:16:48.525600 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"94d7bb01-64a7-442a-aee4-e958d8511b76","Type":"ContainerDied","Data":"3381e3d40939b89f2f49a7f7a930e51358fcd14142eac9225f51202ec1a339a4"} Nov 22 08:16:48 crc kubenswrapper[4789]: I1122 08:16:48.525654 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"94d7bb01-64a7-442a-aee4-e958d8511b76","Type":"ContainerDied","Data":"8463b06d53f122d2dd4bda6d1f5b32b440b64a9e858e15291e72b4e4f847355f"} Nov 22 08:16:48 crc kubenswrapper[4789]: I1122 08:16:48.525666 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"94d7bb01-64a7-442a-aee4-e958d8511b76","Type":"ContainerDied","Data":"5c1bd328b27a7fd1cdca6ff567666e218a70a3b926831af09bf3a542194a3a1b"} Nov 22 08:16:48 crc kubenswrapper[4789]: I1122 08:16:48.531400 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6976474df8-jfq2h" event={"ID":"e382d956-58e8-4c3b-abd9-0732d19bc35e","Type":"ContainerStarted","Data":"612d0ac99e38a37b8ca19ff88131467979ce665edafa1d0125ab68a49d7e8682"} Nov 22 08:16:48 crc kubenswrapper[4789]: I1122 08:16:48.531461 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6976474df8-jfq2h" event={"ID":"e382d956-58e8-4c3b-abd9-0732d19bc35e","Type":"ContainerStarted","Data":"8ea390169168c7b47bb8cac431625afc823ef69ac15d874d12be3d0b2f197034"} Nov 22 08:16:48 crc kubenswrapper[4789]: I1122 08:16:48.540151 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-75d6db87b5-kbchw" event={"ID":"4be1163c-8900-4db9-a041-826526ae8751","Type":"ContainerStarted","Data":"60072d91c68904f56332cc39e6e1b69a937ce8755e065dd8a28e59f01469ede9"} Nov 22 08:16:49 crc kubenswrapper[4789]: I1122 08:16:49.017319 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-68cd7949f9-tqtjq" Nov 22 08:16:49 crc kubenswrapper[4789]: I1122 08:16:49.023074 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-lg59m" Nov 22 08:16:49 crc kubenswrapper[4789]: I1122 08:16:49.175566 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6bf8aa3e-a7e7-4f0c-a542-d668a5623707-combined-ca-bundle\") pod \"6bf8aa3e-a7e7-4f0c-a542-d668a5623707\" (UID: \"6bf8aa3e-a7e7-4f0c-a542-d668a5623707\") " Nov 22 08:16:49 crc kubenswrapper[4789]: I1122 08:16:49.176016 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6bf8aa3e-a7e7-4f0c-a542-d668a5623707-db-sync-config-data\") pod \"6bf8aa3e-a7e7-4f0c-a542-d668a5623707\" (UID: \"6bf8aa3e-a7e7-4f0c-a542-d668a5623707\") " Nov 22 08:16:49 crc kubenswrapper[4789]: I1122 08:16:49.176054 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6bf8aa3e-a7e7-4f0c-a542-d668a5623707-config-data\") pod \"6bf8aa3e-a7e7-4f0c-a542-d668a5623707\" (UID: \"6bf8aa3e-a7e7-4f0c-a542-d668a5623707\") " Nov 22 08:16:49 crc kubenswrapper[4789]: I1122 08:16:49.176094 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t9qz8\" (UniqueName: \"kubernetes.io/projected/6bf8aa3e-a7e7-4f0c-a542-d668a5623707-kube-api-access-t9qz8\") pod \"6bf8aa3e-a7e7-4f0c-a542-d668a5623707\" (UID: \"6bf8aa3e-a7e7-4f0c-a542-d668a5623707\") " Nov 22 08:16:49 crc kubenswrapper[4789]: I1122 08:16:49.176119 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2d8e424e-6fdf-4967-a24c-0a70af3724e3-ovsdbserver-nb\") pod \"2d8e424e-6fdf-4967-a24c-0a70af3724e3\" (UID: \"2d8e424e-6fdf-4967-a24c-0a70af3724e3\") " Nov 22 08:16:49 crc kubenswrapper[4789]: I1122 08:16:49.176831 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2d8e424e-6fdf-4967-a24c-0a70af3724e3-dns-svc\") pod \"2d8e424e-6fdf-4967-a24c-0a70af3724e3\" (UID: \"2d8e424e-6fdf-4967-a24c-0a70af3724e3\") " Nov 22 08:16:49 crc kubenswrapper[4789]: I1122 08:16:49.176907 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6bf8aa3e-a7e7-4f0c-a542-d668a5623707-scripts\") pod \"6bf8aa3e-a7e7-4f0c-a542-d668a5623707\" (UID: \"6bf8aa3e-a7e7-4f0c-a542-d668a5623707\") " Nov 22 08:16:49 crc kubenswrapper[4789]: I1122 08:16:49.176933 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d8e424e-6fdf-4967-a24c-0a70af3724e3-config\") pod \"2d8e424e-6fdf-4967-a24c-0a70af3724e3\" (UID: \"2d8e424e-6fdf-4967-a24c-0a70af3724e3\") " Nov 22 08:16:49 crc kubenswrapper[4789]: I1122 08:16:49.176989 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h6lm7\" (UniqueName: \"kubernetes.io/projected/2d8e424e-6fdf-4967-a24c-0a70af3724e3-kube-api-access-h6lm7\") pod \"2d8e424e-6fdf-4967-a24c-0a70af3724e3\" (UID: \"2d8e424e-6fdf-4967-a24c-0a70af3724e3\") " Nov 22 08:16:49 crc kubenswrapper[4789]: I1122 08:16:49.177035 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6bf8aa3e-a7e7-4f0c-a542-d668a5623707-etc-machine-id\") pod \"6bf8aa3e-a7e7-4f0c-a542-d668a5623707\" (UID: \"6bf8aa3e-a7e7-4f0c-a542-d668a5623707\") " Nov 22 08:16:49 crc kubenswrapper[4789]: I1122 08:16:49.177066 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2d8e424e-6fdf-4967-a24c-0a70af3724e3-ovsdbserver-sb\") pod \"2d8e424e-6fdf-4967-a24c-0a70af3724e3\" (UID: \"2d8e424e-6fdf-4967-a24c-0a70af3724e3\") " Nov 22 08:16:49 crc kubenswrapper[4789]: I1122 08:16:49.178003 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6bf8aa3e-a7e7-4f0c-a542-d668a5623707-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "6bf8aa3e-a7e7-4f0c-a542-d668a5623707" (UID: "6bf8aa3e-a7e7-4f0c-a542-d668a5623707"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 08:16:49 crc kubenswrapper[4789]: I1122 08:16:49.193482 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6bf8aa3e-a7e7-4f0c-a542-d668a5623707-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "6bf8aa3e-a7e7-4f0c-a542-d668a5623707" (UID: "6bf8aa3e-a7e7-4f0c-a542-d668a5623707"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:16:49 crc kubenswrapper[4789]: I1122 08:16:49.194371 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6bf8aa3e-a7e7-4f0c-a542-d668a5623707-kube-api-access-t9qz8" (OuterVolumeSpecName: "kube-api-access-t9qz8") pod "6bf8aa3e-a7e7-4f0c-a542-d668a5623707" (UID: "6bf8aa3e-a7e7-4f0c-a542-d668a5623707"). InnerVolumeSpecName "kube-api-access-t9qz8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:16:49 crc kubenswrapper[4789]: I1122 08:16:49.196547 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d8e424e-6fdf-4967-a24c-0a70af3724e3-kube-api-access-h6lm7" (OuterVolumeSpecName: "kube-api-access-h6lm7") pod "2d8e424e-6fdf-4967-a24c-0a70af3724e3" (UID: "2d8e424e-6fdf-4967-a24c-0a70af3724e3"). InnerVolumeSpecName "kube-api-access-h6lm7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:16:49 crc kubenswrapper[4789]: I1122 08:16:49.206064 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6bf8aa3e-a7e7-4f0c-a542-d668a5623707-scripts" (OuterVolumeSpecName: "scripts") pod "6bf8aa3e-a7e7-4f0c-a542-d668a5623707" (UID: "6bf8aa3e-a7e7-4f0c-a542-d668a5623707"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:16:49 crc kubenswrapper[4789]: I1122 08:16:49.246609 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6bf8aa3e-a7e7-4f0c-a542-d668a5623707-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6bf8aa3e-a7e7-4f0c-a542-d668a5623707" (UID: "6bf8aa3e-a7e7-4f0c-a542-d668a5623707"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:16:49 crc kubenswrapper[4789]: I1122 08:16:49.256813 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d8e424e-6fdf-4967-a24c-0a70af3724e3-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "2d8e424e-6fdf-4967-a24c-0a70af3724e3" (UID: "2d8e424e-6fdf-4967-a24c-0a70af3724e3"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:16:49 crc kubenswrapper[4789]: I1122 08:16:49.282365 4789 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6bf8aa3e-a7e7-4f0c-a542-d668a5623707-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:16:49 crc kubenswrapper[4789]: I1122 08:16:49.282405 4789 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6bf8aa3e-a7e7-4f0c-a542-d668a5623707-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:16:49 crc kubenswrapper[4789]: I1122 08:16:49.282420 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t9qz8\" (UniqueName: \"kubernetes.io/projected/6bf8aa3e-a7e7-4f0c-a542-d668a5623707-kube-api-access-t9qz8\") on node \"crc\" DevicePath \"\"" Nov 22 08:16:49 crc kubenswrapper[4789]: I1122 08:16:49.282432 4789 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2d8e424e-6fdf-4967-a24c-0a70af3724e3-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 22 08:16:49 crc kubenswrapper[4789]: I1122 08:16:49.282442 4789 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6bf8aa3e-a7e7-4f0c-a542-d668a5623707-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:16:49 crc kubenswrapper[4789]: I1122 08:16:49.282452 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h6lm7\" (UniqueName: \"kubernetes.io/projected/2d8e424e-6fdf-4967-a24c-0a70af3724e3-kube-api-access-h6lm7\") on node \"crc\" DevicePath \"\"" Nov 22 08:16:49 crc kubenswrapper[4789]: I1122 08:16:49.282464 4789 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6bf8aa3e-a7e7-4f0c-a542-d668a5623707-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 22 08:16:49 crc kubenswrapper[4789]: I1122 08:16:49.285470 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d8e424e-6fdf-4967-a24c-0a70af3724e3-config" (OuterVolumeSpecName: "config") pod "2d8e424e-6fdf-4967-a24c-0a70af3724e3" (UID: "2d8e424e-6fdf-4967-a24c-0a70af3724e3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:16:49 crc kubenswrapper[4789]: I1122 08:16:49.294550 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d8e424e-6fdf-4967-a24c-0a70af3724e3-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "2d8e424e-6fdf-4967-a24c-0a70af3724e3" (UID: "2d8e424e-6fdf-4967-a24c-0a70af3724e3"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:16:49 crc kubenswrapper[4789]: I1122 08:16:49.297888 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6bf8aa3e-a7e7-4f0c-a542-d668a5623707-config-data" (OuterVolumeSpecName: "config-data") pod "6bf8aa3e-a7e7-4f0c-a542-d668a5623707" (UID: "6bf8aa3e-a7e7-4f0c-a542-d668a5623707"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:16:49 crc kubenswrapper[4789]: I1122 08:16:49.301228 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d8e424e-6fdf-4967-a24c-0a70af3724e3-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "2d8e424e-6fdf-4967-a24c-0a70af3724e3" (UID: "2d8e424e-6fdf-4967-a24c-0a70af3724e3"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:16:49 crc kubenswrapper[4789]: I1122 08:16:49.385835 4789 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2d8e424e-6fdf-4967-a24c-0a70af3724e3-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 22 08:16:49 crc kubenswrapper[4789]: I1122 08:16:49.385879 4789 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d8e424e-6fdf-4967-a24c-0a70af3724e3-config\") on node \"crc\" DevicePath \"\"" Nov 22 08:16:49 crc kubenswrapper[4789]: I1122 08:16:49.385895 4789 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2d8e424e-6fdf-4967-a24c-0a70af3724e3-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 22 08:16:49 crc kubenswrapper[4789]: I1122 08:16:49.385909 4789 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6bf8aa3e-a7e7-4f0c-a542-d668a5623707-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:16:49 crc kubenswrapper[4789]: I1122 08:16:49.386384 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-568cd49f4d-ms9nh"] Nov 22 08:16:49 crc kubenswrapper[4789]: I1122 08:16:49.594542 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68cd7949f9-tqtjq" event={"ID":"2d8e424e-6fdf-4967-a24c-0a70af3724e3","Type":"ContainerDied","Data":"b4ffbb41d12659a706e25cde9d916c4388b3b507070551f1a3086a2530841672"} Nov 22 08:16:49 crc kubenswrapper[4789]: I1122 08:16:49.594607 4789 scope.go:117] "RemoveContainer" containerID="437b6c2ad5d376597aece00254567a75824d4490c507e5e042f2dd3a306bbf83" Nov 22 08:16:49 crc kubenswrapper[4789]: I1122 08:16:49.594725 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-68cd7949f9-tqtjq" Nov 22 08:16:49 crc kubenswrapper[4789]: I1122 08:16:49.599046 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Nov 22 08:16:49 crc kubenswrapper[4789]: E1122 08:16:49.599407 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d8e424e-6fdf-4967-a24c-0a70af3724e3" containerName="init" Nov 22 08:16:49 crc kubenswrapper[4789]: I1122 08:16:49.599419 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d8e424e-6fdf-4967-a24c-0a70af3724e3" containerName="init" Nov 22 08:16:49 crc kubenswrapper[4789]: E1122 08:16:49.599446 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6bf8aa3e-a7e7-4f0c-a542-d668a5623707" containerName="cinder-db-sync" Nov 22 08:16:49 crc kubenswrapper[4789]: I1122 08:16:49.599454 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="6bf8aa3e-a7e7-4f0c-a542-d668a5623707" containerName="cinder-db-sync" Nov 22 08:16:49 crc kubenswrapper[4789]: I1122 08:16:49.599674 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d8e424e-6fdf-4967-a24c-0a70af3724e3" containerName="init" Nov 22 08:16:49 crc kubenswrapper[4789]: I1122 08:16:49.599693 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="6bf8aa3e-a7e7-4f0c-a542-d668a5623707" containerName="cinder-db-sync" Nov 22 08:16:49 crc kubenswrapper[4789]: I1122 08:16:49.600616 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 22 08:16:49 crc kubenswrapper[4789]: I1122 08:16:49.607105 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-lg59m" event={"ID":"6bf8aa3e-a7e7-4f0c-a542-d668a5623707","Type":"ContainerDied","Data":"835d05ccf6aeb44f1550e11dc4e9f94affbfd2055349d047da78d8dda6b514d6"} Nov 22 08:16:49 crc kubenswrapper[4789]: I1122 08:16:49.607139 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="835d05ccf6aeb44f1550e11dc4e9f94affbfd2055349d047da78d8dda6b514d6" Nov 22 08:16:49 crc kubenswrapper[4789]: I1122 08:16:49.607181 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-lg59m" Nov 22 08:16:49 crc kubenswrapper[4789]: I1122 08:16:49.612148 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Nov 22 08:16:49 crc kubenswrapper[4789]: I1122 08:16:49.621936 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 22 08:16:49 crc kubenswrapper[4789]: I1122 08:16:49.627256 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6976474df8-jfq2h" event={"ID":"e382d956-58e8-4c3b-abd9-0732d19bc35e","Type":"ContainerStarted","Data":"753f281ef5ab8c90d5710f4a0e8721ac8c9ebc938a789a976d5a6573e0dbcc20"} Nov 22 08:16:49 crc kubenswrapper[4789]: I1122 08:16:49.628221 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6976474df8-jfq2h" Nov 22 08:16:49 crc kubenswrapper[4789]: I1122 08:16:49.628256 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6976474df8-jfq2h" Nov 22 08:16:49 crc kubenswrapper[4789]: I1122 08:16:49.698713 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e1b1a2ae-3c67-4ada-b478-92ee0a9bb30d-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"e1b1a2ae-3c67-4ada-b478-92ee0a9bb30d\") " pod="openstack/cinder-scheduler-0" Nov 22 08:16:49 crc kubenswrapper[4789]: I1122 08:16:49.700321 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-844b557b9c-lvrgl"] Nov 22 08:16:49 crc kubenswrapper[4789]: I1122 08:16:49.700397 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tqlfl\" (UniqueName: \"kubernetes.io/projected/e1b1a2ae-3c67-4ada-b478-92ee0a9bb30d-kube-api-access-tqlfl\") pod \"cinder-scheduler-0\" (UID: \"e1b1a2ae-3c67-4ada-b478-92ee0a9bb30d\") " pod="openstack/cinder-scheduler-0" Nov 22 08:16:49 crc kubenswrapper[4789]: I1122 08:16:49.700428 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1b1a2ae-3c67-4ada-b478-92ee0a9bb30d-config-data\") pod \"cinder-scheduler-0\" (UID: \"e1b1a2ae-3c67-4ada-b478-92ee0a9bb30d\") " pod="openstack/cinder-scheduler-0" Nov 22 08:16:49 crc kubenswrapper[4789]: I1122 08:16:49.700482 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e1b1a2ae-3c67-4ada-b478-92ee0a9bb30d-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"e1b1a2ae-3c67-4ada-b478-92ee0a9bb30d\") " pod="openstack/cinder-scheduler-0" Nov 22 08:16:49 crc kubenswrapper[4789]: I1122 08:16:49.700543 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1b1a2ae-3c67-4ada-b478-92ee0a9bb30d-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"e1b1a2ae-3c67-4ada-b478-92ee0a9bb30d\") " pod="openstack/cinder-scheduler-0" Nov 22 08:16:49 crc kubenswrapper[4789]: I1122 08:16:49.700642 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e1b1a2ae-3c67-4ada-b478-92ee0a9bb30d-scripts\") pod \"cinder-scheduler-0\" (UID: \"e1b1a2ae-3c67-4ada-b478-92ee0a9bb30d\") " pod="openstack/cinder-scheduler-0" Nov 22 08:16:49 crc kubenswrapper[4789]: I1122 08:16:49.722986 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-6976474df8-jfq2h" podStartSLOduration=3.7229605770000003 podStartE2EDuration="3.722960577s" podCreationTimestamp="2025-11-22 08:16:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:16:49.673522299 +0000 UTC m=+1223.907922602" watchObservedRunningTime="2025-11-22 08:16:49.722960577 +0000 UTC m=+1223.957360850" Nov 22 08:16:49 crc kubenswrapper[4789]: I1122 08:16:49.743131 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-775457b975-x8hvd"] Nov 22 08:16:49 crc kubenswrapper[4789]: I1122 08:16:49.744687 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-775457b975-x8hvd" Nov 22 08:16:49 crc kubenswrapper[4789]: I1122 08:16:49.778480 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-775457b975-x8hvd"] Nov 22 08:16:49 crc kubenswrapper[4789]: I1122 08:16:49.804464 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tqlfl\" (UniqueName: \"kubernetes.io/projected/e1b1a2ae-3c67-4ada-b478-92ee0a9bb30d-kube-api-access-tqlfl\") pod \"cinder-scheduler-0\" (UID: \"e1b1a2ae-3c67-4ada-b478-92ee0a9bb30d\") " pod="openstack/cinder-scheduler-0" Nov 22 08:16:49 crc kubenswrapper[4789]: I1122 08:16:49.804500 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e1b1a2ae-3c67-4ada-b478-92ee0a9bb30d-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"e1b1a2ae-3c67-4ada-b478-92ee0a9bb30d\") " pod="openstack/cinder-scheduler-0" Nov 22 08:16:49 crc kubenswrapper[4789]: I1122 08:16:49.804520 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1b1a2ae-3c67-4ada-b478-92ee0a9bb30d-config-data\") pod \"cinder-scheduler-0\" (UID: \"e1b1a2ae-3c67-4ada-b478-92ee0a9bb30d\") " pod="openstack/cinder-scheduler-0" Nov 22 08:16:49 crc kubenswrapper[4789]: I1122 08:16:49.804539 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e1b1a2ae-3c67-4ada-b478-92ee0a9bb30d-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"e1b1a2ae-3c67-4ada-b478-92ee0a9bb30d\") " pod="openstack/cinder-scheduler-0" Nov 22 08:16:49 crc kubenswrapper[4789]: I1122 08:16:49.804594 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1b1a2ae-3c67-4ada-b478-92ee0a9bb30d-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"e1b1a2ae-3c67-4ada-b478-92ee0a9bb30d\") " pod="openstack/cinder-scheduler-0" Nov 22 08:16:49 crc kubenswrapper[4789]: I1122 08:16:49.804680 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e1b1a2ae-3c67-4ada-b478-92ee0a9bb30d-scripts\") pod \"cinder-scheduler-0\" (UID: \"e1b1a2ae-3c67-4ada-b478-92ee0a9bb30d\") " pod="openstack/cinder-scheduler-0" Nov 22 08:16:49 crc kubenswrapper[4789]: I1122 08:16:49.805600 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e1b1a2ae-3c67-4ada-b478-92ee0a9bb30d-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"e1b1a2ae-3c67-4ada-b478-92ee0a9bb30d\") " pod="openstack/cinder-scheduler-0" Nov 22 08:16:49 crc kubenswrapper[4789]: I1122 08:16:49.838596 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tqlfl\" (UniqueName: \"kubernetes.io/projected/e1b1a2ae-3c67-4ada-b478-92ee0a9bb30d-kube-api-access-tqlfl\") pod \"cinder-scheduler-0\" (UID: \"e1b1a2ae-3c67-4ada-b478-92ee0a9bb30d\") " pod="openstack/cinder-scheduler-0" Nov 22 08:16:49 crc kubenswrapper[4789]: I1122 08:16:49.838680 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-68cd7949f9-tqtjq"] Nov 22 08:16:49 crc kubenswrapper[4789]: I1122 08:16:49.853410 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-68cd7949f9-tqtjq"] Nov 22 08:16:49 crc kubenswrapper[4789]: I1122 08:16:49.854638 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e1b1a2ae-3c67-4ada-b478-92ee0a9bb30d-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"e1b1a2ae-3c67-4ada-b478-92ee0a9bb30d\") " pod="openstack/cinder-scheduler-0" Nov 22 08:16:49 crc kubenswrapper[4789]: I1122 08:16:49.855693 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1b1a2ae-3c67-4ada-b478-92ee0a9bb30d-config-data\") pod \"cinder-scheduler-0\" (UID: \"e1b1a2ae-3c67-4ada-b478-92ee0a9bb30d\") " pod="openstack/cinder-scheduler-0" Nov 22 08:16:49 crc kubenswrapper[4789]: I1122 08:16:49.855995 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1b1a2ae-3c67-4ada-b478-92ee0a9bb30d-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"e1b1a2ae-3c67-4ada-b478-92ee0a9bb30d\") " pod="openstack/cinder-scheduler-0" Nov 22 08:16:49 crc kubenswrapper[4789]: I1122 08:16:49.861038 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e1b1a2ae-3c67-4ada-b478-92ee0a9bb30d-scripts\") pod \"cinder-scheduler-0\" (UID: \"e1b1a2ae-3c67-4ada-b478-92ee0a9bb30d\") " pod="openstack/cinder-scheduler-0" Nov 22 08:16:49 crc kubenswrapper[4789]: I1122 08:16:49.873158 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Nov 22 08:16:49 crc kubenswrapper[4789]: I1122 08:16:49.876360 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 22 08:16:49 crc kubenswrapper[4789]: I1122 08:16:49.879774 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 22 08:16:49 crc kubenswrapper[4789]: I1122 08:16:49.884601 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Nov 22 08:16:49 crc kubenswrapper[4789]: I1122 08:16:49.906837 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/63d8628c-ab9f-4c35-8550-08d647beffa4-ovsdbserver-sb\") pod \"dnsmasq-dns-775457b975-x8hvd\" (UID: \"63d8628c-ab9f-4c35-8550-08d647beffa4\") " pod="openstack/dnsmasq-dns-775457b975-x8hvd" Nov 22 08:16:49 crc kubenswrapper[4789]: I1122 08:16:49.906889 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/63d8628c-ab9f-4c35-8550-08d647beffa4-ovsdbserver-nb\") pod \"dnsmasq-dns-775457b975-x8hvd\" (UID: \"63d8628c-ab9f-4c35-8550-08d647beffa4\") " pod="openstack/dnsmasq-dns-775457b975-x8hvd" Nov 22 08:16:49 crc kubenswrapper[4789]: I1122 08:16:49.906913 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/63d8628c-ab9f-4c35-8550-08d647beffa4-config\") pod \"dnsmasq-dns-775457b975-x8hvd\" (UID: \"63d8628c-ab9f-4c35-8550-08d647beffa4\") " pod="openstack/dnsmasq-dns-775457b975-x8hvd" Nov 22 08:16:49 crc kubenswrapper[4789]: I1122 08:16:49.906939 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p49cm\" (UniqueName: \"kubernetes.io/projected/63d8628c-ab9f-4c35-8550-08d647beffa4-kube-api-access-p49cm\") pod \"dnsmasq-dns-775457b975-x8hvd\" (UID: \"63d8628c-ab9f-4c35-8550-08d647beffa4\") " pod="openstack/dnsmasq-dns-775457b975-x8hvd" Nov 22 08:16:49 crc kubenswrapper[4789]: I1122 08:16:49.906978 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/63d8628c-ab9f-4c35-8550-08d647beffa4-dns-svc\") pod \"dnsmasq-dns-775457b975-x8hvd\" (UID: \"63d8628c-ab9f-4c35-8550-08d647beffa4\") " pod="openstack/dnsmasq-dns-775457b975-x8hvd" Nov 22 08:16:49 crc kubenswrapper[4789]: I1122 08:16:49.965520 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 22 08:16:49 crc kubenswrapper[4789]: I1122 08:16:49.976885 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2d8e424e-6fdf-4967-a24c-0a70af3724e3" path="/var/lib/kubelet/pods/2d8e424e-6fdf-4967-a24c-0a70af3724e3/volumes" Nov 22 08:16:50 crc kubenswrapper[4789]: I1122 08:16:50.012211 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/63d8628c-ab9f-4c35-8550-08d647beffa4-dns-svc\") pod \"dnsmasq-dns-775457b975-x8hvd\" (UID: \"63d8628c-ab9f-4c35-8550-08d647beffa4\") " pod="openstack/dnsmasq-dns-775457b975-x8hvd" Nov 22 08:16:50 crc kubenswrapper[4789]: I1122 08:16:50.012267 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60e8e2c6-6de6-48c4-a004-a15eb39f4d84-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"60e8e2c6-6de6-48c4-a004-a15eb39f4d84\") " pod="openstack/cinder-api-0" Nov 22 08:16:50 crc kubenswrapper[4789]: I1122 08:16:50.012336 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/60e8e2c6-6de6-48c4-a004-a15eb39f4d84-etc-machine-id\") pod \"cinder-api-0\" (UID: \"60e8e2c6-6de6-48c4-a004-a15eb39f4d84\") " pod="openstack/cinder-api-0" Nov 22 08:16:50 crc kubenswrapper[4789]: I1122 08:16:50.012365 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-777ct\" (UniqueName: \"kubernetes.io/projected/60e8e2c6-6de6-48c4-a004-a15eb39f4d84-kube-api-access-777ct\") pod \"cinder-api-0\" (UID: \"60e8e2c6-6de6-48c4-a004-a15eb39f4d84\") " pod="openstack/cinder-api-0" Nov 22 08:16:50 crc kubenswrapper[4789]: I1122 08:16:50.012390 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/60e8e2c6-6de6-48c4-a004-a15eb39f4d84-logs\") pod \"cinder-api-0\" (UID: \"60e8e2c6-6de6-48c4-a004-a15eb39f4d84\") " pod="openstack/cinder-api-0" Nov 22 08:16:50 crc kubenswrapper[4789]: I1122 08:16:50.012449 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60e8e2c6-6de6-48c4-a004-a15eb39f4d84-config-data\") pod \"cinder-api-0\" (UID: \"60e8e2c6-6de6-48c4-a004-a15eb39f4d84\") " pod="openstack/cinder-api-0" Nov 22 08:16:50 crc kubenswrapper[4789]: I1122 08:16:50.012511 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/63d8628c-ab9f-4c35-8550-08d647beffa4-ovsdbserver-sb\") pod \"dnsmasq-dns-775457b975-x8hvd\" (UID: \"63d8628c-ab9f-4c35-8550-08d647beffa4\") " pod="openstack/dnsmasq-dns-775457b975-x8hvd" Nov 22 08:16:50 crc kubenswrapper[4789]: I1122 08:16:50.012553 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/63d8628c-ab9f-4c35-8550-08d647beffa4-ovsdbserver-nb\") pod \"dnsmasq-dns-775457b975-x8hvd\" (UID: \"63d8628c-ab9f-4c35-8550-08d647beffa4\") " pod="openstack/dnsmasq-dns-775457b975-x8hvd" Nov 22 08:16:50 crc kubenswrapper[4789]: I1122 08:16:50.012579 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/63d8628c-ab9f-4c35-8550-08d647beffa4-config\") pod \"dnsmasq-dns-775457b975-x8hvd\" (UID: \"63d8628c-ab9f-4c35-8550-08d647beffa4\") " pod="openstack/dnsmasq-dns-775457b975-x8hvd" Nov 22 08:16:50 crc kubenswrapper[4789]: I1122 08:16:50.012610 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p49cm\" (UniqueName: \"kubernetes.io/projected/63d8628c-ab9f-4c35-8550-08d647beffa4-kube-api-access-p49cm\") pod \"dnsmasq-dns-775457b975-x8hvd\" (UID: \"63d8628c-ab9f-4c35-8550-08d647beffa4\") " pod="openstack/dnsmasq-dns-775457b975-x8hvd" Nov 22 08:16:50 crc kubenswrapper[4789]: I1122 08:16:50.012633 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/60e8e2c6-6de6-48c4-a004-a15eb39f4d84-scripts\") pod \"cinder-api-0\" (UID: \"60e8e2c6-6de6-48c4-a004-a15eb39f4d84\") " pod="openstack/cinder-api-0" Nov 22 08:16:50 crc kubenswrapper[4789]: I1122 08:16:50.012672 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/60e8e2c6-6de6-48c4-a004-a15eb39f4d84-config-data-custom\") pod \"cinder-api-0\" (UID: \"60e8e2c6-6de6-48c4-a004-a15eb39f4d84\") " pod="openstack/cinder-api-0" Nov 22 08:16:50 crc kubenswrapper[4789]: I1122 08:16:50.013709 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/63d8628c-ab9f-4c35-8550-08d647beffa4-dns-svc\") pod \"dnsmasq-dns-775457b975-x8hvd\" (UID: \"63d8628c-ab9f-4c35-8550-08d647beffa4\") " pod="openstack/dnsmasq-dns-775457b975-x8hvd" Nov 22 08:16:50 crc kubenswrapper[4789]: I1122 08:16:50.013896 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/63d8628c-ab9f-4c35-8550-08d647beffa4-ovsdbserver-nb\") pod \"dnsmasq-dns-775457b975-x8hvd\" (UID: \"63d8628c-ab9f-4c35-8550-08d647beffa4\") " pod="openstack/dnsmasq-dns-775457b975-x8hvd" Nov 22 08:16:50 crc kubenswrapper[4789]: I1122 08:16:50.013936 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/63d8628c-ab9f-4c35-8550-08d647beffa4-ovsdbserver-sb\") pod \"dnsmasq-dns-775457b975-x8hvd\" (UID: \"63d8628c-ab9f-4c35-8550-08d647beffa4\") " pod="openstack/dnsmasq-dns-775457b975-x8hvd" Nov 22 08:16:50 crc kubenswrapper[4789]: I1122 08:16:50.014248 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/63d8628c-ab9f-4c35-8550-08d647beffa4-config\") pod \"dnsmasq-dns-775457b975-x8hvd\" (UID: \"63d8628c-ab9f-4c35-8550-08d647beffa4\") " pod="openstack/dnsmasq-dns-775457b975-x8hvd" Nov 22 08:16:50 crc kubenswrapper[4789]: I1122 08:16:50.036293 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p49cm\" (UniqueName: \"kubernetes.io/projected/63d8628c-ab9f-4c35-8550-08d647beffa4-kube-api-access-p49cm\") pod \"dnsmasq-dns-775457b975-x8hvd\" (UID: \"63d8628c-ab9f-4c35-8550-08d647beffa4\") " pod="openstack/dnsmasq-dns-775457b975-x8hvd" Nov 22 08:16:50 crc kubenswrapper[4789]: I1122 08:16:50.076460 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-775457b975-x8hvd" Nov 22 08:16:50 crc kubenswrapper[4789]: I1122 08:16:50.115199 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/60e8e2c6-6de6-48c4-a004-a15eb39f4d84-etc-machine-id\") pod \"cinder-api-0\" (UID: \"60e8e2c6-6de6-48c4-a004-a15eb39f4d84\") " pod="openstack/cinder-api-0" Nov 22 08:16:50 crc kubenswrapper[4789]: I1122 08:16:50.115678 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-777ct\" (UniqueName: \"kubernetes.io/projected/60e8e2c6-6de6-48c4-a004-a15eb39f4d84-kube-api-access-777ct\") pod \"cinder-api-0\" (UID: \"60e8e2c6-6de6-48c4-a004-a15eb39f4d84\") " pod="openstack/cinder-api-0" Nov 22 08:16:50 crc kubenswrapper[4789]: I1122 08:16:50.115807 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/60e8e2c6-6de6-48c4-a004-a15eb39f4d84-logs\") pod \"cinder-api-0\" (UID: \"60e8e2c6-6de6-48c4-a004-a15eb39f4d84\") " pod="openstack/cinder-api-0" Nov 22 08:16:50 crc kubenswrapper[4789]: I1122 08:16:50.115928 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60e8e2c6-6de6-48c4-a004-a15eb39f4d84-config-data\") pod \"cinder-api-0\" (UID: \"60e8e2c6-6de6-48c4-a004-a15eb39f4d84\") " pod="openstack/cinder-api-0" Nov 22 08:16:50 crc kubenswrapper[4789]: I1122 08:16:50.115386 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/60e8e2c6-6de6-48c4-a004-a15eb39f4d84-etc-machine-id\") pod \"cinder-api-0\" (UID: \"60e8e2c6-6de6-48c4-a004-a15eb39f4d84\") " pod="openstack/cinder-api-0" Nov 22 08:16:50 crc kubenswrapper[4789]: I1122 08:16:50.116222 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/60e8e2c6-6de6-48c4-a004-a15eb39f4d84-scripts\") pod \"cinder-api-0\" (UID: \"60e8e2c6-6de6-48c4-a004-a15eb39f4d84\") " pod="openstack/cinder-api-0" Nov 22 08:16:50 crc kubenswrapper[4789]: I1122 08:16:50.116321 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/60e8e2c6-6de6-48c4-a004-a15eb39f4d84-config-data-custom\") pod \"cinder-api-0\" (UID: \"60e8e2c6-6de6-48c4-a004-a15eb39f4d84\") " pod="openstack/cinder-api-0" Nov 22 08:16:50 crc kubenswrapper[4789]: I1122 08:16:50.116441 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60e8e2c6-6de6-48c4-a004-a15eb39f4d84-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"60e8e2c6-6de6-48c4-a004-a15eb39f4d84\") " pod="openstack/cinder-api-0" Nov 22 08:16:50 crc kubenswrapper[4789]: I1122 08:16:50.116680 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/60e8e2c6-6de6-48c4-a004-a15eb39f4d84-logs\") pod \"cinder-api-0\" (UID: \"60e8e2c6-6de6-48c4-a004-a15eb39f4d84\") " pod="openstack/cinder-api-0" Nov 22 08:16:50 crc kubenswrapper[4789]: W1122 08:16:50.117333 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf3130d2e_6671_4a4b_978b_eac10b62cdc8.slice/crio-4e594f7116255277649b2f47d7fffe951d7791097ca30e8eb7a91043e41420c5 WatchSource:0}: Error finding container 4e594f7116255277649b2f47d7fffe951d7791097ca30e8eb7a91043e41420c5: Status 404 returned error can't find the container with id 4e594f7116255277649b2f47d7fffe951d7791097ca30e8eb7a91043e41420c5 Nov 22 08:16:50 crc kubenswrapper[4789]: I1122 08:16:50.122611 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60e8e2c6-6de6-48c4-a004-a15eb39f4d84-config-data\") pod \"cinder-api-0\" (UID: \"60e8e2c6-6de6-48c4-a004-a15eb39f4d84\") " pod="openstack/cinder-api-0" Nov 22 08:16:50 crc kubenswrapper[4789]: I1122 08:16:50.122909 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/60e8e2c6-6de6-48c4-a004-a15eb39f4d84-config-data-custom\") pod \"cinder-api-0\" (UID: \"60e8e2c6-6de6-48c4-a004-a15eb39f4d84\") " pod="openstack/cinder-api-0" Nov 22 08:16:50 crc kubenswrapper[4789]: I1122 08:16:50.124373 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60e8e2c6-6de6-48c4-a004-a15eb39f4d84-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"60e8e2c6-6de6-48c4-a004-a15eb39f4d84\") " pod="openstack/cinder-api-0" Nov 22 08:16:50 crc kubenswrapper[4789]: I1122 08:16:50.131920 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/60e8e2c6-6de6-48c4-a004-a15eb39f4d84-scripts\") pod \"cinder-api-0\" (UID: \"60e8e2c6-6de6-48c4-a004-a15eb39f4d84\") " pod="openstack/cinder-api-0" Nov 22 08:16:50 crc kubenswrapper[4789]: I1122 08:16:50.138140 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-777ct\" (UniqueName: \"kubernetes.io/projected/60e8e2c6-6de6-48c4-a004-a15eb39f4d84-kube-api-access-777ct\") pod \"cinder-api-0\" (UID: \"60e8e2c6-6de6-48c4-a004-a15eb39f4d84\") " pod="openstack/cinder-api-0" Nov 22 08:16:50 crc kubenswrapper[4789]: I1122 08:16:50.196717 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 22 08:16:50 crc kubenswrapper[4789]: I1122 08:16:50.564283 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 22 08:16:50 crc kubenswrapper[4789]: I1122 08:16:50.644471 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-844b557b9c-lvrgl" event={"ID":"f3130d2e-6671-4a4b-978b-eac10b62cdc8","Type":"ContainerStarted","Data":"4e594f7116255277649b2f47d7fffe951d7791097ca30e8eb7a91043e41420c5"} Nov 22 08:16:50 crc kubenswrapper[4789]: I1122 08:16:50.648549 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-568cd49f4d-ms9nh" event={"ID":"2949ebe1-7541-42a2-894c-2588c2bd3396","Type":"ContainerStarted","Data":"fde8262fa183d29bdeda97160e0de154acc0b30e8d06355d1c99fb047c5f1ba0"} Nov 22 08:16:50 crc kubenswrapper[4789]: I1122 08:16:50.654503 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-775457b975-x8hvd"] Nov 22 08:16:50 crc kubenswrapper[4789]: I1122 08:16:50.655409 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e1b1a2ae-3c67-4ada-b478-92ee0a9bb30d","Type":"ContainerStarted","Data":"7c37e3313ff0e05a96d71fc2869b8fa55eb2be4d7d3c1d5459a19d0aa134398e"} Nov 22 08:16:50 crc kubenswrapper[4789]: W1122 08:16:50.666976 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod63d8628c_ab9f_4c35_8550_08d647beffa4.slice/crio-4e97eb89335c67cd4648cf09d9cf3029ec59539eee8a712ea386948b010051d1 WatchSource:0}: Error finding container 4e97eb89335c67cd4648cf09d9cf3029ec59539eee8a712ea386948b010051d1: Status 404 returned error can't find the container with id 4e97eb89335c67cd4648cf09d9cf3029ec59539eee8a712ea386948b010051d1 Nov 22 08:16:51 crc kubenswrapper[4789]: I1122 08:16:51.275600 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-5b56df9788-wsmhq" Nov 22 08:16:51 crc kubenswrapper[4789]: I1122 08:16:51.404996 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 22 08:16:51 crc kubenswrapper[4789]: W1122 08:16:51.560882 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod60e8e2c6_6de6_48c4_a004_a15eb39f4d84.slice/crio-d5cef0478dc4534ff63796e77ead5a16b224a869d6bf8f9221f03be991f6ce9d WatchSource:0}: Error finding container d5cef0478dc4534ff63796e77ead5a16b224a869d6bf8f9221f03be991f6ce9d: Status 404 returned error can't find the container with id d5cef0478dc4534ff63796e77ead5a16b224a869d6bf8f9221f03be991f6ce9d Nov 22 08:16:51 crc kubenswrapper[4789]: I1122 08:16:51.584977 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-7b778f8c8-6rl4f" Nov 22 08:16:51 crc kubenswrapper[4789]: I1122 08:16:51.666966 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-775457b975-x8hvd" event={"ID":"63d8628c-ab9f-4c35-8550-08d647beffa4","Type":"ContainerStarted","Data":"4e97eb89335c67cd4648cf09d9cf3029ec59539eee8a712ea386948b010051d1"} Nov 22 08:16:51 crc kubenswrapper[4789]: I1122 08:16:51.669044 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-844b557b9c-lvrgl" event={"ID":"f3130d2e-6671-4a4b-978b-eac10b62cdc8","Type":"ContainerStarted","Data":"7880bf06c4e8110d8b726cf2edb81c414055cf50551ba7e85ff7962a65a61cc7"} Nov 22 08:16:51 crc kubenswrapper[4789]: I1122 08:16:51.670784 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"60e8e2c6-6de6-48c4-a004-a15eb39f4d84","Type":"ContainerStarted","Data":"d5cef0478dc4534ff63796e77ead5a16b224a869d6bf8f9221f03be991f6ce9d"} Nov 22 08:16:52 crc kubenswrapper[4789]: I1122 08:16:52.633558 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 08:16:52 crc kubenswrapper[4789]: I1122 08:16:52.696372 4789 generic.go:334] "Generic (PLEG): container finished" podID="63d8628c-ab9f-4c35-8550-08d647beffa4" containerID="8d5b43af47dc8c714c823ec10bfe8850e47ceb36a28d5618349498082600cfa2" exitCode=0 Nov 22 08:16:52 crc kubenswrapper[4789]: I1122 08:16:52.697815 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-775457b975-x8hvd" event={"ID":"63d8628c-ab9f-4c35-8550-08d647beffa4","Type":"ContainerDied","Data":"8d5b43af47dc8c714c823ec10bfe8850e47ceb36a28d5618349498082600cfa2"} Nov 22 08:16:52 crc kubenswrapper[4789]: I1122 08:16:52.703170 4789 generic.go:334] "Generic (PLEG): container finished" podID="f3130d2e-6671-4a4b-978b-eac10b62cdc8" containerID="7880bf06c4e8110d8b726cf2edb81c414055cf50551ba7e85ff7962a65a61cc7" exitCode=0 Nov 22 08:16:52 crc kubenswrapper[4789]: I1122 08:16:52.703448 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-844b557b9c-lvrgl" event={"ID":"f3130d2e-6671-4a4b-978b-eac10b62cdc8","Type":"ContainerDied","Data":"7880bf06c4e8110d8b726cf2edb81c414055cf50551ba7e85ff7962a65a61cc7"} Nov 22 08:16:52 crc kubenswrapper[4789]: I1122 08:16:52.711887 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7ccbbd9f98-jd5vz" event={"ID":"affd3f11-c0d4-41a2-888b-f4ae0c7e3d41","Type":"ContainerStarted","Data":"060b7689887c0141a77d903382401f5c544a019810a5de27d50084f7b80957e1"} Nov 22 08:16:52 crc kubenswrapper[4789]: I1122 08:16:52.711929 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7ccbbd9f98-jd5vz" event={"ID":"affd3f11-c0d4-41a2-888b-f4ae0c7e3d41","Type":"ContainerStarted","Data":"893a8d6379161f1b995c3c8f0508f2ad01ad50ee319ade6b6296c929ef3162ba"} Nov 22 08:16:52 crc kubenswrapper[4789]: I1122 08:16:52.727582 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-568cd49f4d-ms9nh" event={"ID":"2949ebe1-7541-42a2-894c-2588c2bd3396","Type":"ContainerStarted","Data":"9aec8c709e300ab397822f5fff34aba93e0915d5c8707aac0584b703bc60df1a"} Nov 22 08:16:52 crc kubenswrapper[4789]: I1122 08:16:52.727636 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-568cd49f4d-ms9nh" event={"ID":"2949ebe1-7541-42a2-894c-2588c2bd3396","Type":"ContainerStarted","Data":"77b48a1183833a07016416adb6c59b89b3874df9589e6f5b1c3dba3285e81de8"} Nov 22 08:16:52 crc kubenswrapper[4789]: I1122 08:16:52.727699 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-568cd49f4d-ms9nh" Nov 22 08:16:52 crc kubenswrapper[4789]: I1122 08:16:52.748702 4789 generic.go:334] "Generic (PLEG): container finished" podID="94d7bb01-64a7-442a-aee4-e958d8511b76" containerID="b8f37fba2551f2fc622fc288b287229e43574f1d9231f7967747f9645db2d0e9" exitCode=0 Nov 22 08:16:52 crc kubenswrapper[4789]: I1122 08:16:52.749057 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"94d7bb01-64a7-442a-aee4-e958d8511b76","Type":"ContainerDied","Data":"b8f37fba2551f2fc622fc288b287229e43574f1d9231f7967747f9645db2d0e9"} Nov 22 08:16:52 crc kubenswrapper[4789]: I1122 08:16:52.749215 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"94d7bb01-64a7-442a-aee4-e958d8511b76","Type":"ContainerDied","Data":"414dc84d92b52f295bade29141d15ec375784cdc75fb3cbb8ce807e50fe81343"} Nov 22 08:16:52 crc kubenswrapper[4789]: I1122 08:16:52.749256 4789 scope.go:117] "RemoveContainer" containerID="3381e3d40939b89f2f49a7f7a930e51358fcd14142eac9225f51202ec1a339a4" Nov 22 08:16:52 crc kubenswrapper[4789]: I1122 08:16:52.749992 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 08:16:52 crc kubenswrapper[4789]: I1122 08:16:52.764871 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/94d7bb01-64a7-442a-aee4-e958d8511b76-sg-core-conf-yaml\") pod \"94d7bb01-64a7-442a-aee4-e958d8511b76\" (UID: \"94d7bb01-64a7-442a-aee4-e958d8511b76\") " Nov 22 08:16:52 crc kubenswrapper[4789]: I1122 08:16:52.764956 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/94d7bb01-64a7-442a-aee4-e958d8511b76-scripts\") pod \"94d7bb01-64a7-442a-aee4-e958d8511b76\" (UID: \"94d7bb01-64a7-442a-aee4-e958d8511b76\") " Nov 22 08:16:52 crc kubenswrapper[4789]: I1122 08:16:52.765964 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94d7bb01-64a7-442a-aee4-e958d8511b76-combined-ca-bundle\") pod \"94d7bb01-64a7-442a-aee4-e958d8511b76\" (UID: \"94d7bb01-64a7-442a-aee4-e958d8511b76\") " Nov 22 08:16:52 crc kubenswrapper[4789]: I1122 08:16:52.766052 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/94d7bb01-64a7-442a-aee4-e958d8511b76-log-httpd\") pod \"94d7bb01-64a7-442a-aee4-e958d8511b76\" (UID: \"94d7bb01-64a7-442a-aee4-e958d8511b76\") " Nov 22 08:16:52 crc kubenswrapper[4789]: I1122 08:16:52.766111 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94d7bb01-64a7-442a-aee4-e958d8511b76-config-data\") pod \"94d7bb01-64a7-442a-aee4-e958d8511b76\" (UID: \"94d7bb01-64a7-442a-aee4-e958d8511b76\") " Nov 22 08:16:52 crc kubenswrapper[4789]: I1122 08:16:52.766187 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rxhrx\" (UniqueName: \"kubernetes.io/projected/94d7bb01-64a7-442a-aee4-e958d8511b76-kube-api-access-rxhrx\") pod \"94d7bb01-64a7-442a-aee4-e958d8511b76\" (UID: \"94d7bb01-64a7-442a-aee4-e958d8511b76\") " Nov 22 08:16:52 crc kubenswrapper[4789]: I1122 08:16:52.766234 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/94d7bb01-64a7-442a-aee4-e958d8511b76-run-httpd\") pod \"94d7bb01-64a7-442a-aee4-e958d8511b76\" (UID: \"94d7bb01-64a7-442a-aee4-e958d8511b76\") " Nov 22 08:16:52 crc kubenswrapper[4789]: I1122 08:16:52.771625 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/94d7bb01-64a7-442a-aee4-e958d8511b76-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "94d7bb01-64a7-442a-aee4-e958d8511b76" (UID: "94d7bb01-64a7-442a-aee4-e958d8511b76"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:16:52 crc kubenswrapper[4789]: I1122 08:16:52.773036 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/94d7bb01-64a7-442a-aee4-e958d8511b76-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "94d7bb01-64a7-442a-aee4-e958d8511b76" (UID: "94d7bb01-64a7-442a-aee4-e958d8511b76"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:16:52 crc kubenswrapper[4789]: I1122 08:16:52.775665 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-75d6db87b5-kbchw" event={"ID":"4be1163c-8900-4db9-a041-826526ae8751","Type":"ContainerStarted","Data":"17ba8642c853d3d8a0ca8d5623af64aa2442b875e8027b88f3f535d5a4dd6b9a"} Nov 22 08:16:52 crc kubenswrapper[4789]: I1122 08:16:52.775724 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-75d6db87b5-kbchw" event={"ID":"4be1163c-8900-4db9-a041-826526ae8751","Type":"ContainerStarted","Data":"ff913d5f5069ef18f843931f940fa8280eedf1469e9d3dfa1c16334f4e9469d0"} Nov 22 08:16:52 crc kubenswrapper[4789]: I1122 08:16:52.783805 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94d7bb01-64a7-442a-aee4-e958d8511b76-scripts" (OuterVolumeSpecName: "scripts") pod "94d7bb01-64a7-442a-aee4-e958d8511b76" (UID: "94d7bb01-64a7-442a-aee4-e958d8511b76"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:16:52 crc kubenswrapper[4789]: I1122 08:16:52.784532 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-568cd49f4d-ms9nh" podStartSLOduration=5.784513074 podStartE2EDuration="5.784513074s" podCreationTimestamp="2025-11-22 08:16:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:16:52.776189676 +0000 UTC m=+1227.010589959" watchObservedRunningTime="2025-11-22 08:16:52.784513074 +0000 UTC m=+1227.018913347" Nov 22 08:16:52 crc kubenswrapper[4789]: I1122 08:16:52.790272 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94d7bb01-64a7-442a-aee4-e958d8511b76-kube-api-access-rxhrx" (OuterVolumeSpecName: "kube-api-access-rxhrx") pod "94d7bb01-64a7-442a-aee4-e958d8511b76" (UID: "94d7bb01-64a7-442a-aee4-e958d8511b76"). InnerVolumeSpecName "kube-api-access-rxhrx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:16:52 crc kubenswrapper[4789]: I1122 08:16:52.798894 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-75d6db87b5-kbchw" podStartSLOduration=3.989787366 podStartE2EDuration="6.798871466s" podCreationTimestamp="2025-11-22 08:16:46 +0000 UTC" firstStartedPulling="2025-11-22 08:16:47.521177711 +0000 UTC m=+1221.755577984" lastFinishedPulling="2025-11-22 08:16:50.330261811 +0000 UTC m=+1224.564662084" observedRunningTime="2025-11-22 08:16:52.796697307 +0000 UTC m=+1227.031097590" watchObservedRunningTime="2025-11-22 08:16:52.798871466 +0000 UTC m=+1227.033271739" Nov 22 08:16:52 crc kubenswrapper[4789]: I1122 08:16:52.823860 4789 scope.go:117] "RemoveContainer" containerID="8463b06d53f122d2dd4bda6d1f5b32b440b64a9e858e15291e72b4e4f847355f" Nov 22 08:16:52 crc kubenswrapper[4789]: I1122 08:16:52.868287 4789 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/94d7bb01-64a7-442a-aee4-e958d8511b76-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 22 08:16:52 crc kubenswrapper[4789]: I1122 08:16:52.868312 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rxhrx\" (UniqueName: \"kubernetes.io/projected/94d7bb01-64a7-442a-aee4-e958d8511b76-kube-api-access-rxhrx\") on node \"crc\" DevicePath \"\"" Nov 22 08:16:52 crc kubenswrapper[4789]: I1122 08:16:52.868322 4789 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/94d7bb01-64a7-442a-aee4-e958d8511b76-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 22 08:16:52 crc kubenswrapper[4789]: I1122 08:16:52.868332 4789 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/94d7bb01-64a7-442a-aee4-e958d8511b76-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:16:52 crc kubenswrapper[4789]: I1122 08:16:52.871467 4789 scope.go:117] "RemoveContainer" containerID="b8f37fba2551f2fc622fc288b287229e43574f1d9231f7967747f9645db2d0e9" Nov 22 08:16:52 crc kubenswrapper[4789]: I1122 08:16:52.929967 4789 scope.go:117] "RemoveContainer" containerID="5c1bd328b27a7fd1cdca6ff567666e218a70a3b926831af09bf3a542194a3a1b" Nov 22 08:16:52 crc kubenswrapper[4789]: I1122 08:16:52.936373 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94d7bb01-64a7-442a-aee4-e958d8511b76-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "94d7bb01-64a7-442a-aee4-e958d8511b76" (UID: "94d7bb01-64a7-442a-aee4-e958d8511b76"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:16:52 crc kubenswrapper[4789]: I1122 08:16:52.966487 4789 scope.go:117] "RemoveContainer" containerID="3381e3d40939b89f2f49a7f7a930e51358fcd14142eac9225f51202ec1a339a4" Nov 22 08:16:52 crc kubenswrapper[4789]: E1122 08:16:52.967241 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3381e3d40939b89f2f49a7f7a930e51358fcd14142eac9225f51202ec1a339a4\": container with ID starting with 3381e3d40939b89f2f49a7f7a930e51358fcd14142eac9225f51202ec1a339a4 not found: ID does not exist" containerID="3381e3d40939b89f2f49a7f7a930e51358fcd14142eac9225f51202ec1a339a4" Nov 22 08:16:52 crc kubenswrapper[4789]: I1122 08:16:52.967339 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3381e3d40939b89f2f49a7f7a930e51358fcd14142eac9225f51202ec1a339a4"} err="failed to get container status \"3381e3d40939b89f2f49a7f7a930e51358fcd14142eac9225f51202ec1a339a4\": rpc error: code = NotFound desc = could not find container \"3381e3d40939b89f2f49a7f7a930e51358fcd14142eac9225f51202ec1a339a4\": container with ID starting with 3381e3d40939b89f2f49a7f7a930e51358fcd14142eac9225f51202ec1a339a4 not found: ID does not exist" Nov 22 08:16:52 crc kubenswrapper[4789]: I1122 08:16:52.967412 4789 scope.go:117] "RemoveContainer" containerID="8463b06d53f122d2dd4bda6d1f5b32b440b64a9e858e15291e72b4e4f847355f" Nov 22 08:16:52 crc kubenswrapper[4789]: I1122 08:16:52.971381 4789 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/94d7bb01-64a7-442a-aee4-e958d8511b76-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 22 08:16:52 crc kubenswrapper[4789]: E1122 08:16:52.971882 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8463b06d53f122d2dd4bda6d1f5b32b440b64a9e858e15291e72b4e4f847355f\": container with ID starting with 8463b06d53f122d2dd4bda6d1f5b32b440b64a9e858e15291e72b4e4f847355f not found: ID does not exist" containerID="8463b06d53f122d2dd4bda6d1f5b32b440b64a9e858e15291e72b4e4f847355f" Nov 22 08:16:52 crc kubenswrapper[4789]: I1122 08:16:52.972345 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8463b06d53f122d2dd4bda6d1f5b32b440b64a9e858e15291e72b4e4f847355f"} err="failed to get container status \"8463b06d53f122d2dd4bda6d1f5b32b440b64a9e858e15291e72b4e4f847355f\": rpc error: code = NotFound desc = could not find container \"8463b06d53f122d2dd4bda6d1f5b32b440b64a9e858e15291e72b4e4f847355f\": container with ID starting with 8463b06d53f122d2dd4bda6d1f5b32b440b64a9e858e15291e72b4e4f847355f not found: ID does not exist" Nov 22 08:16:52 crc kubenswrapper[4789]: I1122 08:16:52.972440 4789 scope.go:117] "RemoveContainer" containerID="b8f37fba2551f2fc622fc288b287229e43574f1d9231f7967747f9645db2d0e9" Nov 22 08:16:52 crc kubenswrapper[4789]: E1122 08:16:52.973089 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b8f37fba2551f2fc622fc288b287229e43574f1d9231f7967747f9645db2d0e9\": container with ID starting with b8f37fba2551f2fc622fc288b287229e43574f1d9231f7967747f9645db2d0e9 not found: ID does not exist" containerID="b8f37fba2551f2fc622fc288b287229e43574f1d9231f7967747f9645db2d0e9" Nov 22 08:16:52 crc kubenswrapper[4789]: I1122 08:16:52.979027 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b8f37fba2551f2fc622fc288b287229e43574f1d9231f7967747f9645db2d0e9"} err="failed to get container status \"b8f37fba2551f2fc622fc288b287229e43574f1d9231f7967747f9645db2d0e9\": rpc error: code = NotFound desc = could not find container \"b8f37fba2551f2fc622fc288b287229e43574f1d9231f7967747f9645db2d0e9\": container with ID starting with b8f37fba2551f2fc622fc288b287229e43574f1d9231f7967747f9645db2d0e9 not found: ID does not exist" Nov 22 08:16:52 crc kubenswrapper[4789]: I1122 08:16:52.979087 4789 scope.go:117] "RemoveContainer" containerID="5c1bd328b27a7fd1cdca6ff567666e218a70a3b926831af09bf3a542194a3a1b" Nov 22 08:16:52 crc kubenswrapper[4789]: E1122 08:16:52.979889 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5c1bd328b27a7fd1cdca6ff567666e218a70a3b926831af09bf3a542194a3a1b\": container with ID starting with 5c1bd328b27a7fd1cdca6ff567666e218a70a3b926831af09bf3a542194a3a1b not found: ID does not exist" containerID="5c1bd328b27a7fd1cdca6ff567666e218a70a3b926831af09bf3a542194a3a1b" Nov 22 08:16:52 crc kubenswrapper[4789]: I1122 08:16:52.979926 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c1bd328b27a7fd1cdca6ff567666e218a70a3b926831af09bf3a542194a3a1b"} err="failed to get container status \"5c1bd328b27a7fd1cdca6ff567666e218a70a3b926831af09bf3a542194a3a1b\": rpc error: code = NotFound desc = could not find container \"5c1bd328b27a7fd1cdca6ff567666e218a70a3b926831af09bf3a542194a3a1b\": container with ID starting with 5c1bd328b27a7fd1cdca6ff567666e218a70a3b926831af09bf3a542194a3a1b not found: ID does not exist" Nov 22 08:16:53 crc kubenswrapper[4789]: I1122 08:16:53.051500 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-844b557b9c-lvrgl" Nov 22 08:16:53 crc kubenswrapper[4789]: I1122 08:16:53.127550 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94d7bb01-64a7-442a-aee4-e958d8511b76-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "94d7bb01-64a7-442a-aee4-e958d8511b76" (UID: "94d7bb01-64a7-442a-aee4-e958d8511b76"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:16:53 crc kubenswrapper[4789]: I1122 08:16:53.177744 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f3130d2e-6671-4a4b-978b-eac10b62cdc8-config\") pod \"f3130d2e-6671-4a4b-978b-eac10b62cdc8\" (UID: \"f3130d2e-6671-4a4b-978b-eac10b62cdc8\") " Nov 22 08:16:53 crc kubenswrapper[4789]: I1122 08:16:53.177885 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f3130d2e-6671-4a4b-978b-eac10b62cdc8-ovsdbserver-nb\") pod \"f3130d2e-6671-4a4b-978b-eac10b62cdc8\" (UID: \"f3130d2e-6671-4a4b-978b-eac10b62cdc8\") " Nov 22 08:16:53 crc kubenswrapper[4789]: I1122 08:16:53.177950 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f3130d2e-6671-4a4b-978b-eac10b62cdc8-ovsdbserver-sb\") pod \"f3130d2e-6671-4a4b-978b-eac10b62cdc8\" (UID: \"f3130d2e-6671-4a4b-978b-eac10b62cdc8\") " Nov 22 08:16:53 crc kubenswrapper[4789]: I1122 08:16:53.178105 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lk4xm\" (UniqueName: \"kubernetes.io/projected/f3130d2e-6671-4a4b-978b-eac10b62cdc8-kube-api-access-lk4xm\") pod \"f3130d2e-6671-4a4b-978b-eac10b62cdc8\" (UID: \"f3130d2e-6671-4a4b-978b-eac10b62cdc8\") " Nov 22 08:16:53 crc kubenswrapper[4789]: I1122 08:16:53.178181 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f3130d2e-6671-4a4b-978b-eac10b62cdc8-dns-svc\") pod \"f3130d2e-6671-4a4b-978b-eac10b62cdc8\" (UID: \"f3130d2e-6671-4a4b-978b-eac10b62cdc8\") " Nov 22 08:16:53 crc kubenswrapper[4789]: I1122 08:16:53.178611 4789 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94d7bb01-64a7-442a-aee4-e958d8511b76-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:16:53 crc kubenswrapper[4789]: I1122 08:16:53.195893 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f3130d2e-6671-4a4b-978b-eac10b62cdc8-kube-api-access-lk4xm" (OuterVolumeSpecName: "kube-api-access-lk4xm") pod "f3130d2e-6671-4a4b-978b-eac10b62cdc8" (UID: "f3130d2e-6671-4a4b-978b-eac10b62cdc8"). InnerVolumeSpecName "kube-api-access-lk4xm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:16:53 crc kubenswrapper[4789]: I1122 08:16:53.199472 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94d7bb01-64a7-442a-aee4-e958d8511b76-config-data" (OuterVolumeSpecName: "config-data") pod "94d7bb01-64a7-442a-aee4-e958d8511b76" (UID: "94d7bb01-64a7-442a-aee4-e958d8511b76"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:16:53 crc kubenswrapper[4789]: I1122 08:16:53.221899 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f3130d2e-6671-4a4b-978b-eac10b62cdc8-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f3130d2e-6671-4a4b-978b-eac10b62cdc8" (UID: "f3130d2e-6671-4a4b-978b-eac10b62cdc8"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:16:53 crc kubenswrapper[4789]: I1122 08:16:53.223038 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f3130d2e-6671-4a4b-978b-eac10b62cdc8-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "f3130d2e-6671-4a4b-978b-eac10b62cdc8" (UID: "f3130d2e-6671-4a4b-978b-eac10b62cdc8"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:16:53 crc kubenswrapper[4789]: I1122 08:16:53.227214 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f3130d2e-6671-4a4b-978b-eac10b62cdc8-config" (OuterVolumeSpecName: "config") pod "f3130d2e-6671-4a4b-978b-eac10b62cdc8" (UID: "f3130d2e-6671-4a4b-978b-eac10b62cdc8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:16:53 crc kubenswrapper[4789]: I1122 08:16:53.253794 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f3130d2e-6671-4a4b-978b-eac10b62cdc8-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "f3130d2e-6671-4a4b-978b-eac10b62cdc8" (UID: "f3130d2e-6671-4a4b-978b-eac10b62cdc8"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:16:53 crc kubenswrapper[4789]: I1122 08:16:53.280433 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lk4xm\" (UniqueName: \"kubernetes.io/projected/f3130d2e-6671-4a4b-978b-eac10b62cdc8-kube-api-access-lk4xm\") on node \"crc\" DevicePath \"\"" Nov 22 08:16:53 crc kubenswrapper[4789]: I1122 08:16:53.280473 4789 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f3130d2e-6671-4a4b-978b-eac10b62cdc8-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 22 08:16:53 crc kubenswrapper[4789]: I1122 08:16:53.280484 4789 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94d7bb01-64a7-442a-aee4-e958d8511b76-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:16:53 crc kubenswrapper[4789]: I1122 08:16:53.280493 4789 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f3130d2e-6671-4a4b-978b-eac10b62cdc8-config\") on node \"crc\" DevicePath \"\"" Nov 22 08:16:53 crc kubenswrapper[4789]: I1122 08:16:53.280503 4789 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f3130d2e-6671-4a4b-978b-eac10b62cdc8-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 22 08:16:53 crc kubenswrapper[4789]: I1122 08:16:53.280517 4789 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f3130d2e-6671-4a4b-978b-eac10b62cdc8-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 22 08:16:53 crc kubenswrapper[4789]: I1122 08:16:53.399517 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:16:53 crc kubenswrapper[4789]: I1122 08:16:53.424593 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:16:53 crc kubenswrapper[4789]: I1122 08:16:53.432557 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:16:53 crc kubenswrapper[4789]: E1122 08:16:53.433991 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94d7bb01-64a7-442a-aee4-e958d8511b76" containerName="sg-core" Nov 22 08:16:53 crc kubenswrapper[4789]: I1122 08:16:53.434010 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="94d7bb01-64a7-442a-aee4-e958d8511b76" containerName="sg-core" Nov 22 08:16:53 crc kubenswrapper[4789]: E1122 08:16:53.434030 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94d7bb01-64a7-442a-aee4-e958d8511b76" containerName="ceilometer-notification-agent" Nov 22 08:16:53 crc kubenswrapper[4789]: I1122 08:16:53.434037 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="94d7bb01-64a7-442a-aee4-e958d8511b76" containerName="ceilometer-notification-agent" Nov 22 08:16:53 crc kubenswrapper[4789]: E1122 08:16:53.434055 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3130d2e-6671-4a4b-978b-eac10b62cdc8" containerName="init" Nov 22 08:16:53 crc kubenswrapper[4789]: I1122 08:16:53.434061 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3130d2e-6671-4a4b-978b-eac10b62cdc8" containerName="init" Nov 22 08:16:53 crc kubenswrapper[4789]: E1122 08:16:53.434071 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94d7bb01-64a7-442a-aee4-e958d8511b76" containerName="ceilometer-central-agent" Nov 22 08:16:53 crc kubenswrapper[4789]: I1122 08:16:53.434077 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="94d7bb01-64a7-442a-aee4-e958d8511b76" containerName="ceilometer-central-agent" Nov 22 08:16:53 crc kubenswrapper[4789]: E1122 08:16:53.434096 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94d7bb01-64a7-442a-aee4-e958d8511b76" containerName="proxy-httpd" Nov 22 08:16:53 crc kubenswrapper[4789]: I1122 08:16:53.434101 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="94d7bb01-64a7-442a-aee4-e958d8511b76" containerName="proxy-httpd" Nov 22 08:16:53 crc kubenswrapper[4789]: I1122 08:16:53.434261 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="94d7bb01-64a7-442a-aee4-e958d8511b76" containerName="sg-core" Nov 22 08:16:53 crc kubenswrapper[4789]: I1122 08:16:53.434275 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="94d7bb01-64a7-442a-aee4-e958d8511b76" containerName="ceilometer-notification-agent" Nov 22 08:16:53 crc kubenswrapper[4789]: I1122 08:16:53.434289 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3130d2e-6671-4a4b-978b-eac10b62cdc8" containerName="init" Nov 22 08:16:53 crc kubenswrapper[4789]: I1122 08:16:53.434301 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="94d7bb01-64a7-442a-aee4-e958d8511b76" containerName="proxy-httpd" Nov 22 08:16:53 crc kubenswrapper[4789]: I1122 08:16:53.434312 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="94d7bb01-64a7-442a-aee4-e958d8511b76" containerName="ceilometer-central-agent" Nov 22 08:16:53 crc kubenswrapper[4789]: I1122 08:16:53.436093 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 08:16:53 crc kubenswrapper[4789]: I1122 08:16:53.439089 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 22 08:16:53 crc kubenswrapper[4789]: I1122 08:16:53.448722 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 22 08:16:53 crc kubenswrapper[4789]: I1122 08:16:53.501566 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:16:53 crc kubenswrapper[4789]: I1122 08:16:53.588013 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de991bf6-8f96-4c99-bf7a-b8909c0c80d2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"de991bf6-8f96-4c99-bf7a-b8909c0c80d2\") " pod="openstack/ceilometer-0" Nov 22 08:16:53 crc kubenswrapper[4789]: I1122 08:16:53.588059 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/de991bf6-8f96-4c99-bf7a-b8909c0c80d2-scripts\") pod \"ceilometer-0\" (UID: \"de991bf6-8f96-4c99-bf7a-b8909c0c80d2\") " pod="openstack/ceilometer-0" Nov 22 08:16:53 crc kubenswrapper[4789]: I1122 08:16:53.588106 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de991bf6-8f96-4c99-bf7a-b8909c0c80d2-config-data\") pod \"ceilometer-0\" (UID: \"de991bf6-8f96-4c99-bf7a-b8909c0c80d2\") " pod="openstack/ceilometer-0" Nov 22 08:16:53 crc kubenswrapper[4789]: I1122 08:16:53.588128 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/de991bf6-8f96-4c99-bf7a-b8909c0c80d2-run-httpd\") pod \"ceilometer-0\" (UID: \"de991bf6-8f96-4c99-bf7a-b8909c0c80d2\") " pod="openstack/ceilometer-0" Nov 22 08:16:53 crc kubenswrapper[4789]: I1122 08:16:53.588190 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ppk69\" (UniqueName: \"kubernetes.io/projected/de991bf6-8f96-4c99-bf7a-b8909c0c80d2-kube-api-access-ppk69\") pod \"ceilometer-0\" (UID: \"de991bf6-8f96-4c99-bf7a-b8909c0c80d2\") " pod="openstack/ceilometer-0" Nov 22 08:16:53 crc kubenswrapper[4789]: I1122 08:16:53.588225 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/de991bf6-8f96-4c99-bf7a-b8909c0c80d2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"de991bf6-8f96-4c99-bf7a-b8909c0c80d2\") " pod="openstack/ceilometer-0" Nov 22 08:16:53 crc kubenswrapper[4789]: I1122 08:16:53.588308 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/de991bf6-8f96-4c99-bf7a-b8909c0c80d2-log-httpd\") pod \"ceilometer-0\" (UID: \"de991bf6-8f96-4c99-bf7a-b8909c0c80d2\") " pod="openstack/ceilometer-0" Nov 22 08:16:53 crc kubenswrapper[4789]: I1122 08:16:53.689654 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de991bf6-8f96-4c99-bf7a-b8909c0c80d2-config-data\") pod \"ceilometer-0\" (UID: \"de991bf6-8f96-4c99-bf7a-b8909c0c80d2\") " pod="openstack/ceilometer-0" Nov 22 08:16:53 crc kubenswrapper[4789]: I1122 08:16:53.689704 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/de991bf6-8f96-4c99-bf7a-b8909c0c80d2-run-httpd\") pod \"ceilometer-0\" (UID: \"de991bf6-8f96-4c99-bf7a-b8909c0c80d2\") " pod="openstack/ceilometer-0" Nov 22 08:16:53 crc kubenswrapper[4789]: I1122 08:16:53.689913 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ppk69\" (UniqueName: \"kubernetes.io/projected/de991bf6-8f96-4c99-bf7a-b8909c0c80d2-kube-api-access-ppk69\") pod \"ceilometer-0\" (UID: \"de991bf6-8f96-4c99-bf7a-b8909c0c80d2\") " pod="openstack/ceilometer-0" Nov 22 08:16:53 crc kubenswrapper[4789]: I1122 08:16:53.689959 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/de991bf6-8f96-4c99-bf7a-b8909c0c80d2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"de991bf6-8f96-4c99-bf7a-b8909c0c80d2\") " pod="openstack/ceilometer-0" Nov 22 08:16:53 crc kubenswrapper[4789]: I1122 08:16:53.690001 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/de991bf6-8f96-4c99-bf7a-b8909c0c80d2-log-httpd\") pod \"ceilometer-0\" (UID: \"de991bf6-8f96-4c99-bf7a-b8909c0c80d2\") " pod="openstack/ceilometer-0" Nov 22 08:16:53 crc kubenswrapper[4789]: I1122 08:16:53.690152 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de991bf6-8f96-4c99-bf7a-b8909c0c80d2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"de991bf6-8f96-4c99-bf7a-b8909c0c80d2\") " pod="openstack/ceilometer-0" Nov 22 08:16:53 crc kubenswrapper[4789]: I1122 08:16:53.690180 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/de991bf6-8f96-4c99-bf7a-b8909c0c80d2-scripts\") pod \"ceilometer-0\" (UID: \"de991bf6-8f96-4c99-bf7a-b8909c0c80d2\") " pod="openstack/ceilometer-0" Nov 22 08:16:53 crc kubenswrapper[4789]: I1122 08:16:53.690514 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/de991bf6-8f96-4c99-bf7a-b8909c0c80d2-run-httpd\") pod \"ceilometer-0\" (UID: \"de991bf6-8f96-4c99-bf7a-b8909c0c80d2\") " pod="openstack/ceilometer-0" Nov 22 08:16:53 crc kubenswrapper[4789]: I1122 08:16:53.690724 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/de991bf6-8f96-4c99-bf7a-b8909c0c80d2-log-httpd\") pod \"ceilometer-0\" (UID: \"de991bf6-8f96-4c99-bf7a-b8909c0c80d2\") " pod="openstack/ceilometer-0" Nov 22 08:16:53 crc kubenswrapper[4789]: I1122 08:16:53.695509 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de991bf6-8f96-4c99-bf7a-b8909c0c80d2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"de991bf6-8f96-4c99-bf7a-b8909c0c80d2\") " pod="openstack/ceilometer-0" Nov 22 08:16:53 crc kubenswrapper[4789]: I1122 08:16:53.700056 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de991bf6-8f96-4c99-bf7a-b8909c0c80d2-config-data\") pod \"ceilometer-0\" (UID: \"de991bf6-8f96-4c99-bf7a-b8909c0c80d2\") " pod="openstack/ceilometer-0" Nov 22 08:16:53 crc kubenswrapper[4789]: I1122 08:16:53.700429 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/de991bf6-8f96-4c99-bf7a-b8909c0c80d2-scripts\") pod \"ceilometer-0\" (UID: \"de991bf6-8f96-4c99-bf7a-b8909c0c80d2\") " pod="openstack/ceilometer-0" Nov 22 08:16:53 crc kubenswrapper[4789]: I1122 08:16:53.701122 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/de991bf6-8f96-4c99-bf7a-b8909c0c80d2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"de991bf6-8f96-4c99-bf7a-b8909c0c80d2\") " pod="openstack/ceilometer-0" Nov 22 08:16:53 crc kubenswrapper[4789]: I1122 08:16:53.720622 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ppk69\" (UniqueName: \"kubernetes.io/projected/de991bf6-8f96-4c99-bf7a-b8909c0c80d2-kube-api-access-ppk69\") pod \"ceilometer-0\" (UID: \"de991bf6-8f96-4c99-bf7a-b8909c0c80d2\") " pod="openstack/ceilometer-0" Nov 22 08:16:53 crc kubenswrapper[4789]: I1122 08:16:53.777286 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 08:16:53 crc kubenswrapper[4789]: I1122 08:16:53.792178 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-775457b975-x8hvd" event={"ID":"63d8628c-ab9f-4c35-8550-08d647beffa4","Type":"ContainerStarted","Data":"edb4df9f694719c6d72cb115eb8dcf575170c30a9b7954588cb0ade5ba02390f"} Nov 22 08:16:53 crc kubenswrapper[4789]: I1122 08:16:53.792527 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-775457b975-x8hvd" Nov 22 08:16:53 crc kubenswrapper[4789]: I1122 08:16:53.798832 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-844b557b9c-lvrgl" event={"ID":"f3130d2e-6671-4a4b-978b-eac10b62cdc8","Type":"ContainerDied","Data":"4e594f7116255277649b2f47d7fffe951d7791097ca30e8eb7a91043e41420c5"} Nov 22 08:16:53 crc kubenswrapper[4789]: I1122 08:16:53.798883 4789 scope.go:117] "RemoveContainer" containerID="7880bf06c4e8110d8b726cf2edb81c414055cf50551ba7e85ff7962a65a61cc7" Nov 22 08:16:53 crc kubenswrapper[4789]: I1122 08:16:53.799025 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-844b557b9c-lvrgl" Nov 22 08:16:53 crc kubenswrapper[4789]: I1122 08:16:53.809535 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-5b56df9788-wsmhq" Nov 22 08:16:53 crc kubenswrapper[4789]: I1122 08:16:53.814423 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"60e8e2c6-6de6-48c4-a004-a15eb39f4d84","Type":"ContainerStarted","Data":"869cba33126819ca1457ce4eb9f4170e778c6028cdba96e83c23055f9117517c"} Nov 22 08:16:53 crc kubenswrapper[4789]: I1122 08:16:53.814470 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-775457b975-x8hvd" podStartSLOduration=4.814450588 podStartE2EDuration="4.814450588s" podCreationTimestamp="2025-11-22 08:16:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:16:53.813672056 +0000 UTC m=+1228.048072329" watchObservedRunningTime="2025-11-22 08:16:53.814450588 +0000 UTC m=+1228.048850861" Nov 22 08:16:53 crc kubenswrapper[4789]: I1122 08:16:53.828081 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e1b1a2ae-3c67-4ada-b478-92ee0a9bb30d","Type":"ContainerStarted","Data":"0994c405912329d7244b21305bd0c8f8ec9e21db14d6162835da6b124c7b3674"} Nov 22 08:16:53 crc kubenswrapper[4789]: I1122 08:16:53.913859 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-844b557b9c-lvrgl"] Nov 22 08:16:53 crc kubenswrapper[4789]: I1122 08:16:53.916875 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-844b557b9c-lvrgl"] Nov 22 08:16:53 crc kubenswrapper[4789]: I1122 08:16:53.931569 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-7ccbbd9f98-jd5vz" podStartSLOduration=5.295896791 podStartE2EDuration="7.931542846s" podCreationTimestamp="2025-11-22 08:16:46 +0000 UTC" firstStartedPulling="2025-11-22 08:16:47.645260882 +0000 UTC m=+1221.879661155" lastFinishedPulling="2025-11-22 08:16:50.280906937 +0000 UTC m=+1224.515307210" observedRunningTime="2025-11-22 08:16:53.919944568 +0000 UTC m=+1228.154344851" watchObservedRunningTime="2025-11-22 08:16:53.931542846 +0000 UTC m=+1228.165943119" Nov 22 08:16:54 crc kubenswrapper[4789]: I1122 08:16:54.001382 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="94d7bb01-64a7-442a-aee4-e958d8511b76" path="/var/lib/kubelet/pods/94d7bb01-64a7-442a-aee4-e958d8511b76/volumes" Nov 22 08:16:54 crc kubenswrapper[4789]: I1122 08:16:54.004324 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f3130d2e-6671-4a4b-978b-eac10b62cdc8" path="/var/lib/kubelet/pods/f3130d2e-6671-4a4b-978b-eac10b62cdc8/volumes" Nov 22 08:16:54 crc kubenswrapper[4789]: I1122 08:16:54.425390 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-7b778f8c8-6rl4f" Nov 22 08:16:54 crc kubenswrapper[4789]: I1122 08:16:54.503596 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5b56df9788-wsmhq"] Nov 22 08:16:54 crc kubenswrapper[4789]: I1122 08:16:54.538218 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Nov 22 08:16:54 crc kubenswrapper[4789]: I1122 08:16:54.570161 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:16:54 crc kubenswrapper[4789]: I1122 08:16:54.726699 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-5fc55b98d9-vqxcz"] Nov 22 08:16:54 crc kubenswrapper[4789]: I1122 08:16:54.728432 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5fc55b98d9-vqxcz" Nov 22 08:16:54 crc kubenswrapper[4789]: I1122 08:16:54.730945 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Nov 22 08:16:54 crc kubenswrapper[4789]: I1122 08:16:54.731903 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Nov 22 08:16:54 crc kubenswrapper[4789]: I1122 08:16:54.738590 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5fc55b98d9-vqxcz"] Nov 22 08:16:54 crc kubenswrapper[4789]: I1122 08:16:54.818024 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/a5636b40-2e27-4073-8f3a-1d429a885936-httpd-config\") pod \"neutron-5fc55b98d9-vqxcz\" (UID: \"a5636b40-2e27-4073-8f3a-1d429a885936\") " pod="openstack/neutron-5fc55b98d9-vqxcz" Nov 22 08:16:54 crc kubenswrapper[4789]: I1122 08:16:54.818088 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a5636b40-2e27-4073-8f3a-1d429a885936-ovndb-tls-certs\") pod \"neutron-5fc55b98d9-vqxcz\" (UID: \"a5636b40-2e27-4073-8f3a-1d429a885936\") " pod="openstack/neutron-5fc55b98d9-vqxcz" Nov 22 08:16:54 crc kubenswrapper[4789]: I1122 08:16:54.818119 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5636b40-2e27-4073-8f3a-1d429a885936-combined-ca-bundle\") pod \"neutron-5fc55b98d9-vqxcz\" (UID: \"a5636b40-2e27-4073-8f3a-1d429a885936\") " pod="openstack/neutron-5fc55b98d9-vqxcz" Nov 22 08:16:54 crc kubenswrapper[4789]: I1122 08:16:54.818150 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z6xx2\" (UniqueName: \"kubernetes.io/projected/a5636b40-2e27-4073-8f3a-1d429a885936-kube-api-access-z6xx2\") pod \"neutron-5fc55b98d9-vqxcz\" (UID: \"a5636b40-2e27-4073-8f3a-1d429a885936\") " pod="openstack/neutron-5fc55b98d9-vqxcz" Nov 22 08:16:54 crc kubenswrapper[4789]: I1122 08:16:54.818168 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a5636b40-2e27-4073-8f3a-1d429a885936-internal-tls-certs\") pod \"neutron-5fc55b98d9-vqxcz\" (UID: \"a5636b40-2e27-4073-8f3a-1d429a885936\") " pod="openstack/neutron-5fc55b98d9-vqxcz" Nov 22 08:16:54 crc kubenswrapper[4789]: I1122 08:16:54.818214 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/a5636b40-2e27-4073-8f3a-1d429a885936-config\") pod \"neutron-5fc55b98d9-vqxcz\" (UID: \"a5636b40-2e27-4073-8f3a-1d429a885936\") " pod="openstack/neutron-5fc55b98d9-vqxcz" Nov 22 08:16:54 crc kubenswrapper[4789]: I1122 08:16:54.818268 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a5636b40-2e27-4073-8f3a-1d429a885936-public-tls-certs\") pod \"neutron-5fc55b98d9-vqxcz\" (UID: \"a5636b40-2e27-4073-8f3a-1d429a885936\") " pod="openstack/neutron-5fc55b98d9-vqxcz" Nov 22 08:16:54 crc kubenswrapper[4789]: I1122 08:16:54.837735 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"de991bf6-8f96-4c99-bf7a-b8909c0c80d2","Type":"ContainerStarted","Data":"6aa40ca2ed1346bdebd7b6c732bc0d934c01d47edfac917c83aafa706fec5731"} Nov 22 08:16:54 crc kubenswrapper[4789]: I1122 08:16:54.837933 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-5b56df9788-wsmhq" podUID="56fd3074-a15c-4024-baf9-f784a50e14a5" containerName="horizon-log" containerID="cri-o://93ff7a096adcf0bb8be87de88e1df3ddfb208ee02cbae6c682c5da91f25f03f9" gracePeriod=30 Nov 22 08:16:54 crc kubenswrapper[4789]: I1122 08:16:54.837992 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-5b56df9788-wsmhq" podUID="56fd3074-a15c-4024-baf9-f784a50e14a5" containerName="horizon" containerID="cri-o://2e99dc56ab31ddc560842a39626551056b3b1db94ece09fb54c781cd0d891832" gracePeriod=30 Nov 22 08:16:54 crc kubenswrapper[4789]: I1122 08:16:54.919706 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5636b40-2e27-4073-8f3a-1d429a885936-combined-ca-bundle\") pod \"neutron-5fc55b98d9-vqxcz\" (UID: \"a5636b40-2e27-4073-8f3a-1d429a885936\") " pod="openstack/neutron-5fc55b98d9-vqxcz" Nov 22 08:16:54 crc kubenswrapper[4789]: I1122 08:16:54.919765 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z6xx2\" (UniqueName: \"kubernetes.io/projected/a5636b40-2e27-4073-8f3a-1d429a885936-kube-api-access-z6xx2\") pod \"neutron-5fc55b98d9-vqxcz\" (UID: \"a5636b40-2e27-4073-8f3a-1d429a885936\") " pod="openstack/neutron-5fc55b98d9-vqxcz" Nov 22 08:16:54 crc kubenswrapper[4789]: I1122 08:16:54.919782 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a5636b40-2e27-4073-8f3a-1d429a885936-internal-tls-certs\") pod \"neutron-5fc55b98d9-vqxcz\" (UID: \"a5636b40-2e27-4073-8f3a-1d429a885936\") " pod="openstack/neutron-5fc55b98d9-vqxcz" Nov 22 08:16:54 crc kubenswrapper[4789]: I1122 08:16:54.919844 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/a5636b40-2e27-4073-8f3a-1d429a885936-config\") pod \"neutron-5fc55b98d9-vqxcz\" (UID: \"a5636b40-2e27-4073-8f3a-1d429a885936\") " pod="openstack/neutron-5fc55b98d9-vqxcz" Nov 22 08:16:54 crc kubenswrapper[4789]: I1122 08:16:54.919929 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a5636b40-2e27-4073-8f3a-1d429a885936-public-tls-certs\") pod \"neutron-5fc55b98d9-vqxcz\" (UID: \"a5636b40-2e27-4073-8f3a-1d429a885936\") " pod="openstack/neutron-5fc55b98d9-vqxcz" Nov 22 08:16:54 crc kubenswrapper[4789]: I1122 08:16:54.919958 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/a5636b40-2e27-4073-8f3a-1d429a885936-httpd-config\") pod \"neutron-5fc55b98d9-vqxcz\" (UID: \"a5636b40-2e27-4073-8f3a-1d429a885936\") " pod="openstack/neutron-5fc55b98d9-vqxcz" Nov 22 08:16:54 crc kubenswrapper[4789]: I1122 08:16:54.919993 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a5636b40-2e27-4073-8f3a-1d429a885936-ovndb-tls-certs\") pod \"neutron-5fc55b98d9-vqxcz\" (UID: \"a5636b40-2e27-4073-8f3a-1d429a885936\") " pod="openstack/neutron-5fc55b98d9-vqxcz" Nov 22 08:16:54 crc kubenswrapper[4789]: I1122 08:16:54.926080 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/a5636b40-2e27-4073-8f3a-1d429a885936-httpd-config\") pod \"neutron-5fc55b98d9-vqxcz\" (UID: \"a5636b40-2e27-4073-8f3a-1d429a885936\") " pod="openstack/neutron-5fc55b98d9-vqxcz" Nov 22 08:16:54 crc kubenswrapper[4789]: I1122 08:16:54.931673 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a5636b40-2e27-4073-8f3a-1d429a885936-public-tls-certs\") pod \"neutron-5fc55b98d9-vqxcz\" (UID: \"a5636b40-2e27-4073-8f3a-1d429a885936\") " pod="openstack/neutron-5fc55b98d9-vqxcz" Nov 22 08:16:54 crc kubenswrapper[4789]: I1122 08:16:54.931973 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/a5636b40-2e27-4073-8f3a-1d429a885936-config\") pod \"neutron-5fc55b98d9-vqxcz\" (UID: \"a5636b40-2e27-4073-8f3a-1d429a885936\") " pod="openstack/neutron-5fc55b98d9-vqxcz" Nov 22 08:16:54 crc kubenswrapper[4789]: I1122 08:16:54.932242 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a5636b40-2e27-4073-8f3a-1d429a885936-internal-tls-certs\") pod \"neutron-5fc55b98d9-vqxcz\" (UID: \"a5636b40-2e27-4073-8f3a-1d429a885936\") " pod="openstack/neutron-5fc55b98d9-vqxcz" Nov 22 08:16:54 crc kubenswrapper[4789]: I1122 08:16:54.933295 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5636b40-2e27-4073-8f3a-1d429a885936-combined-ca-bundle\") pod \"neutron-5fc55b98d9-vqxcz\" (UID: \"a5636b40-2e27-4073-8f3a-1d429a885936\") " pod="openstack/neutron-5fc55b98d9-vqxcz" Nov 22 08:16:54 crc kubenswrapper[4789]: I1122 08:16:54.940080 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a5636b40-2e27-4073-8f3a-1d429a885936-ovndb-tls-certs\") pod \"neutron-5fc55b98d9-vqxcz\" (UID: \"a5636b40-2e27-4073-8f3a-1d429a885936\") " pod="openstack/neutron-5fc55b98d9-vqxcz" Nov 22 08:16:54 crc kubenswrapper[4789]: I1122 08:16:54.951508 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z6xx2\" (UniqueName: \"kubernetes.io/projected/a5636b40-2e27-4073-8f3a-1d429a885936-kube-api-access-z6xx2\") pod \"neutron-5fc55b98d9-vqxcz\" (UID: \"a5636b40-2e27-4073-8f3a-1d429a885936\") " pod="openstack/neutron-5fc55b98d9-vqxcz" Nov 22 08:16:55 crc kubenswrapper[4789]: I1122 08:16:55.054922 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5fc55b98d9-vqxcz" Nov 22 08:16:55 crc kubenswrapper[4789]: I1122 08:16:55.582711 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5fc55b98d9-vqxcz"] Nov 22 08:16:55 crc kubenswrapper[4789]: W1122 08:16:55.601225 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda5636b40_2e27_4073_8f3a_1d429a885936.slice/crio-afa4fd5954dc4d9d12a1c60805d99558cdbc6d3dfc5c8359cc97115135c1311e WatchSource:0}: Error finding container afa4fd5954dc4d9d12a1c60805d99558cdbc6d3dfc5c8359cc97115135c1311e: Status 404 returned error can't find the container with id afa4fd5954dc4d9d12a1c60805d99558cdbc6d3dfc5c8359cc97115135c1311e Nov 22 08:16:55 crc kubenswrapper[4789]: I1122 08:16:55.847306 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"60e8e2c6-6de6-48c4-a004-a15eb39f4d84","Type":"ContainerStarted","Data":"94dde7e5cd6bb5a623fda65cf743e4fb3dadd9222bf47cffff2138bc308a4677"} Nov 22 08:16:55 crc kubenswrapper[4789]: I1122 08:16:55.848856 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Nov 22 08:16:55 crc kubenswrapper[4789]: I1122 08:16:55.847533 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="60e8e2c6-6de6-48c4-a004-a15eb39f4d84" containerName="cinder-api-log" containerID="cri-o://869cba33126819ca1457ce4eb9f4170e778c6028cdba96e83c23055f9117517c" gracePeriod=30 Nov 22 08:16:55 crc kubenswrapper[4789]: I1122 08:16:55.847501 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="60e8e2c6-6de6-48c4-a004-a15eb39f4d84" containerName="cinder-api" containerID="cri-o://94dde7e5cd6bb5a623fda65cf743e4fb3dadd9222bf47cffff2138bc308a4677" gracePeriod=30 Nov 22 08:16:55 crc kubenswrapper[4789]: I1122 08:16:55.850277 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5fc55b98d9-vqxcz" event={"ID":"a5636b40-2e27-4073-8f3a-1d429a885936","Type":"ContainerStarted","Data":"afa4fd5954dc4d9d12a1c60805d99558cdbc6d3dfc5c8359cc97115135c1311e"} Nov 22 08:16:55 crc kubenswrapper[4789]: I1122 08:16:55.854390 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e1b1a2ae-3c67-4ada-b478-92ee0a9bb30d","Type":"ContainerStarted","Data":"de49ff9e3613638a69b7d9cc20c5d4372e86c6ad0721d1b9f28300cccc0089ed"} Nov 22 08:16:55 crc kubenswrapper[4789]: I1122 08:16:55.872305 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=6.87228442 podStartE2EDuration="6.87228442s" podCreationTimestamp="2025-11-22 08:16:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:16:55.86498667 +0000 UTC m=+1230.099386943" watchObservedRunningTime="2025-11-22 08:16:55.87228442 +0000 UTC m=+1230.106684693" Nov 22 08:16:55 crc kubenswrapper[4789]: I1122 08:16:55.893890 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=5.518318073 podStartE2EDuration="6.893872669s" podCreationTimestamp="2025-11-22 08:16:49 +0000 UTC" firstStartedPulling="2025-11-22 08:16:50.586290927 +0000 UTC m=+1224.820691200" lastFinishedPulling="2025-11-22 08:16:51.961845523 +0000 UTC m=+1226.196245796" observedRunningTime="2025-11-22 08:16:55.888318157 +0000 UTC m=+1230.122718430" watchObservedRunningTime="2025-11-22 08:16:55.893872669 +0000 UTC m=+1230.128272942" Nov 22 08:16:56 crc kubenswrapper[4789]: I1122 08:16:56.496503 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 22 08:16:56 crc kubenswrapper[4789]: I1122 08:16:56.647674 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/60e8e2c6-6de6-48c4-a004-a15eb39f4d84-logs\") pod \"60e8e2c6-6de6-48c4-a004-a15eb39f4d84\" (UID: \"60e8e2c6-6de6-48c4-a004-a15eb39f4d84\") " Nov 22 08:16:56 crc kubenswrapper[4789]: I1122 08:16:56.647727 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/60e8e2c6-6de6-48c4-a004-a15eb39f4d84-scripts\") pod \"60e8e2c6-6de6-48c4-a004-a15eb39f4d84\" (UID: \"60e8e2c6-6de6-48c4-a004-a15eb39f4d84\") " Nov 22 08:16:56 crc kubenswrapper[4789]: I1122 08:16:56.647796 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/60e8e2c6-6de6-48c4-a004-a15eb39f4d84-config-data-custom\") pod \"60e8e2c6-6de6-48c4-a004-a15eb39f4d84\" (UID: \"60e8e2c6-6de6-48c4-a004-a15eb39f4d84\") " Nov 22 08:16:56 crc kubenswrapper[4789]: I1122 08:16:56.647853 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60e8e2c6-6de6-48c4-a004-a15eb39f4d84-config-data\") pod \"60e8e2c6-6de6-48c4-a004-a15eb39f4d84\" (UID: \"60e8e2c6-6de6-48c4-a004-a15eb39f4d84\") " Nov 22 08:16:56 crc kubenswrapper[4789]: I1122 08:16:56.647916 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/60e8e2c6-6de6-48c4-a004-a15eb39f4d84-etc-machine-id\") pod \"60e8e2c6-6de6-48c4-a004-a15eb39f4d84\" (UID: \"60e8e2c6-6de6-48c4-a004-a15eb39f4d84\") " Nov 22 08:16:56 crc kubenswrapper[4789]: I1122 08:16:56.648042 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-777ct\" (UniqueName: \"kubernetes.io/projected/60e8e2c6-6de6-48c4-a004-a15eb39f4d84-kube-api-access-777ct\") pod \"60e8e2c6-6de6-48c4-a004-a15eb39f4d84\" (UID: \"60e8e2c6-6de6-48c4-a004-a15eb39f4d84\") " Nov 22 08:16:56 crc kubenswrapper[4789]: I1122 08:16:56.648086 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60e8e2c6-6de6-48c4-a004-a15eb39f4d84-combined-ca-bundle\") pod \"60e8e2c6-6de6-48c4-a004-a15eb39f4d84\" (UID: \"60e8e2c6-6de6-48c4-a004-a15eb39f4d84\") " Nov 22 08:16:56 crc kubenswrapper[4789]: I1122 08:16:56.648505 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/60e8e2c6-6de6-48c4-a004-a15eb39f4d84-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "60e8e2c6-6de6-48c4-a004-a15eb39f4d84" (UID: "60e8e2c6-6de6-48c4-a004-a15eb39f4d84"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 08:16:56 crc kubenswrapper[4789]: I1122 08:16:56.648990 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/60e8e2c6-6de6-48c4-a004-a15eb39f4d84-logs" (OuterVolumeSpecName: "logs") pod "60e8e2c6-6de6-48c4-a004-a15eb39f4d84" (UID: "60e8e2c6-6de6-48c4-a004-a15eb39f4d84"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:16:56 crc kubenswrapper[4789]: I1122 08:16:56.649814 4789 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/60e8e2c6-6de6-48c4-a004-a15eb39f4d84-logs\") on node \"crc\" DevicePath \"\"" Nov 22 08:16:56 crc kubenswrapper[4789]: I1122 08:16:56.649844 4789 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/60e8e2c6-6de6-48c4-a004-a15eb39f4d84-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 22 08:16:56 crc kubenswrapper[4789]: I1122 08:16:56.655127 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60e8e2c6-6de6-48c4-a004-a15eb39f4d84-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "60e8e2c6-6de6-48c4-a004-a15eb39f4d84" (UID: "60e8e2c6-6de6-48c4-a004-a15eb39f4d84"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:16:56 crc kubenswrapper[4789]: I1122 08:16:56.656684 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60e8e2c6-6de6-48c4-a004-a15eb39f4d84-scripts" (OuterVolumeSpecName: "scripts") pod "60e8e2c6-6de6-48c4-a004-a15eb39f4d84" (UID: "60e8e2c6-6de6-48c4-a004-a15eb39f4d84"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:16:56 crc kubenswrapper[4789]: I1122 08:16:56.656864 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60e8e2c6-6de6-48c4-a004-a15eb39f4d84-kube-api-access-777ct" (OuterVolumeSpecName: "kube-api-access-777ct") pod "60e8e2c6-6de6-48c4-a004-a15eb39f4d84" (UID: "60e8e2c6-6de6-48c4-a004-a15eb39f4d84"). InnerVolumeSpecName "kube-api-access-777ct". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:16:56 crc kubenswrapper[4789]: I1122 08:16:56.693125 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60e8e2c6-6de6-48c4-a004-a15eb39f4d84-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "60e8e2c6-6de6-48c4-a004-a15eb39f4d84" (UID: "60e8e2c6-6de6-48c4-a004-a15eb39f4d84"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:16:56 crc kubenswrapper[4789]: I1122 08:16:56.735117 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60e8e2c6-6de6-48c4-a004-a15eb39f4d84-config-data" (OuterVolumeSpecName: "config-data") pod "60e8e2c6-6de6-48c4-a004-a15eb39f4d84" (UID: "60e8e2c6-6de6-48c4-a004-a15eb39f4d84"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:16:56 crc kubenswrapper[4789]: I1122 08:16:56.751466 4789 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/60e8e2c6-6de6-48c4-a004-a15eb39f4d84-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:16:56 crc kubenswrapper[4789]: I1122 08:16:56.751509 4789 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/60e8e2c6-6de6-48c4-a004-a15eb39f4d84-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 22 08:16:56 crc kubenswrapper[4789]: I1122 08:16:56.751526 4789 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60e8e2c6-6de6-48c4-a004-a15eb39f4d84-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:16:56 crc kubenswrapper[4789]: I1122 08:16:56.751541 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-777ct\" (UniqueName: \"kubernetes.io/projected/60e8e2c6-6de6-48c4-a004-a15eb39f4d84-kube-api-access-777ct\") on node \"crc\" DevicePath \"\"" Nov 22 08:16:56 crc kubenswrapper[4789]: I1122 08:16:56.751553 4789 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60e8e2c6-6de6-48c4-a004-a15eb39f4d84-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:16:56 crc kubenswrapper[4789]: I1122 08:16:56.863823 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"de991bf6-8f96-4c99-bf7a-b8909c0c80d2","Type":"ContainerStarted","Data":"60c80dac437908b708a2f24653527c9e713047e6113c781463535e492f681d00"} Nov 22 08:16:56 crc kubenswrapper[4789]: I1122 08:16:56.867202 4789 generic.go:334] "Generic (PLEG): container finished" podID="60e8e2c6-6de6-48c4-a004-a15eb39f4d84" containerID="94dde7e5cd6bb5a623fda65cf743e4fb3dadd9222bf47cffff2138bc308a4677" exitCode=0 Nov 22 08:16:56 crc kubenswrapper[4789]: I1122 08:16:56.867233 4789 generic.go:334] "Generic (PLEG): container finished" podID="60e8e2c6-6de6-48c4-a004-a15eb39f4d84" containerID="869cba33126819ca1457ce4eb9f4170e778c6028cdba96e83c23055f9117517c" exitCode=143 Nov 22 08:16:56 crc kubenswrapper[4789]: I1122 08:16:56.867264 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 22 08:16:56 crc kubenswrapper[4789]: I1122 08:16:56.867276 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"60e8e2c6-6de6-48c4-a004-a15eb39f4d84","Type":"ContainerDied","Data":"94dde7e5cd6bb5a623fda65cf743e4fb3dadd9222bf47cffff2138bc308a4677"} Nov 22 08:16:56 crc kubenswrapper[4789]: I1122 08:16:56.867302 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"60e8e2c6-6de6-48c4-a004-a15eb39f4d84","Type":"ContainerDied","Data":"869cba33126819ca1457ce4eb9f4170e778c6028cdba96e83c23055f9117517c"} Nov 22 08:16:56 crc kubenswrapper[4789]: I1122 08:16:56.867312 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"60e8e2c6-6de6-48c4-a004-a15eb39f4d84","Type":"ContainerDied","Data":"d5cef0478dc4534ff63796e77ead5a16b224a869d6bf8f9221f03be991f6ce9d"} Nov 22 08:16:56 crc kubenswrapper[4789]: I1122 08:16:56.867329 4789 scope.go:117] "RemoveContainer" containerID="94dde7e5cd6bb5a623fda65cf743e4fb3dadd9222bf47cffff2138bc308a4677" Nov 22 08:16:56 crc kubenswrapper[4789]: I1122 08:16:56.870117 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5fc55b98d9-vqxcz" event={"ID":"a5636b40-2e27-4073-8f3a-1d429a885936","Type":"ContainerStarted","Data":"f22a76f5a413a3f97c0e6ce74e1bec05eb9aa50921e3027bda7a22b6a8761020"} Nov 22 08:16:56 crc kubenswrapper[4789]: I1122 08:16:56.870158 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5fc55b98d9-vqxcz" event={"ID":"a5636b40-2e27-4073-8f3a-1d429a885936","Type":"ContainerStarted","Data":"49f90e09b7feb84562a31027c00c72d7f0b0ce67bc61fabbb2d3652ea1b34c64"} Nov 22 08:16:56 crc kubenswrapper[4789]: I1122 08:16:56.870204 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-5fc55b98d9-vqxcz" Nov 22 08:16:56 crc kubenswrapper[4789]: I1122 08:16:56.934494 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-5fc55b98d9-vqxcz" podStartSLOduration=2.934479604 podStartE2EDuration="2.934479604s" podCreationTimestamp="2025-11-22 08:16:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:16:56.930398342 +0000 UTC m=+1231.164798615" watchObservedRunningTime="2025-11-22 08:16:56.934479604 +0000 UTC m=+1231.168879877" Nov 22 08:16:56 crc kubenswrapper[4789]: I1122 08:16:56.968082 4789 scope.go:117] "RemoveContainer" containerID="869cba33126819ca1457ce4eb9f4170e778c6028cdba96e83c23055f9117517c" Nov 22 08:16:56 crc kubenswrapper[4789]: I1122 08:16:56.982825 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Nov 22 08:16:57 crc kubenswrapper[4789]: I1122 08:16:57.013577 4789 scope.go:117] "RemoveContainer" containerID="94dde7e5cd6bb5a623fda65cf743e4fb3dadd9222bf47cffff2138bc308a4677" Nov 22 08:16:57 crc kubenswrapper[4789]: E1122 08:16:57.014060 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"94dde7e5cd6bb5a623fda65cf743e4fb3dadd9222bf47cffff2138bc308a4677\": container with ID starting with 94dde7e5cd6bb5a623fda65cf743e4fb3dadd9222bf47cffff2138bc308a4677 not found: ID does not exist" containerID="94dde7e5cd6bb5a623fda65cf743e4fb3dadd9222bf47cffff2138bc308a4677" Nov 22 08:16:57 crc kubenswrapper[4789]: I1122 08:16:57.014090 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"94dde7e5cd6bb5a623fda65cf743e4fb3dadd9222bf47cffff2138bc308a4677"} err="failed to get container status \"94dde7e5cd6bb5a623fda65cf743e4fb3dadd9222bf47cffff2138bc308a4677\": rpc error: code = NotFound desc = could not find container \"94dde7e5cd6bb5a623fda65cf743e4fb3dadd9222bf47cffff2138bc308a4677\": container with ID starting with 94dde7e5cd6bb5a623fda65cf743e4fb3dadd9222bf47cffff2138bc308a4677 not found: ID does not exist" Nov 22 08:16:57 crc kubenswrapper[4789]: I1122 08:16:57.014109 4789 scope.go:117] "RemoveContainer" containerID="869cba33126819ca1457ce4eb9f4170e778c6028cdba96e83c23055f9117517c" Nov 22 08:16:57 crc kubenswrapper[4789]: E1122 08:16:57.014270 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"869cba33126819ca1457ce4eb9f4170e778c6028cdba96e83c23055f9117517c\": container with ID starting with 869cba33126819ca1457ce4eb9f4170e778c6028cdba96e83c23055f9117517c not found: ID does not exist" containerID="869cba33126819ca1457ce4eb9f4170e778c6028cdba96e83c23055f9117517c" Nov 22 08:16:57 crc kubenswrapper[4789]: I1122 08:16:57.014289 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"869cba33126819ca1457ce4eb9f4170e778c6028cdba96e83c23055f9117517c"} err="failed to get container status \"869cba33126819ca1457ce4eb9f4170e778c6028cdba96e83c23055f9117517c\": rpc error: code = NotFound desc = could not find container \"869cba33126819ca1457ce4eb9f4170e778c6028cdba96e83c23055f9117517c\": container with ID starting with 869cba33126819ca1457ce4eb9f4170e778c6028cdba96e83c23055f9117517c not found: ID does not exist" Nov 22 08:16:57 crc kubenswrapper[4789]: I1122 08:16:57.014301 4789 scope.go:117] "RemoveContainer" containerID="94dde7e5cd6bb5a623fda65cf743e4fb3dadd9222bf47cffff2138bc308a4677" Nov 22 08:16:57 crc kubenswrapper[4789]: I1122 08:16:57.014476 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"94dde7e5cd6bb5a623fda65cf743e4fb3dadd9222bf47cffff2138bc308a4677"} err="failed to get container status \"94dde7e5cd6bb5a623fda65cf743e4fb3dadd9222bf47cffff2138bc308a4677\": rpc error: code = NotFound desc = could not find container \"94dde7e5cd6bb5a623fda65cf743e4fb3dadd9222bf47cffff2138bc308a4677\": container with ID starting with 94dde7e5cd6bb5a623fda65cf743e4fb3dadd9222bf47cffff2138bc308a4677 not found: ID does not exist" Nov 22 08:16:57 crc kubenswrapper[4789]: I1122 08:16:57.014493 4789 scope.go:117] "RemoveContainer" containerID="869cba33126819ca1457ce4eb9f4170e778c6028cdba96e83c23055f9117517c" Nov 22 08:16:57 crc kubenswrapper[4789]: I1122 08:16:57.015022 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"869cba33126819ca1457ce4eb9f4170e778c6028cdba96e83c23055f9117517c"} err="failed to get container status \"869cba33126819ca1457ce4eb9f4170e778c6028cdba96e83c23055f9117517c\": rpc error: code = NotFound desc = could not find container \"869cba33126819ca1457ce4eb9f4170e778c6028cdba96e83c23055f9117517c\": container with ID starting with 869cba33126819ca1457ce4eb9f4170e778c6028cdba96e83c23055f9117517c not found: ID does not exist" Nov 22 08:16:57 crc kubenswrapper[4789]: I1122 08:16:57.049849 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Nov 22 08:16:57 crc kubenswrapper[4789]: I1122 08:16:57.060808 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Nov 22 08:16:57 crc kubenswrapper[4789]: E1122 08:16:57.061238 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60e8e2c6-6de6-48c4-a004-a15eb39f4d84" containerName="cinder-api-log" Nov 22 08:16:57 crc kubenswrapper[4789]: I1122 08:16:57.061255 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="60e8e2c6-6de6-48c4-a004-a15eb39f4d84" containerName="cinder-api-log" Nov 22 08:16:57 crc kubenswrapper[4789]: E1122 08:16:57.061269 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60e8e2c6-6de6-48c4-a004-a15eb39f4d84" containerName="cinder-api" Nov 22 08:16:57 crc kubenswrapper[4789]: I1122 08:16:57.061275 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="60e8e2c6-6de6-48c4-a004-a15eb39f4d84" containerName="cinder-api" Nov 22 08:16:57 crc kubenswrapper[4789]: I1122 08:16:57.061435 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="60e8e2c6-6de6-48c4-a004-a15eb39f4d84" containerName="cinder-api-log" Nov 22 08:16:57 crc kubenswrapper[4789]: I1122 08:16:57.061460 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="60e8e2c6-6de6-48c4-a004-a15eb39f4d84" containerName="cinder-api" Nov 22 08:16:57 crc kubenswrapper[4789]: I1122 08:16:57.064336 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 22 08:16:57 crc kubenswrapper[4789]: I1122 08:16:57.066090 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 22 08:16:57 crc kubenswrapper[4789]: I1122 08:16:57.081666 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Nov 22 08:16:57 crc kubenswrapper[4789]: I1122 08:16:57.081883 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Nov 22 08:16:57 crc kubenswrapper[4789]: I1122 08:16:57.082031 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Nov 22 08:16:57 crc kubenswrapper[4789]: I1122 08:16:57.166800 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/27920801-df7f-48f4-8830-6b190dea7234-config-data-custom\") pod \"cinder-api-0\" (UID: \"27920801-df7f-48f4-8830-6b190dea7234\") " pod="openstack/cinder-api-0" Nov 22 08:16:57 crc kubenswrapper[4789]: I1122 08:16:57.168512 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27920801-df7f-48f4-8830-6b190dea7234-config-data\") pod \"cinder-api-0\" (UID: \"27920801-df7f-48f4-8830-6b190dea7234\") " pod="openstack/cinder-api-0" Nov 22 08:16:57 crc kubenswrapper[4789]: I1122 08:16:57.168561 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27920801-df7f-48f4-8830-6b190dea7234-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"27920801-df7f-48f4-8830-6b190dea7234\") " pod="openstack/cinder-api-0" Nov 22 08:16:57 crc kubenswrapper[4789]: I1122 08:16:57.168611 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/27920801-df7f-48f4-8830-6b190dea7234-etc-machine-id\") pod \"cinder-api-0\" (UID: \"27920801-df7f-48f4-8830-6b190dea7234\") " pod="openstack/cinder-api-0" Nov 22 08:16:57 crc kubenswrapper[4789]: I1122 08:16:57.168696 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/27920801-df7f-48f4-8830-6b190dea7234-logs\") pod \"cinder-api-0\" (UID: \"27920801-df7f-48f4-8830-6b190dea7234\") " pod="openstack/cinder-api-0" Nov 22 08:16:57 crc kubenswrapper[4789]: I1122 08:16:57.168731 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/27920801-df7f-48f4-8830-6b190dea7234-scripts\") pod \"cinder-api-0\" (UID: \"27920801-df7f-48f4-8830-6b190dea7234\") " pod="openstack/cinder-api-0" Nov 22 08:16:57 crc kubenswrapper[4789]: I1122 08:16:57.168788 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/27920801-df7f-48f4-8830-6b190dea7234-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"27920801-df7f-48f4-8830-6b190dea7234\") " pod="openstack/cinder-api-0" Nov 22 08:16:57 crc kubenswrapper[4789]: I1122 08:16:57.168872 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2sxbd\" (UniqueName: \"kubernetes.io/projected/27920801-df7f-48f4-8830-6b190dea7234-kube-api-access-2sxbd\") pod \"cinder-api-0\" (UID: \"27920801-df7f-48f4-8830-6b190dea7234\") " pod="openstack/cinder-api-0" Nov 22 08:16:57 crc kubenswrapper[4789]: I1122 08:16:57.168974 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/27920801-df7f-48f4-8830-6b190dea7234-public-tls-certs\") pod \"cinder-api-0\" (UID: \"27920801-df7f-48f4-8830-6b190dea7234\") " pod="openstack/cinder-api-0" Nov 22 08:16:57 crc kubenswrapper[4789]: I1122 08:16:57.270776 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2sxbd\" (UniqueName: \"kubernetes.io/projected/27920801-df7f-48f4-8830-6b190dea7234-kube-api-access-2sxbd\") pod \"cinder-api-0\" (UID: \"27920801-df7f-48f4-8830-6b190dea7234\") " pod="openstack/cinder-api-0" Nov 22 08:16:57 crc kubenswrapper[4789]: I1122 08:16:57.270873 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/27920801-df7f-48f4-8830-6b190dea7234-public-tls-certs\") pod \"cinder-api-0\" (UID: \"27920801-df7f-48f4-8830-6b190dea7234\") " pod="openstack/cinder-api-0" Nov 22 08:16:57 crc kubenswrapper[4789]: I1122 08:16:57.270939 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/27920801-df7f-48f4-8830-6b190dea7234-config-data-custom\") pod \"cinder-api-0\" (UID: \"27920801-df7f-48f4-8830-6b190dea7234\") " pod="openstack/cinder-api-0" Nov 22 08:16:57 crc kubenswrapper[4789]: I1122 08:16:57.270957 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27920801-df7f-48f4-8830-6b190dea7234-config-data\") pod \"cinder-api-0\" (UID: \"27920801-df7f-48f4-8830-6b190dea7234\") " pod="openstack/cinder-api-0" Nov 22 08:16:57 crc kubenswrapper[4789]: I1122 08:16:57.271009 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27920801-df7f-48f4-8830-6b190dea7234-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"27920801-df7f-48f4-8830-6b190dea7234\") " pod="openstack/cinder-api-0" Nov 22 08:16:57 crc kubenswrapper[4789]: I1122 08:16:57.271030 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/27920801-df7f-48f4-8830-6b190dea7234-etc-machine-id\") pod \"cinder-api-0\" (UID: \"27920801-df7f-48f4-8830-6b190dea7234\") " pod="openstack/cinder-api-0" Nov 22 08:16:57 crc kubenswrapper[4789]: I1122 08:16:57.271079 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/27920801-df7f-48f4-8830-6b190dea7234-logs\") pod \"cinder-api-0\" (UID: \"27920801-df7f-48f4-8830-6b190dea7234\") " pod="openstack/cinder-api-0" Nov 22 08:16:57 crc kubenswrapper[4789]: I1122 08:16:57.271105 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/27920801-df7f-48f4-8830-6b190dea7234-scripts\") pod \"cinder-api-0\" (UID: \"27920801-df7f-48f4-8830-6b190dea7234\") " pod="openstack/cinder-api-0" Nov 22 08:16:57 crc kubenswrapper[4789]: I1122 08:16:57.271125 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/27920801-df7f-48f4-8830-6b190dea7234-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"27920801-df7f-48f4-8830-6b190dea7234\") " pod="openstack/cinder-api-0" Nov 22 08:16:57 crc kubenswrapper[4789]: I1122 08:16:57.275352 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/27920801-df7f-48f4-8830-6b190dea7234-etc-machine-id\") pod \"cinder-api-0\" (UID: \"27920801-df7f-48f4-8830-6b190dea7234\") " pod="openstack/cinder-api-0" Nov 22 08:16:57 crc kubenswrapper[4789]: I1122 08:16:57.277049 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/27920801-df7f-48f4-8830-6b190dea7234-logs\") pod \"cinder-api-0\" (UID: \"27920801-df7f-48f4-8830-6b190dea7234\") " pod="openstack/cinder-api-0" Nov 22 08:16:57 crc kubenswrapper[4789]: I1122 08:16:57.284189 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27920801-df7f-48f4-8830-6b190dea7234-config-data\") pod \"cinder-api-0\" (UID: \"27920801-df7f-48f4-8830-6b190dea7234\") " pod="openstack/cinder-api-0" Nov 22 08:16:57 crc kubenswrapper[4789]: I1122 08:16:57.287971 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/27920801-df7f-48f4-8830-6b190dea7234-scripts\") pod \"cinder-api-0\" (UID: \"27920801-df7f-48f4-8830-6b190dea7234\") " pod="openstack/cinder-api-0" Nov 22 08:16:57 crc kubenswrapper[4789]: I1122 08:16:57.288431 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/27920801-df7f-48f4-8830-6b190dea7234-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"27920801-df7f-48f4-8830-6b190dea7234\") " pod="openstack/cinder-api-0" Nov 22 08:16:57 crc kubenswrapper[4789]: I1122 08:16:57.289611 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/27920801-df7f-48f4-8830-6b190dea7234-public-tls-certs\") pod \"cinder-api-0\" (UID: \"27920801-df7f-48f4-8830-6b190dea7234\") " pod="openstack/cinder-api-0" Nov 22 08:16:57 crc kubenswrapper[4789]: I1122 08:16:57.296798 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27920801-df7f-48f4-8830-6b190dea7234-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"27920801-df7f-48f4-8830-6b190dea7234\") " pod="openstack/cinder-api-0" Nov 22 08:16:57 crc kubenswrapper[4789]: I1122 08:16:57.304377 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/27920801-df7f-48f4-8830-6b190dea7234-config-data-custom\") pod \"cinder-api-0\" (UID: \"27920801-df7f-48f4-8830-6b190dea7234\") " pod="openstack/cinder-api-0" Nov 22 08:16:57 crc kubenswrapper[4789]: I1122 08:16:57.321267 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2sxbd\" (UniqueName: \"kubernetes.io/projected/27920801-df7f-48f4-8830-6b190dea7234-kube-api-access-2sxbd\") pod \"cinder-api-0\" (UID: \"27920801-df7f-48f4-8830-6b190dea7234\") " pod="openstack/cinder-api-0" Nov 22 08:16:57 crc kubenswrapper[4789]: I1122 08:16:57.574153 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 22 08:16:57 crc kubenswrapper[4789]: I1122 08:16:57.809329 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-54cbbcb976-cxb6c"] Nov 22 08:16:57 crc kubenswrapper[4789]: I1122 08:16:57.813448 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-54cbbcb976-cxb6c" Nov 22 08:16:57 crc kubenswrapper[4789]: I1122 08:16:57.832911 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Nov 22 08:16:57 crc kubenswrapper[4789]: I1122 08:16:57.833168 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Nov 22 08:16:57 crc kubenswrapper[4789]: I1122 08:16:57.865140 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-54cbbcb976-cxb6c"] Nov 22 08:16:57 crc kubenswrapper[4789]: I1122 08:16:57.885150 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee6a65b2-109d-44d8-a4e0-62ab28d392e4-config-data\") pod \"barbican-api-54cbbcb976-cxb6c\" (UID: \"ee6a65b2-109d-44d8-a4e0-62ab28d392e4\") " pod="openstack/barbican-api-54cbbcb976-cxb6c" Nov 22 08:16:57 crc kubenswrapper[4789]: I1122 08:16:57.885475 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ee6a65b2-109d-44d8-a4e0-62ab28d392e4-internal-tls-certs\") pod \"barbican-api-54cbbcb976-cxb6c\" (UID: \"ee6a65b2-109d-44d8-a4e0-62ab28d392e4\") " pod="openstack/barbican-api-54cbbcb976-cxb6c" Nov 22 08:16:57 crc kubenswrapper[4789]: I1122 08:16:57.885576 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g9rps\" (UniqueName: \"kubernetes.io/projected/ee6a65b2-109d-44d8-a4e0-62ab28d392e4-kube-api-access-g9rps\") pod \"barbican-api-54cbbcb976-cxb6c\" (UID: \"ee6a65b2-109d-44d8-a4e0-62ab28d392e4\") " pod="openstack/barbican-api-54cbbcb976-cxb6c" Nov 22 08:16:57 crc kubenswrapper[4789]: I1122 08:16:57.885615 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee6a65b2-109d-44d8-a4e0-62ab28d392e4-combined-ca-bundle\") pod \"barbican-api-54cbbcb976-cxb6c\" (UID: \"ee6a65b2-109d-44d8-a4e0-62ab28d392e4\") " pod="openstack/barbican-api-54cbbcb976-cxb6c" Nov 22 08:16:57 crc kubenswrapper[4789]: I1122 08:16:57.885665 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ee6a65b2-109d-44d8-a4e0-62ab28d392e4-public-tls-certs\") pod \"barbican-api-54cbbcb976-cxb6c\" (UID: \"ee6a65b2-109d-44d8-a4e0-62ab28d392e4\") " pod="openstack/barbican-api-54cbbcb976-cxb6c" Nov 22 08:16:57 crc kubenswrapper[4789]: I1122 08:16:57.885719 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ee6a65b2-109d-44d8-a4e0-62ab28d392e4-logs\") pod \"barbican-api-54cbbcb976-cxb6c\" (UID: \"ee6a65b2-109d-44d8-a4e0-62ab28d392e4\") " pod="openstack/barbican-api-54cbbcb976-cxb6c" Nov 22 08:16:57 crc kubenswrapper[4789]: I1122 08:16:57.885790 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ee6a65b2-109d-44d8-a4e0-62ab28d392e4-config-data-custom\") pod \"barbican-api-54cbbcb976-cxb6c\" (UID: \"ee6a65b2-109d-44d8-a4e0-62ab28d392e4\") " pod="openstack/barbican-api-54cbbcb976-cxb6c" Nov 22 08:16:57 crc kubenswrapper[4789]: I1122 08:16:57.910397 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"de991bf6-8f96-4c99-bf7a-b8909c0c80d2","Type":"ContainerStarted","Data":"63d2e8a684de408f5cfa4728b23b07803ae5286c76ec7a0b812bf69230a7922f"} Nov 22 08:16:57 crc kubenswrapper[4789]: I1122 08:16:57.980481 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="60e8e2c6-6de6-48c4-a004-a15eb39f4d84" path="/var/lib/kubelet/pods/60e8e2c6-6de6-48c4-a004-a15eb39f4d84/volumes" Nov 22 08:16:57 crc kubenswrapper[4789]: I1122 08:16:57.988467 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g9rps\" (UniqueName: \"kubernetes.io/projected/ee6a65b2-109d-44d8-a4e0-62ab28d392e4-kube-api-access-g9rps\") pod \"barbican-api-54cbbcb976-cxb6c\" (UID: \"ee6a65b2-109d-44d8-a4e0-62ab28d392e4\") " pod="openstack/barbican-api-54cbbcb976-cxb6c" Nov 22 08:16:57 crc kubenswrapper[4789]: I1122 08:16:57.988560 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee6a65b2-109d-44d8-a4e0-62ab28d392e4-combined-ca-bundle\") pod \"barbican-api-54cbbcb976-cxb6c\" (UID: \"ee6a65b2-109d-44d8-a4e0-62ab28d392e4\") " pod="openstack/barbican-api-54cbbcb976-cxb6c" Nov 22 08:16:57 crc kubenswrapper[4789]: I1122 08:16:57.988621 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ee6a65b2-109d-44d8-a4e0-62ab28d392e4-public-tls-certs\") pod \"barbican-api-54cbbcb976-cxb6c\" (UID: \"ee6a65b2-109d-44d8-a4e0-62ab28d392e4\") " pod="openstack/barbican-api-54cbbcb976-cxb6c" Nov 22 08:16:57 crc kubenswrapper[4789]: I1122 08:16:57.988700 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ee6a65b2-109d-44d8-a4e0-62ab28d392e4-logs\") pod \"barbican-api-54cbbcb976-cxb6c\" (UID: \"ee6a65b2-109d-44d8-a4e0-62ab28d392e4\") " pod="openstack/barbican-api-54cbbcb976-cxb6c" Nov 22 08:16:57 crc kubenswrapper[4789]: I1122 08:16:57.988802 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ee6a65b2-109d-44d8-a4e0-62ab28d392e4-config-data-custom\") pod \"barbican-api-54cbbcb976-cxb6c\" (UID: \"ee6a65b2-109d-44d8-a4e0-62ab28d392e4\") " pod="openstack/barbican-api-54cbbcb976-cxb6c" Nov 22 08:16:57 crc kubenswrapper[4789]: I1122 08:16:57.988955 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee6a65b2-109d-44d8-a4e0-62ab28d392e4-config-data\") pod \"barbican-api-54cbbcb976-cxb6c\" (UID: \"ee6a65b2-109d-44d8-a4e0-62ab28d392e4\") " pod="openstack/barbican-api-54cbbcb976-cxb6c" Nov 22 08:16:57 crc kubenswrapper[4789]: I1122 08:16:57.989016 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ee6a65b2-109d-44d8-a4e0-62ab28d392e4-internal-tls-certs\") pod \"barbican-api-54cbbcb976-cxb6c\" (UID: \"ee6a65b2-109d-44d8-a4e0-62ab28d392e4\") " pod="openstack/barbican-api-54cbbcb976-cxb6c" Nov 22 08:16:57 crc kubenswrapper[4789]: I1122 08:16:57.990722 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ee6a65b2-109d-44d8-a4e0-62ab28d392e4-logs\") pod \"barbican-api-54cbbcb976-cxb6c\" (UID: \"ee6a65b2-109d-44d8-a4e0-62ab28d392e4\") " pod="openstack/barbican-api-54cbbcb976-cxb6c" Nov 22 08:16:57 crc kubenswrapper[4789]: I1122 08:16:57.998563 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ee6a65b2-109d-44d8-a4e0-62ab28d392e4-config-data-custom\") pod \"barbican-api-54cbbcb976-cxb6c\" (UID: \"ee6a65b2-109d-44d8-a4e0-62ab28d392e4\") " pod="openstack/barbican-api-54cbbcb976-cxb6c" Nov 22 08:16:58 crc kubenswrapper[4789]: I1122 08:16:58.002469 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee6a65b2-109d-44d8-a4e0-62ab28d392e4-combined-ca-bundle\") pod \"barbican-api-54cbbcb976-cxb6c\" (UID: \"ee6a65b2-109d-44d8-a4e0-62ab28d392e4\") " pod="openstack/barbican-api-54cbbcb976-cxb6c" Nov 22 08:16:58 crc kubenswrapper[4789]: I1122 08:16:58.003744 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ee6a65b2-109d-44d8-a4e0-62ab28d392e4-internal-tls-certs\") pod \"barbican-api-54cbbcb976-cxb6c\" (UID: \"ee6a65b2-109d-44d8-a4e0-62ab28d392e4\") " pod="openstack/barbican-api-54cbbcb976-cxb6c" Nov 22 08:16:58 crc kubenswrapper[4789]: I1122 08:16:58.004210 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ee6a65b2-109d-44d8-a4e0-62ab28d392e4-public-tls-certs\") pod \"barbican-api-54cbbcb976-cxb6c\" (UID: \"ee6a65b2-109d-44d8-a4e0-62ab28d392e4\") " pod="openstack/barbican-api-54cbbcb976-cxb6c" Nov 22 08:16:58 crc kubenswrapper[4789]: I1122 08:16:58.008636 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee6a65b2-109d-44d8-a4e0-62ab28d392e4-config-data\") pod \"barbican-api-54cbbcb976-cxb6c\" (UID: \"ee6a65b2-109d-44d8-a4e0-62ab28d392e4\") " pod="openstack/barbican-api-54cbbcb976-cxb6c" Nov 22 08:16:58 crc kubenswrapper[4789]: I1122 08:16:58.011325 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g9rps\" (UniqueName: \"kubernetes.io/projected/ee6a65b2-109d-44d8-a4e0-62ab28d392e4-kube-api-access-g9rps\") pod \"barbican-api-54cbbcb976-cxb6c\" (UID: \"ee6a65b2-109d-44d8-a4e0-62ab28d392e4\") " pod="openstack/barbican-api-54cbbcb976-cxb6c" Nov 22 08:16:58 crc kubenswrapper[4789]: I1122 08:16:58.152254 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-54cbbcb976-cxb6c" Nov 22 08:16:58 crc kubenswrapper[4789]: I1122 08:16:58.248320 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 22 08:16:58 crc kubenswrapper[4789]: I1122 08:16:58.452981 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-5b56df9788-wsmhq" podUID="56fd3074-a15c-4024-baf9-f784a50e14a5" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.140:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.140:8443: connect: connection refused" Nov 22 08:16:58 crc kubenswrapper[4789]: I1122 08:16:58.681228 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-54cbbcb976-cxb6c"] Nov 22 08:16:58 crc kubenswrapper[4789]: I1122 08:16:58.921263 4789 generic.go:334] "Generic (PLEG): container finished" podID="56fd3074-a15c-4024-baf9-f784a50e14a5" containerID="2e99dc56ab31ddc560842a39626551056b3b1db94ece09fb54c781cd0d891832" exitCode=0 Nov 22 08:16:58 crc kubenswrapper[4789]: I1122 08:16:58.921302 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5b56df9788-wsmhq" event={"ID":"56fd3074-a15c-4024-baf9-f784a50e14a5","Type":"ContainerDied","Data":"2e99dc56ab31ddc560842a39626551056b3b1db94ece09fb54c781cd0d891832"} Nov 22 08:16:58 crc kubenswrapper[4789]: I1122 08:16:58.923308 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-54cbbcb976-cxb6c" event={"ID":"ee6a65b2-109d-44d8-a4e0-62ab28d392e4","Type":"ContainerStarted","Data":"332932402b6ceae06b5e5d5cd130fcbc00419ce5b5411d8500faba5773213c43"} Nov 22 08:16:58 crc kubenswrapper[4789]: I1122 08:16:58.924720 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"27920801-df7f-48f4-8830-6b190dea7234","Type":"ContainerStarted","Data":"570debd40643361dab7e32451b90270a09281abd7cde0020f4c74598451e4339"} Nov 22 08:16:58 crc kubenswrapper[4789]: I1122 08:16:58.927440 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"de991bf6-8f96-4c99-bf7a-b8909c0c80d2","Type":"ContainerStarted","Data":"6168d229a82e8b6c67cebc486d47d83616de74ed1dcc8f74ca5a8473c64386f3"} Nov 22 08:16:59 crc kubenswrapper[4789]: I1122 08:16:59.011272 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6976474df8-jfq2h" Nov 22 08:16:59 crc kubenswrapper[4789]: I1122 08:16:59.081407 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6976474df8-jfq2h" Nov 22 08:16:59 crc kubenswrapper[4789]: I1122 08:16:59.945377 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-54cbbcb976-cxb6c" event={"ID":"ee6a65b2-109d-44d8-a4e0-62ab28d392e4","Type":"ContainerStarted","Data":"7148be946217dc772db3065285ecf8521ff440d9b4df65e42a5aa53f9d636900"} Nov 22 08:16:59 crc kubenswrapper[4789]: I1122 08:16:59.945914 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-54cbbcb976-cxb6c" event={"ID":"ee6a65b2-109d-44d8-a4e0-62ab28d392e4","Type":"ContainerStarted","Data":"57a6e0a5fd603dd2ceca9727488b0d861fbcca647087059fd30d938a478826bd"} Nov 22 08:16:59 crc kubenswrapper[4789]: I1122 08:16:59.947111 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-54cbbcb976-cxb6c" Nov 22 08:16:59 crc kubenswrapper[4789]: I1122 08:16:59.947142 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-54cbbcb976-cxb6c" Nov 22 08:16:59 crc kubenswrapper[4789]: I1122 08:16:59.952538 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"27920801-df7f-48f4-8830-6b190dea7234","Type":"ContainerStarted","Data":"2260a84f51c6bf5437573b6f16ffe4d5c1289bc16585ccec1e13ba11e1bdf716"} Nov 22 08:16:59 crc kubenswrapper[4789]: I1122 08:16:59.955006 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 22 08:16:59 crc kubenswrapper[4789]: I1122 08:16:59.978013 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-54cbbcb976-cxb6c" podStartSLOduration=2.977997311 podStartE2EDuration="2.977997311s" podCreationTimestamp="2025-11-22 08:16:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:16:59.973145498 +0000 UTC m=+1234.207545771" watchObservedRunningTime="2025-11-22 08:16:59.977997311 +0000 UTC m=+1234.212397584" Nov 22 08:16:59 crc kubenswrapper[4789]: I1122 08:16:59.991040 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Nov 22 08:17:00 crc kubenswrapper[4789]: I1122 08:17:00.059496 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.8993765040000001 podStartE2EDuration="7.059472716s" podCreationTimestamp="2025-11-22 08:16:53 +0000 UTC" firstStartedPulling="2025-11-22 08:16:54.574825941 +0000 UTC m=+1228.809226214" lastFinishedPulling="2025-11-22 08:16:59.734922163 +0000 UTC m=+1233.969322426" observedRunningTime="2025-11-22 08:17:00.030077253 +0000 UTC m=+1234.264477536" watchObservedRunningTime="2025-11-22 08:17:00.059472716 +0000 UTC m=+1234.293872989" Nov 22 08:17:00 crc kubenswrapper[4789]: I1122 08:17:00.101705 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-775457b975-x8hvd" Nov 22 08:17:00 crc kubenswrapper[4789]: I1122 08:17:00.278071 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-66f4bdbdb7-tzwtz"] Nov 22 08:17:00 crc kubenswrapper[4789]: I1122 08:17:00.278331 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-66f4bdbdb7-tzwtz" podUID="17e7e1f9-63dc-4917-b15c-c25cfca9ee85" containerName="dnsmasq-dns" containerID="cri-o://da986e372351234f811d9c57b91629f3b44dd049dff52d99223a0d6584a78bfc" gracePeriod=10 Nov 22 08:17:00 crc kubenswrapper[4789]: I1122 08:17:00.581734 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Nov 22 08:17:00 crc kubenswrapper[4789]: I1122 08:17:00.963810 4789 generic.go:334] "Generic (PLEG): container finished" podID="17e7e1f9-63dc-4917-b15c-c25cfca9ee85" containerID="da986e372351234f811d9c57b91629f3b44dd049dff52d99223a0d6584a78bfc" exitCode=0 Nov 22 08:17:00 crc kubenswrapper[4789]: I1122 08:17:00.963876 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-66f4bdbdb7-tzwtz" event={"ID":"17e7e1f9-63dc-4917-b15c-c25cfca9ee85","Type":"ContainerDied","Data":"da986e372351234f811d9c57b91629f3b44dd049dff52d99223a0d6584a78bfc"} Nov 22 08:17:00 crc kubenswrapper[4789]: I1122 08:17:00.966193 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"27920801-df7f-48f4-8830-6b190dea7234","Type":"ContainerStarted","Data":"76b3070f56d685ba5c41f29e150547da2d25c918f004224f0cfa58063448698d"} Nov 22 08:17:00 crc kubenswrapper[4789]: I1122 08:17:00.966532 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Nov 22 08:17:00 crc kubenswrapper[4789]: I1122 08:17:00.968450 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"de991bf6-8f96-4c99-bf7a-b8909c0c80d2","Type":"ContainerStarted","Data":"4b4d49b0d5e06a9d2af814c8237d522b72d09606a55f523f315c307cfa7163bf"} Nov 22 08:17:00 crc kubenswrapper[4789]: I1122 08:17:00.985438 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.985418009 podStartE2EDuration="4.985418009s" podCreationTimestamp="2025-11-22 08:16:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:17:00.983343283 +0000 UTC m=+1235.217743566" watchObservedRunningTime="2025-11-22 08:17:00.985418009 +0000 UTC m=+1235.219818282" Nov 22 08:17:01 crc kubenswrapper[4789]: I1122 08:17:01.023734 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 22 08:17:01 crc kubenswrapper[4789]: I1122 08:17:01.459643 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-66f4bdbdb7-tzwtz" Nov 22 08:17:01 crc kubenswrapper[4789]: I1122 08:17:01.594877 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/17e7e1f9-63dc-4917-b15c-c25cfca9ee85-dns-svc\") pod \"17e7e1f9-63dc-4917-b15c-c25cfca9ee85\" (UID: \"17e7e1f9-63dc-4917-b15c-c25cfca9ee85\") " Nov 22 08:17:01 crc kubenswrapper[4789]: I1122 08:17:01.594938 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/17e7e1f9-63dc-4917-b15c-c25cfca9ee85-ovsdbserver-nb\") pod \"17e7e1f9-63dc-4917-b15c-c25cfca9ee85\" (UID: \"17e7e1f9-63dc-4917-b15c-c25cfca9ee85\") " Nov 22 08:17:01 crc kubenswrapper[4789]: I1122 08:17:01.595016 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6lj5x\" (UniqueName: \"kubernetes.io/projected/17e7e1f9-63dc-4917-b15c-c25cfca9ee85-kube-api-access-6lj5x\") pod \"17e7e1f9-63dc-4917-b15c-c25cfca9ee85\" (UID: \"17e7e1f9-63dc-4917-b15c-c25cfca9ee85\") " Nov 22 08:17:01 crc kubenswrapper[4789]: I1122 08:17:01.595174 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/17e7e1f9-63dc-4917-b15c-c25cfca9ee85-ovsdbserver-sb\") pod \"17e7e1f9-63dc-4917-b15c-c25cfca9ee85\" (UID: \"17e7e1f9-63dc-4917-b15c-c25cfca9ee85\") " Nov 22 08:17:01 crc kubenswrapper[4789]: I1122 08:17:01.595199 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17e7e1f9-63dc-4917-b15c-c25cfca9ee85-config\") pod \"17e7e1f9-63dc-4917-b15c-c25cfca9ee85\" (UID: \"17e7e1f9-63dc-4917-b15c-c25cfca9ee85\") " Nov 22 08:17:01 crc kubenswrapper[4789]: I1122 08:17:01.603963 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17e7e1f9-63dc-4917-b15c-c25cfca9ee85-kube-api-access-6lj5x" (OuterVolumeSpecName: "kube-api-access-6lj5x") pod "17e7e1f9-63dc-4917-b15c-c25cfca9ee85" (UID: "17e7e1f9-63dc-4917-b15c-c25cfca9ee85"). InnerVolumeSpecName "kube-api-access-6lj5x". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:17:01 crc kubenswrapper[4789]: I1122 08:17:01.648291 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/17e7e1f9-63dc-4917-b15c-c25cfca9ee85-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "17e7e1f9-63dc-4917-b15c-c25cfca9ee85" (UID: "17e7e1f9-63dc-4917-b15c-c25cfca9ee85"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:17:01 crc kubenswrapper[4789]: I1122 08:17:01.648303 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/17e7e1f9-63dc-4917-b15c-c25cfca9ee85-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "17e7e1f9-63dc-4917-b15c-c25cfca9ee85" (UID: "17e7e1f9-63dc-4917-b15c-c25cfca9ee85"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:17:01 crc kubenswrapper[4789]: I1122 08:17:01.651985 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/17e7e1f9-63dc-4917-b15c-c25cfca9ee85-config" (OuterVolumeSpecName: "config") pod "17e7e1f9-63dc-4917-b15c-c25cfca9ee85" (UID: "17e7e1f9-63dc-4917-b15c-c25cfca9ee85"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:17:01 crc kubenswrapper[4789]: I1122 08:17:01.661425 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/17e7e1f9-63dc-4917-b15c-c25cfca9ee85-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "17e7e1f9-63dc-4917-b15c-c25cfca9ee85" (UID: "17e7e1f9-63dc-4917-b15c-c25cfca9ee85"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:17:01 crc kubenswrapper[4789]: I1122 08:17:01.697470 4789 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/17e7e1f9-63dc-4917-b15c-c25cfca9ee85-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 22 08:17:01 crc kubenswrapper[4789]: I1122 08:17:01.697519 4789 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/17e7e1f9-63dc-4917-b15c-c25cfca9ee85-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 22 08:17:01 crc kubenswrapper[4789]: I1122 08:17:01.697534 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6lj5x\" (UniqueName: \"kubernetes.io/projected/17e7e1f9-63dc-4917-b15c-c25cfca9ee85-kube-api-access-6lj5x\") on node \"crc\" DevicePath \"\"" Nov 22 08:17:01 crc kubenswrapper[4789]: I1122 08:17:01.697548 4789 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/17e7e1f9-63dc-4917-b15c-c25cfca9ee85-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 22 08:17:01 crc kubenswrapper[4789]: I1122 08:17:01.697559 4789 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17e7e1f9-63dc-4917-b15c-c25cfca9ee85-config\") on node \"crc\" DevicePath \"\"" Nov 22 08:17:01 crc kubenswrapper[4789]: I1122 08:17:01.978476 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-66f4bdbdb7-tzwtz" Nov 22 08:17:01 crc kubenswrapper[4789]: I1122 08:17:01.978501 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-66f4bdbdb7-tzwtz" event={"ID":"17e7e1f9-63dc-4917-b15c-c25cfca9ee85","Type":"ContainerDied","Data":"166a96122b7fd3d01ff1eaedc0e61753faac0db24d9c0ccb069993333588ad5f"} Nov 22 08:17:01 crc kubenswrapper[4789]: I1122 08:17:01.978554 4789 scope.go:117] "RemoveContainer" containerID="da986e372351234f811d9c57b91629f3b44dd049dff52d99223a0d6584a78bfc" Nov 22 08:17:01 crc kubenswrapper[4789]: I1122 08:17:01.978620 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="e1b1a2ae-3c67-4ada-b478-92ee0a9bb30d" containerName="cinder-scheduler" containerID="cri-o://0994c405912329d7244b21305bd0c8f8ec9e21db14d6162835da6b124c7b3674" gracePeriod=30 Nov 22 08:17:01 crc kubenswrapper[4789]: I1122 08:17:01.978688 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="e1b1a2ae-3c67-4ada-b478-92ee0a9bb30d" containerName="probe" containerID="cri-o://de49ff9e3613638a69b7d9cc20c5d4372e86c6ad0721d1b9f28300cccc0089ed" gracePeriod=30 Nov 22 08:17:02 crc kubenswrapper[4789]: I1122 08:17:02.005292 4789 scope.go:117] "RemoveContainer" containerID="849eea141aca67d8500974a142cb8c034c08a66d073fdef2c2686355a607fda1" Nov 22 08:17:02 crc kubenswrapper[4789]: I1122 08:17:02.021130 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-66f4bdbdb7-tzwtz"] Nov 22 08:17:02 crc kubenswrapper[4789]: I1122 08:17:02.034185 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-66f4bdbdb7-tzwtz"] Nov 22 08:17:02 crc kubenswrapper[4789]: I1122 08:17:02.988177 4789 generic.go:334] "Generic (PLEG): container finished" podID="e1b1a2ae-3c67-4ada-b478-92ee0a9bb30d" containerID="de49ff9e3613638a69b7d9cc20c5d4372e86c6ad0721d1b9f28300cccc0089ed" exitCode=0 Nov 22 08:17:02 crc kubenswrapper[4789]: I1122 08:17:02.988230 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e1b1a2ae-3c67-4ada-b478-92ee0a9bb30d","Type":"ContainerDied","Data":"de49ff9e3613638a69b7d9cc20c5d4372e86c6ad0721d1b9f28300cccc0089ed"} Nov 22 08:17:03 crc kubenswrapper[4789]: I1122 08:17:03.990730 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="17e7e1f9-63dc-4917-b15c-c25cfca9ee85" path="/var/lib/kubelet/pods/17e7e1f9-63dc-4917-b15c-c25cfca9ee85/volumes" Nov 22 08:17:05 crc kubenswrapper[4789]: I1122 08:17:05.020104 4789 generic.go:334] "Generic (PLEG): container finished" podID="e1b1a2ae-3c67-4ada-b478-92ee0a9bb30d" containerID="0994c405912329d7244b21305bd0c8f8ec9e21db14d6162835da6b124c7b3674" exitCode=0 Nov 22 08:17:05 crc kubenswrapper[4789]: I1122 08:17:05.020176 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e1b1a2ae-3c67-4ada-b478-92ee0a9bb30d","Type":"ContainerDied","Data":"0994c405912329d7244b21305bd0c8f8ec9e21db14d6162835da6b124c7b3674"} Nov 22 08:17:05 crc kubenswrapper[4789]: I1122 08:17:05.127301 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 22 08:17:05 crc kubenswrapper[4789]: I1122 08:17:05.278911 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e1b1a2ae-3c67-4ada-b478-92ee0a9bb30d-config-data-custom\") pod \"e1b1a2ae-3c67-4ada-b478-92ee0a9bb30d\" (UID: \"e1b1a2ae-3c67-4ada-b478-92ee0a9bb30d\") " Nov 22 08:17:05 crc kubenswrapper[4789]: I1122 08:17:05.279052 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1b1a2ae-3c67-4ada-b478-92ee0a9bb30d-combined-ca-bundle\") pod \"e1b1a2ae-3c67-4ada-b478-92ee0a9bb30d\" (UID: \"e1b1a2ae-3c67-4ada-b478-92ee0a9bb30d\") " Nov 22 08:17:05 crc kubenswrapper[4789]: I1122 08:17:05.279111 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1b1a2ae-3c67-4ada-b478-92ee0a9bb30d-config-data\") pod \"e1b1a2ae-3c67-4ada-b478-92ee0a9bb30d\" (UID: \"e1b1a2ae-3c67-4ada-b478-92ee0a9bb30d\") " Nov 22 08:17:05 crc kubenswrapper[4789]: I1122 08:17:05.279144 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e1b1a2ae-3c67-4ada-b478-92ee0a9bb30d-etc-machine-id\") pod \"e1b1a2ae-3c67-4ada-b478-92ee0a9bb30d\" (UID: \"e1b1a2ae-3c67-4ada-b478-92ee0a9bb30d\") " Nov 22 08:17:05 crc kubenswrapper[4789]: I1122 08:17:05.279190 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e1b1a2ae-3c67-4ada-b478-92ee0a9bb30d-scripts\") pod \"e1b1a2ae-3c67-4ada-b478-92ee0a9bb30d\" (UID: \"e1b1a2ae-3c67-4ada-b478-92ee0a9bb30d\") " Nov 22 08:17:05 crc kubenswrapper[4789]: I1122 08:17:05.279250 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tqlfl\" (UniqueName: \"kubernetes.io/projected/e1b1a2ae-3c67-4ada-b478-92ee0a9bb30d-kube-api-access-tqlfl\") pod \"e1b1a2ae-3c67-4ada-b478-92ee0a9bb30d\" (UID: \"e1b1a2ae-3c67-4ada-b478-92ee0a9bb30d\") " Nov 22 08:17:05 crc kubenswrapper[4789]: I1122 08:17:05.279324 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e1b1a2ae-3c67-4ada-b478-92ee0a9bb30d-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "e1b1a2ae-3c67-4ada-b478-92ee0a9bb30d" (UID: "e1b1a2ae-3c67-4ada-b478-92ee0a9bb30d"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 08:17:05 crc kubenswrapper[4789]: I1122 08:17:05.279645 4789 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e1b1a2ae-3c67-4ada-b478-92ee0a9bb30d-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 22 08:17:05 crc kubenswrapper[4789]: I1122 08:17:05.284702 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1b1a2ae-3c67-4ada-b478-92ee0a9bb30d-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "e1b1a2ae-3c67-4ada-b478-92ee0a9bb30d" (UID: "e1b1a2ae-3c67-4ada-b478-92ee0a9bb30d"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:17:05 crc kubenswrapper[4789]: I1122 08:17:05.285592 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1b1a2ae-3c67-4ada-b478-92ee0a9bb30d-kube-api-access-tqlfl" (OuterVolumeSpecName: "kube-api-access-tqlfl") pod "e1b1a2ae-3c67-4ada-b478-92ee0a9bb30d" (UID: "e1b1a2ae-3c67-4ada-b478-92ee0a9bb30d"). InnerVolumeSpecName "kube-api-access-tqlfl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:17:05 crc kubenswrapper[4789]: I1122 08:17:05.286486 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1b1a2ae-3c67-4ada-b478-92ee0a9bb30d-scripts" (OuterVolumeSpecName: "scripts") pod "e1b1a2ae-3c67-4ada-b478-92ee0a9bb30d" (UID: "e1b1a2ae-3c67-4ada-b478-92ee0a9bb30d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:17:05 crc kubenswrapper[4789]: I1122 08:17:05.324614 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1b1a2ae-3c67-4ada-b478-92ee0a9bb30d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e1b1a2ae-3c67-4ada-b478-92ee0a9bb30d" (UID: "e1b1a2ae-3c67-4ada-b478-92ee0a9bb30d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:17:05 crc kubenswrapper[4789]: I1122 08:17:05.377867 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1b1a2ae-3c67-4ada-b478-92ee0a9bb30d-config-data" (OuterVolumeSpecName: "config-data") pod "e1b1a2ae-3c67-4ada-b478-92ee0a9bb30d" (UID: "e1b1a2ae-3c67-4ada-b478-92ee0a9bb30d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:17:05 crc kubenswrapper[4789]: I1122 08:17:05.386477 4789 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1b1a2ae-3c67-4ada-b478-92ee0a9bb30d-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:17:05 crc kubenswrapper[4789]: I1122 08:17:05.386507 4789 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e1b1a2ae-3c67-4ada-b478-92ee0a9bb30d-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:17:05 crc kubenswrapper[4789]: I1122 08:17:05.386518 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tqlfl\" (UniqueName: \"kubernetes.io/projected/e1b1a2ae-3c67-4ada-b478-92ee0a9bb30d-kube-api-access-tqlfl\") on node \"crc\" DevicePath \"\"" Nov 22 08:17:05 crc kubenswrapper[4789]: I1122 08:17:05.386528 4789 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e1b1a2ae-3c67-4ada-b478-92ee0a9bb30d-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 22 08:17:05 crc kubenswrapper[4789]: I1122 08:17:05.386536 4789 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1b1a2ae-3c67-4ada-b478-92ee0a9bb30d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:17:06 crc kubenswrapper[4789]: I1122 08:17:06.030278 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e1b1a2ae-3c67-4ada-b478-92ee0a9bb30d","Type":"ContainerDied","Data":"7c37e3313ff0e05a96d71fc2869b8fa55eb2be4d7d3c1d5459a19d0aa134398e"} Nov 22 08:17:06 crc kubenswrapper[4789]: I1122 08:17:06.030356 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 22 08:17:06 crc kubenswrapper[4789]: I1122 08:17:06.031132 4789 scope.go:117] "RemoveContainer" containerID="de49ff9e3613638a69b7d9cc20c5d4372e86c6ad0721d1b9f28300cccc0089ed" Nov 22 08:17:06 crc kubenswrapper[4789]: I1122 08:17:06.050140 4789 scope.go:117] "RemoveContainer" containerID="0994c405912329d7244b21305bd0c8f8ec9e21db14d6162835da6b124c7b3674" Nov 22 08:17:06 crc kubenswrapper[4789]: I1122 08:17:06.053447 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 22 08:17:06 crc kubenswrapper[4789]: I1122 08:17:06.061551 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 22 08:17:06 crc kubenswrapper[4789]: I1122 08:17:06.077277 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Nov 22 08:17:06 crc kubenswrapper[4789]: E1122 08:17:06.077679 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1b1a2ae-3c67-4ada-b478-92ee0a9bb30d" containerName="probe" Nov 22 08:17:06 crc kubenswrapper[4789]: I1122 08:17:06.077697 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1b1a2ae-3c67-4ada-b478-92ee0a9bb30d" containerName="probe" Nov 22 08:17:06 crc kubenswrapper[4789]: E1122 08:17:06.077714 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1b1a2ae-3c67-4ada-b478-92ee0a9bb30d" containerName="cinder-scheduler" Nov 22 08:17:06 crc kubenswrapper[4789]: I1122 08:17:06.077721 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1b1a2ae-3c67-4ada-b478-92ee0a9bb30d" containerName="cinder-scheduler" Nov 22 08:17:06 crc kubenswrapper[4789]: E1122 08:17:06.077747 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17e7e1f9-63dc-4917-b15c-c25cfca9ee85" containerName="init" Nov 22 08:17:06 crc kubenswrapper[4789]: I1122 08:17:06.077768 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="17e7e1f9-63dc-4917-b15c-c25cfca9ee85" containerName="init" Nov 22 08:17:06 crc kubenswrapper[4789]: E1122 08:17:06.077777 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17e7e1f9-63dc-4917-b15c-c25cfca9ee85" containerName="dnsmasq-dns" Nov 22 08:17:06 crc kubenswrapper[4789]: I1122 08:17:06.077782 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="17e7e1f9-63dc-4917-b15c-c25cfca9ee85" containerName="dnsmasq-dns" Nov 22 08:17:06 crc kubenswrapper[4789]: I1122 08:17:06.077947 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="17e7e1f9-63dc-4917-b15c-c25cfca9ee85" containerName="dnsmasq-dns" Nov 22 08:17:06 crc kubenswrapper[4789]: I1122 08:17:06.077959 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1b1a2ae-3c67-4ada-b478-92ee0a9bb30d" containerName="probe" Nov 22 08:17:06 crc kubenswrapper[4789]: I1122 08:17:06.077968 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1b1a2ae-3c67-4ada-b478-92ee0a9bb30d" containerName="cinder-scheduler" Nov 22 08:17:06 crc kubenswrapper[4789]: I1122 08:17:06.078966 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 22 08:17:06 crc kubenswrapper[4789]: I1122 08:17:06.081322 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Nov 22 08:17:06 crc kubenswrapper[4789]: I1122 08:17:06.084260 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 22 08:17:06 crc kubenswrapper[4789]: I1122 08:17:06.197822 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b24f838e-1881-4921-8f51-8bb32afc4177-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"b24f838e-1881-4921-8f51-8bb32afc4177\") " pod="openstack/cinder-scheduler-0" Nov 22 08:17:06 crc kubenswrapper[4789]: I1122 08:17:06.198148 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b24f838e-1881-4921-8f51-8bb32afc4177-scripts\") pod \"cinder-scheduler-0\" (UID: \"b24f838e-1881-4921-8f51-8bb32afc4177\") " pod="openstack/cinder-scheduler-0" Nov 22 08:17:06 crc kubenswrapper[4789]: I1122 08:17:06.198268 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dtvbf\" (UniqueName: \"kubernetes.io/projected/b24f838e-1881-4921-8f51-8bb32afc4177-kube-api-access-dtvbf\") pod \"cinder-scheduler-0\" (UID: \"b24f838e-1881-4921-8f51-8bb32afc4177\") " pod="openstack/cinder-scheduler-0" Nov 22 08:17:06 crc kubenswrapper[4789]: I1122 08:17:06.198400 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b24f838e-1881-4921-8f51-8bb32afc4177-config-data\") pod \"cinder-scheduler-0\" (UID: \"b24f838e-1881-4921-8f51-8bb32afc4177\") " pod="openstack/cinder-scheduler-0" Nov 22 08:17:06 crc kubenswrapper[4789]: I1122 08:17:06.198515 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b24f838e-1881-4921-8f51-8bb32afc4177-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"b24f838e-1881-4921-8f51-8bb32afc4177\") " pod="openstack/cinder-scheduler-0" Nov 22 08:17:06 crc kubenswrapper[4789]: I1122 08:17:06.198550 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b24f838e-1881-4921-8f51-8bb32afc4177-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"b24f838e-1881-4921-8f51-8bb32afc4177\") " pod="openstack/cinder-scheduler-0" Nov 22 08:17:06 crc kubenswrapper[4789]: I1122 08:17:06.300466 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dtvbf\" (UniqueName: \"kubernetes.io/projected/b24f838e-1881-4921-8f51-8bb32afc4177-kube-api-access-dtvbf\") pod \"cinder-scheduler-0\" (UID: \"b24f838e-1881-4921-8f51-8bb32afc4177\") " pod="openstack/cinder-scheduler-0" Nov 22 08:17:06 crc kubenswrapper[4789]: I1122 08:17:06.300533 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b24f838e-1881-4921-8f51-8bb32afc4177-config-data\") pod \"cinder-scheduler-0\" (UID: \"b24f838e-1881-4921-8f51-8bb32afc4177\") " pod="openstack/cinder-scheduler-0" Nov 22 08:17:06 crc kubenswrapper[4789]: I1122 08:17:06.300573 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b24f838e-1881-4921-8f51-8bb32afc4177-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"b24f838e-1881-4921-8f51-8bb32afc4177\") " pod="openstack/cinder-scheduler-0" Nov 22 08:17:06 crc kubenswrapper[4789]: I1122 08:17:06.300602 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b24f838e-1881-4921-8f51-8bb32afc4177-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"b24f838e-1881-4921-8f51-8bb32afc4177\") " pod="openstack/cinder-scheduler-0" Nov 22 08:17:06 crc kubenswrapper[4789]: I1122 08:17:06.300705 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b24f838e-1881-4921-8f51-8bb32afc4177-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"b24f838e-1881-4921-8f51-8bb32afc4177\") " pod="openstack/cinder-scheduler-0" Nov 22 08:17:06 crc kubenswrapper[4789]: I1122 08:17:06.300765 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b24f838e-1881-4921-8f51-8bb32afc4177-scripts\") pod \"cinder-scheduler-0\" (UID: \"b24f838e-1881-4921-8f51-8bb32afc4177\") " pod="openstack/cinder-scheduler-0" Nov 22 08:17:06 crc kubenswrapper[4789]: I1122 08:17:06.300925 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b24f838e-1881-4921-8f51-8bb32afc4177-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"b24f838e-1881-4921-8f51-8bb32afc4177\") " pod="openstack/cinder-scheduler-0" Nov 22 08:17:06 crc kubenswrapper[4789]: I1122 08:17:06.306971 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b24f838e-1881-4921-8f51-8bb32afc4177-scripts\") pod \"cinder-scheduler-0\" (UID: \"b24f838e-1881-4921-8f51-8bb32afc4177\") " pod="openstack/cinder-scheduler-0" Nov 22 08:17:06 crc kubenswrapper[4789]: I1122 08:17:06.307113 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b24f838e-1881-4921-8f51-8bb32afc4177-config-data\") pod \"cinder-scheduler-0\" (UID: \"b24f838e-1881-4921-8f51-8bb32afc4177\") " pod="openstack/cinder-scheduler-0" Nov 22 08:17:06 crc kubenswrapper[4789]: I1122 08:17:06.307212 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b24f838e-1881-4921-8f51-8bb32afc4177-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"b24f838e-1881-4921-8f51-8bb32afc4177\") " pod="openstack/cinder-scheduler-0" Nov 22 08:17:06 crc kubenswrapper[4789]: I1122 08:17:06.307624 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b24f838e-1881-4921-8f51-8bb32afc4177-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"b24f838e-1881-4921-8f51-8bb32afc4177\") " pod="openstack/cinder-scheduler-0" Nov 22 08:17:06 crc kubenswrapper[4789]: I1122 08:17:06.319450 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dtvbf\" (UniqueName: \"kubernetes.io/projected/b24f838e-1881-4921-8f51-8bb32afc4177-kube-api-access-dtvbf\") pod \"cinder-scheduler-0\" (UID: \"b24f838e-1881-4921-8f51-8bb32afc4177\") " pod="openstack/cinder-scheduler-0" Nov 22 08:17:06 crc kubenswrapper[4789]: I1122 08:17:06.398738 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 22 08:17:06 crc kubenswrapper[4789]: W1122 08:17:06.850150 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb24f838e_1881_4921_8f51_8bb32afc4177.slice/crio-214fa152ae875157b6c241c74447cad340e07c4ef53f99b602fdd4954f57f4fb WatchSource:0}: Error finding container 214fa152ae875157b6c241c74447cad340e07c4ef53f99b602fdd4954f57f4fb: Status 404 returned error can't find the container with id 214fa152ae875157b6c241c74447cad340e07c4ef53f99b602fdd4954f57f4fb Nov 22 08:17:06 crc kubenswrapper[4789]: I1122 08:17:06.851553 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 22 08:17:07 crc kubenswrapper[4789]: I1122 08:17:07.056658 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"b24f838e-1881-4921-8f51-8bb32afc4177","Type":"ContainerStarted","Data":"214fa152ae875157b6c241c74447cad340e07c4ef53f99b602fdd4954f57f4fb"} Nov 22 08:17:07 crc kubenswrapper[4789]: I1122 08:17:07.984067 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e1b1a2ae-3c67-4ada-b478-92ee0a9bb30d" path="/var/lib/kubelet/pods/e1b1a2ae-3c67-4ada-b478-92ee0a9bb30d/volumes" Nov 22 08:17:08 crc kubenswrapper[4789]: I1122 08:17:08.071314 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"b24f838e-1881-4921-8f51-8bb32afc4177","Type":"ContainerStarted","Data":"7a19b412bf5ecd94379c21a54e195c48806d49dc9b088e4bba48dfe7f6737c29"} Nov 22 08:17:08 crc kubenswrapper[4789]: I1122 08:17:08.071368 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"b24f838e-1881-4921-8f51-8bb32afc4177","Type":"ContainerStarted","Data":"7550582bcca169482a4abe9f92322c55af7e201346e8dd68f9faba77e3571010"} Nov 22 08:17:08 crc kubenswrapper[4789]: I1122 08:17:08.452693 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-5b56df9788-wsmhq" podUID="56fd3074-a15c-4024-baf9-f784a50e14a5" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.140:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.140:8443: connect: connection refused" Nov 22 08:17:09 crc kubenswrapper[4789]: I1122 08:17:09.189103 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-7846c8b766-kj5xm" Nov 22 08:17:09 crc kubenswrapper[4789]: I1122 08:17:09.211076 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.211056511 podStartE2EDuration="3.211056511s" podCreationTimestamp="2025-11-22 08:17:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:17:08.100051144 +0000 UTC m=+1242.334451437" watchObservedRunningTime="2025-11-22 08:17:09.211056511 +0000 UTC m=+1243.445456784" Nov 22 08:17:09 crc kubenswrapper[4789]: I1122 08:17:09.681548 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-7846c8b766-kj5xm" Nov 22 08:17:09 crc kubenswrapper[4789]: I1122 08:17:09.864599 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-54cbbcb976-cxb6c" Nov 22 08:17:09 crc kubenswrapper[4789]: I1122 08:17:09.959318 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-5c64d54575-bxqxv" Nov 22 08:17:10 crc kubenswrapper[4789]: I1122 08:17:10.103282 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Nov 22 08:17:10 crc kubenswrapper[4789]: I1122 08:17:10.229829 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-54cbbcb976-cxb6c" Nov 22 08:17:10 crc kubenswrapper[4789]: I1122 08:17:10.303408 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-6976474df8-jfq2h"] Nov 22 08:17:10 crc kubenswrapper[4789]: I1122 08:17:10.303633 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-6976474df8-jfq2h" podUID="e382d956-58e8-4c3b-abd9-0732d19bc35e" containerName="barbican-api-log" containerID="cri-o://612d0ac99e38a37b8ca19ff88131467979ce665edafa1d0125ab68a49d7e8682" gracePeriod=30 Nov 22 08:17:10 crc kubenswrapper[4789]: I1122 08:17:10.304877 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-6976474df8-jfq2h" podUID="e382d956-58e8-4c3b-abd9-0732d19bc35e" containerName="barbican-api" containerID="cri-o://753f281ef5ab8c90d5710f4a0e8721ac8c9ebc938a789a976d5a6573e0dbcc20" gracePeriod=30 Nov 22 08:17:11 crc kubenswrapper[4789]: I1122 08:17:11.100614 4789 generic.go:334] "Generic (PLEG): container finished" podID="e382d956-58e8-4c3b-abd9-0732d19bc35e" containerID="612d0ac99e38a37b8ca19ff88131467979ce665edafa1d0125ab68a49d7e8682" exitCode=143 Nov 22 08:17:11 crc kubenswrapper[4789]: I1122 08:17:11.101917 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6976474df8-jfq2h" event={"ID":"e382d956-58e8-4c3b-abd9-0732d19bc35e","Type":"ContainerDied","Data":"612d0ac99e38a37b8ca19ff88131467979ce665edafa1d0125ab68a49d7e8682"} Nov 22 08:17:11 crc kubenswrapper[4789]: I1122 08:17:11.399557 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Nov 22 08:17:13 crc kubenswrapper[4789]: I1122 08:17:13.438664 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6976474df8-jfq2h" podUID="e382d956-58e8-4c3b-abd9-0732d19bc35e" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.148:9311/healthcheck\": read tcp 10.217.0.2:41244->10.217.0.148:9311: read: connection reset by peer" Nov 22 08:17:13 crc kubenswrapper[4789]: I1122 08:17:13.444620 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6976474df8-jfq2h" podUID="e382d956-58e8-4c3b-abd9-0732d19bc35e" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.148:9311/healthcheck\": read tcp 10.217.0.2:41234->10.217.0.148:9311: read: connection reset by peer" Nov 22 08:17:13 crc kubenswrapper[4789]: I1122 08:17:13.864029 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6976474df8-jfq2h" Nov 22 08:17:14 crc kubenswrapper[4789]: I1122 08:17:14.055006 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e382d956-58e8-4c3b-abd9-0732d19bc35e-config-data\") pod \"e382d956-58e8-4c3b-abd9-0732d19bc35e\" (UID: \"e382d956-58e8-4c3b-abd9-0732d19bc35e\") " Nov 22 08:17:14 crc kubenswrapper[4789]: I1122 08:17:14.055138 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lw2d2\" (UniqueName: \"kubernetes.io/projected/e382d956-58e8-4c3b-abd9-0732d19bc35e-kube-api-access-lw2d2\") pod \"e382d956-58e8-4c3b-abd9-0732d19bc35e\" (UID: \"e382d956-58e8-4c3b-abd9-0732d19bc35e\") " Nov 22 08:17:14 crc kubenswrapper[4789]: I1122 08:17:14.055199 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e382d956-58e8-4c3b-abd9-0732d19bc35e-combined-ca-bundle\") pod \"e382d956-58e8-4c3b-abd9-0732d19bc35e\" (UID: \"e382d956-58e8-4c3b-abd9-0732d19bc35e\") " Nov 22 08:17:14 crc kubenswrapper[4789]: I1122 08:17:14.055217 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e382d956-58e8-4c3b-abd9-0732d19bc35e-config-data-custom\") pod \"e382d956-58e8-4c3b-abd9-0732d19bc35e\" (UID: \"e382d956-58e8-4c3b-abd9-0732d19bc35e\") " Nov 22 08:17:14 crc kubenswrapper[4789]: I1122 08:17:14.055270 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e382d956-58e8-4c3b-abd9-0732d19bc35e-logs\") pod \"e382d956-58e8-4c3b-abd9-0732d19bc35e\" (UID: \"e382d956-58e8-4c3b-abd9-0732d19bc35e\") " Nov 22 08:17:14 crc kubenswrapper[4789]: I1122 08:17:14.056101 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e382d956-58e8-4c3b-abd9-0732d19bc35e-logs" (OuterVolumeSpecName: "logs") pod "e382d956-58e8-4c3b-abd9-0732d19bc35e" (UID: "e382d956-58e8-4c3b-abd9-0732d19bc35e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:17:14 crc kubenswrapper[4789]: I1122 08:17:14.064765 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e382d956-58e8-4c3b-abd9-0732d19bc35e-kube-api-access-lw2d2" (OuterVolumeSpecName: "kube-api-access-lw2d2") pod "e382d956-58e8-4c3b-abd9-0732d19bc35e" (UID: "e382d956-58e8-4c3b-abd9-0732d19bc35e"). InnerVolumeSpecName "kube-api-access-lw2d2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:17:14 crc kubenswrapper[4789]: I1122 08:17:14.064866 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e382d956-58e8-4c3b-abd9-0732d19bc35e-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "e382d956-58e8-4c3b-abd9-0732d19bc35e" (UID: "e382d956-58e8-4c3b-abd9-0732d19bc35e"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:17:14 crc kubenswrapper[4789]: I1122 08:17:14.084730 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e382d956-58e8-4c3b-abd9-0732d19bc35e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e382d956-58e8-4c3b-abd9-0732d19bc35e" (UID: "e382d956-58e8-4c3b-abd9-0732d19bc35e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:17:14 crc kubenswrapper[4789]: I1122 08:17:14.117366 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e382d956-58e8-4c3b-abd9-0732d19bc35e-config-data" (OuterVolumeSpecName: "config-data") pod "e382d956-58e8-4c3b-abd9-0732d19bc35e" (UID: "e382d956-58e8-4c3b-abd9-0732d19bc35e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:17:14 crc kubenswrapper[4789]: I1122 08:17:14.127535 4789 generic.go:334] "Generic (PLEG): container finished" podID="e382d956-58e8-4c3b-abd9-0732d19bc35e" containerID="753f281ef5ab8c90d5710f4a0e8721ac8c9ebc938a789a976d5a6573e0dbcc20" exitCode=0 Nov 22 08:17:14 crc kubenswrapper[4789]: I1122 08:17:14.127580 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6976474df8-jfq2h" event={"ID":"e382d956-58e8-4c3b-abd9-0732d19bc35e","Type":"ContainerDied","Data":"753f281ef5ab8c90d5710f4a0e8721ac8c9ebc938a789a976d5a6573e0dbcc20"} Nov 22 08:17:14 crc kubenswrapper[4789]: I1122 08:17:14.127593 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6976474df8-jfq2h" Nov 22 08:17:14 crc kubenswrapper[4789]: I1122 08:17:14.127607 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6976474df8-jfq2h" event={"ID":"e382d956-58e8-4c3b-abd9-0732d19bc35e","Type":"ContainerDied","Data":"8ea390169168c7b47bb8cac431625afc823ef69ac15d874d12be3d0b2f197034"} Nov 22 08:17:14 crc kubenswrapper[4789]: I1122 08:17:14.127668 4789 scope.go:117] "RemoveContainer" containerID="753f281ef5ab8c90d5710f4a0e8721ac8c9ebc938a789a976d5a6573e0dbcc20" Nov 22 08:17:14 crc kubenswrapper[4789]: I1122 08:17:14.156935 4789 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e382d956-58e8-4c3b-abd9-0732d19bc35e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:17:14 crc kubenswrapper[4789]: I1122 08:17:14.156959 4789 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e382d956-58e8-4c3b-abd9-0732d19bc35e-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 22 08:17:14 crc kubenswrapper[4789]: I1122 08:17:14.156968 4789 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e382d956-58e8-4c3b-abd9-0732d19bc35e-logs\") on node \"crc\" DevicePath \"\"" Nov 22 08:17:14 crc kubenswrapper[4789]: I1122 08:17:14.156978 4789 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e382d956-58e8-4c3b-abd9-0732d19bc35e-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:17:14 crc kubenswrapper[4789]: I1122 08:17:14.156987 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lw2d2\" (UniqueName: \"kubernetes.io/projected/e382d956-58e8-4c3b-abd9-0732d19bc35e-kube-api-access-lw2d2\") on node \"crc\" DevicePath \"\"" Nov 22 08:17:14 crc kubenswrapper[4789]: I1122 08:17:14.177695 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-6976474df8-jfq2h"] Nov 22 08:17:14 crc kubenswrapper[4789]: I1122 08:17:14.186279 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-6976474df8-jfq2h"] Nov 22 08:17:14 crc kubenswrapper[4789]: I1122 08:17:14.187844 4789 scope.go:117] "RemoveContainer" containerID="612d0ac99e38a37b8ca19ff88131467979ce665edafa1d0125ab68a49d7e8682" Nov 22 08:17:14 crc kubenswrapper[4789]: I1122 08:17:14.209978 4789 scope.go:117] "RemoveContainer" containerID="753f281ef5ab8c90d5710f4a0e8721ac8c9ebc938a789a976d5a6573e0dbcc20" Nov 22 08:17:14 crc kubenswrapper[4789]: E1122 08:17:14.210503 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"753f281ef5ab8c90d5710f4a0e8721ac8c9ebc938a789a976d5a6573e0dbcc20\": container with ID starting with 753f281ef5ab8c90d5710f4a0e8721ac8c9ebc938a789a976d5a6573e0dbcc20 not found: ID does not exist" containerID="753f281ef5ab8c90d5710f4a0e8721ac8c9ebc938a789a976d5a6573e0dbcc20" Nov 22 08:17:14 crc kubenswrapper[4789]: I1122 08:17:14.210540 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"753f281ef5ab8c90d5710f4a0e8721ac8c9ebc938a789a976d5a6573e0dbcc20"} err="failed to get container status \"753f281ef5ab8c90d5710f4a0e8721ac8c9ebc938a789a976d5a6573e0dbcc20\": rpc error: code = NotFound desc = could not find container \"753f281ef5ab8c90d5710f4a0e8721ac8c9ebc938a789a976d5a6573e0dbcc20\": container with ID starting with 753f281ef5ab8c90d5710f4a0e8721ac8c9ebc938a789a976d5a6573e0dbcc20 not found: ID does not exist" Nov 22 08:17:14 crc kubenswrapper[4789]: I1122 08:17:14.210566 4789 scope.go:117] "RemoveContainer" containerID="612d0ac99e38a37b8ca19ff88131467979ce665edafa1d0125ab68a49d7e8682" Nov 22 08:17:14 crc kubenswrapper[4789]: E1122 08:17:14.210795 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"612d0ac99e38a37b8ca19ff88131467979ce665edafa1d0125ab68a49d7e8682\": container with ID starting with 612d0ac99e38a37b8ca19ff88131467979ce665edafa1d0125ab68a49d7e8682 not found: ID does not exist" containerID="612d0ac99e38a37b8ca19ff88131467979ce665edafa1d0125ab68a49d7e8682" Nov 22 08:17:14 crc kubenswrapper[4789]: I1122 08:17:14.210825 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"612d0ac99e38a37b8ca19ff88131467979ce665edafa1d0125ab68a49d7e8682"} err="failed to get container status \"612d0ac99e38a37b8ca19ff88131467979ce665edafa1d0125ab68a49d7e8682\": rpc error: code = NotFound desc = could not find container \"612d0ac99e38a37b8ca19ff88131467979ce665edafa1d0125ab68a49d7e8682\": container with ID starting with 612d0ac99e38a37b8ca19ff88131467979ce665edafa1d0125ab68a49d7e8682 not found: ID does not exist" Nov 22 08:17:14 crc kubenswrapper[4789]: I1122 08:17:14.316070 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Nov 22 08:17:14 crc kubenswrapper[4789]: E1122 08:17:14.316542 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e382d956-58e8-4c3b-abd9-0732d19bc35e" containerName="barbican-api-log" Nov 22 08:17:14 crc kubenswrapper[4789]: I1122 08:17:14.316567 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="e382d956-58e8-4c3b-abd9-0732d19bc35e" containerName="barbican-api-log" Nov 22 08:17:14 crc kubenswrapper[4789]: E1122 08:17:14.316583 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e382d956-58e8-4c3b-abd9-0732d19bc35e" containerName="barbican-api" Nov 22 08:17:14 crc kubenswrapper[4789]: I1122 08:17:14.316592 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="e382d956-58e8-4c3b-abd9-0732d19bc35e" containerName="barbican-api" Nov 22 08:17:14 crc kubenswrapper[4789]: I1122 08:17:14.316840 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="e382d956-58e8-4c3b-abd9-0732d19bc35e" containerName="barbican-api" Nov 22 08:17:14 crc kubenswrapper[4789]: I1122 08:17:14.316862 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="e382d956-58e8-4c3b-abd9-0732d19bc35e" containerName="barbican-api-log" Nov 22 08:17:14 crc kubenswrapper[4789]: I1122 08:17:14.317423 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 22 08:17:14 crc kubenswrapper[4789]: I1122 08:17:14.322363 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-76p4b" Nov 22 08:17:14 crc kubenswrapper[4789]: I1122 08:17:14.322964 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Nov 22 08:17:14 crc kubenswrapper[4789]: I1122 08:17:14.326873 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Nov 22 08:17:14 crc kubenswrapper[4789]: I1122 08:17:14.347602 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 22 08:17:14 crc kubenswrapper[4789]: I1122 08:17:14.462381 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/13cfa613-fa3b-4a2f-a228-0fbf526e489e-openstack-config\") pod \"openstackclient\" (UID: \"13cfa613-fa3b-4a2f-a228-0fbf526e489e\") " pod="openstack/openstackclient" Nov 22 08:17:14 crc kubenswrapper[4789]: I1122 08:17:14.462499 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/13cfa613-fa3b-4a2f-a228-0fbf526e489e-openstack-config-secret\") pod \"openstackclient\" (UID: \"13cfa613-fa3b-4a2f-a228-0fbf526e489e\") " pod="openstack/openstackclient" Nov 22 08:17:14 crc kubenswrapper[4789]: I1122 08:17:14.462573 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gwwsq\" (UniqueName: \"kubernetes.io/projected/13cfa613-fa3b-4a2f-a228-0fbf526e489e-kube-api-access-gwwsq\") pod \"openstackclient\" (UID: \"13cfa613-fa3b-4a2f-a228-0fbf526e489e\") " pod="openstack/openstackclient" Nov 22 08:17:14 crc kubenswrapper[4789]: I1122 08:17:14.462603 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13cfa613-fa3b-4a2f-a228-0fbf526e489e-combined-ca-bundle\") pod \"openstackclient\" (UID: \"13cfa613-fa3b-4a2f-a228-0fbf526e489e\") " pod="openstack/openstackclient" Nov 22 08:17:14 crc kubenswrapper[4789]: I1122 08:17:14.564114 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/13cfa613-fa3b-4a2f-a228-0fbf526e489e-openstack-config\") pod \"openstackclient\" (UID: \"13cfa613-fa3b-4a2f-a228-0fbf526e489e\") " pod="openstack/openstackclient" Nov 22 08:17:14 crc kubenswrapper[4789]: I1122 08:17:14.564222 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/13cfa613-fa3b-4a2f-a228-0fbf526e489e-openstack-config-secret\") pod \"openstackclient\" (UID: \"13cfa613-fa3b-4a2f-a228-0fbf526e489e\") " pod="openstack/openstackclient" Nov 22 08:17:14 crc kubenswrapper[4789]: I1122 08:17:14.564283 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gwwsq\" (UniqueName: \"kubernetes.io/projected/13cfa613-fa3b-4a2f-a228-0fbf526e489e-kube-api-access-gwwsq\") pod \"openstackclient\" (UID: \"13cfa613-fa3b-4a2f-a228-0fbf526e489e\") " pod="openstack/openstackclient" Nov 22 08:17:14 crc kubenswrapper[4789]: I1122 08:17:14.564319 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13cfa613-fa3b-4a2f-a228-0fbf526e489e-combined-ca-bundle\") pod \"openstackclient\" (UID: \"13cfa613-fa3b-4a2f-a228-0fbf526e489e\") " pod="openstack/openstackclient" Nov 22 08:17:14 crc kubenswrapper[4789]: I1122 08:17:14.565005 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/13cfa613-fa3b-4a2f-a228-0fbf526e489e-openstack-config\") pod \"openstackclient\" (UID: \"13cfa613-fa3b-4a2f-a228-0fbf526e489e\") " pod="openstack/openstackclient" Nov 22 08:17:14 crc kubenswrapper[4789]: I1122 08:17:14.569425 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/13cfa613-fa3b-4a2f-a228-0fbf526e489e-openstack-config-secret\") pod \"openstackclient\" (UID: \"13cfa613-fa3b-4a2f-a228-0fbf526e489e\") " pod="openstack/openstackclient" Nov 22 08:17:14 crc kubenswrapper[4789]: I1122 08:17:14.570669 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13cfa613-fa3b-4a2f-a228-0fbf526e489e-combined-ca-bundle\") pod \"openstackclient\" (UID: \"13cfa613-fa3b-4a2f-a228-0fbf526e489e\") " pod="openstack/openstackclient" Nov 22 08:17:14 crc kubenswrapper[4789]: I1122 08:17:14.583618 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gwwsq\" (UniqueName: \"kubernetes.io/projected/13cfa613-fa3b-4a2f-a228-0fbf526e489e-kube-api-access-gwwsq\") pod \"openstackclient\" (UID: \"13cfa613-fa3b-4a2f-a228-0fbf526e489e\") " pod="openstack/openstackclient" Nov 22 08:17:14 crc kubenswrapper[4789]: I1122 08:17:14.648360 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 22 08:17:15 crc kubenswrapper[4789]: I1122 08:17:15.081911 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 22 08:17:15 crc kubenswrapper[4789]: I1122 08:17:15.136000 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"13cfa613-fa3b-4a2f-a228-0fbf526e489e","Type":"ContainerStarted","Data":"6c359befebb3c4e96414e4384087e96f25b20d05f04315385430b97804de2b65"} Nov 22 08:17:15 crc kubenswrapper[4789]: I1122 08:17:15.977191 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e382d956-58e8-4c3b-abd9-0732d19bc35e" path="/var/lib/kubelet/pods/e382d956-58e8-4c3b-abd9-0732d19bc35e/volumes" Nov 22 08:17:16 crc kubenswrapper[4789]: I1122 08:17:16.653353 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Nov 22 08:17:18 crc kubenswrapper[4789]: I1122 08:17:18.030509 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-568cd49f4d-ms9nh" Nov 22 08:17:18 crc kubenswrapper[4789]: I1122 08:17:18.452685 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-5b56df9788-wsmhq" podUID="56fd3074-a15c-4024-baf9-f784a50e14a5" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.140:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.140:8443: connect: connection refused" Nov 22 08:17:18 crc kubenswrapper[4789]: I1122 08:17:18.452881 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-5b56df9788-wsmhq" Nov 22 08:17:23 crc kubenswrapper[4789]: I1122 08:17:23.784593 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 22 08:17:24 crc kubenswrapper[4789]: I1122 08:17:24.216516 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-vcmgb"] Nov 22 08:17:24 crc kubenswrapper[4789]: I1122 08:17:24.223712 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-vcmgb" Nov 22 08:17:24 crc kubenswrapper[4789]: I1122 08:17:24.234596 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-vcmgb"] Nov 22 08:17:24 crc kubenswrapper[4789]: I1122 08:17:24.244859 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"13cfa613-fa3b-4a2f-a228-0fbf526e489e","Type":"ContainerStarted","Data":"64ec2b236bce9c48579f7025c3be726e8613053ca485a6b16e3deb5f3ce00669"} Nov 22 08:17:24 crc kubenswrapper[4789]: I1122 08:17:24.275845 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=1.650402577 podStartE2EDuration="10.275825694s" podCreationTimestamp="2025-11-22 08:17:14 +0000 UTC" firstStartedPulling="2025-11-22 08:17:15.088255996 +0000 UTC m=+1249.322656269" lastFinishedPulling="2025-11-22 08:17:23.713679113 +0000 UTC m=+1257.948079386" observedRunningTime="2025-11-22 08:17:24.265316766 +0000 UTC m=+1258.499717039" watchObservedRunningTime="2025-11-22 08:17:24.275825694 +0000 UTC m=+1258.510225967" Nov 22 08:17:24 crc kubenswrapper[4789]: I1122 08:17:24.325253 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-frb4x"] Nov 22 08:17:24 crc kubenswrapper[4789]: I1122 08:17:24.326397 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-frb4x" Nov 22 08:17:24 crc kubenswrapper[4789]: I1122 08:17:24.346170 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-frb4x"] Nov 22 08:17:24 crc kubenswrapper[4789]: I1122 08:17:24.358808 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/aa7f415d-4528-41d4-996f-57c061b72885-operator-scripts\") pod \"nova-api-db-create-vcmgb\" (UID: \"aa7f415d-4528-41d4-996f-57c061b72885\") " pod="openstack/nova-api-db-create-vcmgb" Nov 22 08:17:24 crc kubenswrapper[4789]: I1122 08:17:24.358946 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ktvc4\" (UniqueName: \"kubernetes.io/projected/aa7f415d-4528-41d4-996f-57c061b72885-kube-api-access-ktvc4\") pod \"nova-api-db-create-vcmgb\" (UID: \"aa7f415d-4528-41d4-996f-57c061b72885\") " pod="openstack/nova-api-db-create-vcmgb" Nov 22 08:17:24 crc kubenswrapper[4789]: I1122 08:17:24.427109 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-8cdf-account-create-hzl44"] Nov 22 08:17:24 crc kubenswrapper[4789]: I1122 08:17:24.428500 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-8cdf-account-create-hzl44" Nov 22 08:17:24 crc kubenswrapper[4789]: I1122 08:17:24.430626 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Nov 22 08:17:24 crc kubenswrapper[4789]: I1122 08:17:24.440613 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-8cdf-account-create-hzl44"] Nov 22 08:17:24 crc kubenswrapper[4789]: I1122 08:17:24.460413 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/83e51032-30aa-47cb-b856-36b4280c2775-operator-scripts\") pod \"nova-cell0-db-create-frb4x\" (UID: \"83e51032-30aa-47cb-b856-36b4280c2775\") " pod="openstack/nova-cell0-db-create-frb4x" Nov 22 08:17:24 crc kubenswrapper[4789]: I1122 08:17:24.460528 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/aa7f415d-4528-41d4-996f-57c061b72885-operator-scripts\") pod \"nova-api-db-create-vcmgb\" (UID: \"aa7f415d-4528-41d4-996f-57c061b72885\") " pod="openstack/nova-api-db-create-vcmgb" Nov 22 08:17:24 crc kubenswrapper[4789]: I1122 08:17:24.460591 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2xc5d\" (UniqueName: \"kubernetes.io/projected/83e51032-30aa-47cb-b856-36b4280c2775-kube-api-access-2xc5d\") pod \"nova-cell0-db-create-frb4x\" (UID: \"83e51032-30aa-47cb-b856-36b4280c2775\") " pod="openstack/nova-cell0-db-create-frb4x" Nov 22 08:17:24 crc kubenswrapper[4789]: I1122 08:17:24.460613 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ktvc4\" (UniqueName: \"kubernetes.io/projected/aa7f415d-4528-41d4-996f-57c061b72885-kube-api-access-ktvc4\") pod \"nova-api-db-create-vcmgb\" (UID: \"aa7f415d-4528-41d4-996f-57c061b72885\") " pod="openstack/nova-api-db-create-vcmgb" Nov 22 08:17:24 crc kubenswrapper[4789]: I1122 08:17:24.461619 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/aa7f415d-4528-41d4-996f-57c061b72885-operator-scripts\") pod \"nova-api-db-create-vcmgb\" (UID: \"aa7f415d-4528-41d4-996f-57c061b72885\") " pod="openstack/nova-api-db-create-vcmgb" Nov 22 08:17:24 crc kubenswrapper[4789]: I1122 08:17:24.480049 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ktvc4\" (UniqueName: \"kubernetes.io/projected/aa7f415d-4528-41d4-996f-57c061b72885-kube-api-access-ktvc4\") pod \"nova-api-db-create-vcmgb\" (UID: \"aa7f415d-4528-41d4-996f-57c061b72885\") " pod="openstack/nova-api-db-create-vcmgb" Nov 22 08:17:24 crc kubenswrapper[4789]: I1122 08:17:24.528151 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-5mvvx"] Nov 22 08:17:24 crc kubenswrapper[4789]: I1122 08:17:24.529508 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-5mvvx" Nov 22 08:17:24 crc kubenswrapper[4789]: I1122 08:17:24.540395 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-5mvvx"] Nov 22 08:17:24 crc kubenswrapper[4789]: I1122 08:17:24.554123 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-vcmgb" Nov 22 08:17:24 crc kubenswrapper[4789]: I1122 08:17:24.562226 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2xc5d\" (UniqueName: \"kubernetes.io/projected/83e51032-30aa-47cb-b856-36b4280c2775-kube-api-access-2xc5d\") pod \"nova-cell0-db-create-frb4x\" (UID: \"83e51032-30aa-47cb-b856-36b4280c2775\") " pod="openstack/nova-cell0-db-create-frb4x" Nov 22 08:17:24 crc kubenswrapper[4789]: I1122 08:17:24.562324 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9e886f6c-babc-4118-b532-50b72dacb370-operator-scripts\") pod \"nova-api-8cdf-account-create-hzl44\" (UID: \"9e886f6c-babc-4118-b532-50b72dacb370\") " pod="openstack/nova-api-8cdf-account-create-hzl44" Nov 22 08:17:24 crc kubenswrapper[4789]: I1122 08:17:24.562378 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8vfvc\" (UniqueName: \"kubernetes.io/projected/9e886f6c-babc-4118-b532-50b72dacb370-kube-api-access-8vfvc\") pod \"nova-api-8cdf-account-create-hzl44\" (UID: \"9e886f6c-babc-4118-b532-50b72dacb370\") " pod="openstack/nova-api-8cdf-account-create-hzl44" Nov 22 08:17:24 crc kubenswrapper[4789]: I1122 08:17:24.562436 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/83e51032-30aa-47cb-b856-36b4280c2775-operator-scripts\") pod \"nova-cell0-db-create-frb4x\" (UID: \"83e51032-30aa-47cb-b856-36b4280c2775\") " pod="openstack/nova-cell0-db-create-frb4x" Nov 22 08:17:24 crc kubenswrapper[4789]: I1122 08:17:24.563153 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/83e51032-30aa-47cb-b856-36b4280c2775-operator-scripts\") pod \"nova-cell0-db-create-frb4x\" (UID: \"83e51032-30aa-47cb-b856-36b4280c2775\") " pod="openstack/nova-cell0-db-create-frb4x" Nov 22 08:17:24 crc kubenswrapper[4789]: I1122 08:17:24.580572 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2xc5d\" (UniqueName: \"kubernetes.io/projected/83e51032-30aa-47cb-b856-36b4280c2775-kube-api-access-2xc5d\") pod \"nova-cell0-db-create-frb4x\" (UID: \"83e51032-30aa-47cb-b856-36b4280c2775\") " pod="openstack/nova-cell0-db-create-frb4x" Nov 22 08:17:24 crc kubenswrapper[4789]: I1122 08:17:24.632282 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-8958-account-create-626c2"] Nov 22 08:17:24 crc kubenswrapper[4789]: I1122 08:17:24.634126 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-8958-account-create-626c2" Nov 22 08:17:24 crc kubenswrapper[4789]: I1122 08:17:24.635928 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Nov 22 08:17:24 crc kubenswrapper[4789]: I1122 08:17:24.643548 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-frb4x" Nov 22 08:17:24 crc kubenswrapper[4789]: I1122 08:17:24.646973 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-8958-account-create-626c2"] Nov 22 08:17:24 crc kubenswrapper[4789]: I1122 08:17:24.664908 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8vfvc\" (UniqueName: \"kubernetes.io/projected/9e886f6c-babc-4118-b532-50b72dacb370-kube-api-access-8vfvc\") pod \"nova-api-8cdf-account-create-hzl44\" (UID: \"9e886f6c-babc-4118-b532-50b72dacb370\") " pod="openstack/nova-api-8cdf-account-create-hzl44" Nov 22 08:17:24 crc kubenswrapper[4789]: I1122 08:17:24.664983 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/07941151-cfeb-4356-ad4e-439fdb875687-operator-scripts\") pod \"nova-cell1-db-create-5mvvx\" (UID: \"07941151-cfeb-4356-ad4e-439fdb875687\") " pod="openstack/nova-cell1-db-create-5mvvx" Nov 22 08:17:24 crc kubenswrapper[4789]: I1122 08:17:24.665027 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xqjbz\" (UniqueName: \"kubernetes.io/projected/07941151-cfeb-4356-ad4e-439fdb875687-kube-api-access-xqjbz\") pod \"nova-cell1-db-create-5mvvx\" (UID: \"07941151-cfeb-4356-ad4e-439fdb875687\") " pod="openstack/nova-cell1-db-create-5mvvx" Nov 22 08:17:24 crc kubenswrapper[4789]: I1122 08:17:24.665107 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9e886f6c-babc-4118-b532-50b72dacb370-operator-scripts\") pod \"nova-api-8cdf-account-create-hzl44\" (UID: \"9e886f6c-babc-4118-b532-50b72dacb370\") " pod="openstack/nova-api-8cdf-account-create-hzl44" Nov 22 08:17:24 crc kubenswrapper[4789]: I1122 08:17:24.666144 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9e886f6c-babc-4118-b532-50b72dacb370-operator-scripts\") pod \"nova-api-8cdf-account-create-hzl44\" (UID: \"9e886f6c-babc-4118-b532-50b72dacb370\") " pod="openstack/nova-api-8cdf-account-create-hzl44" Nov 22 08:17:24 crc kubenswrapper[4789]: I1122 08:17:24.681513 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8vfvc\" (UniqueName: \"kubernetes.io/projected/9e886f6c-babc-4118-b532-50b72dacb370-kube-api-access-8vfvc\") pod \"nova-api-8cdf-account-create-hzl44\" (UID: \"9e886f6c-babc-4118-b532-50b72dacb370\") " pod="openstack/nova-api-8cdf-account-create-hzl44" Nov 22 08:17:24 crc kubenswrapper[4789]: I1122 08:17:24.747698 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-8cdf-account-create-hzl44" Nov 22 08:17:24 crc kubenswrapper[4789]: I1122 08:17:24.773852 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xqjbz\" (UniqueName: \"kubernetes.io/projected/07941151-cfeb-4356-ad4e-439fdb875687-kube-api-access-xqjbz\") pod \"nova-cell1-db-create-5mvvx\" (UID: \"07941151-cfeb-4356-ad4e-439fdb875687\") " pod="openstack/nova-cell1-db-create-5mvvx" Nov 22 08:17:24 crc kubenswrapper[4789]: I1122 08:17:24.773941 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vpgtl\" (UniqueName: \"kubernetes.io/projected/c61136a6-9398-4fec-89a0-e4d7291b99fc-kube-api-access-vpgtl\") pod \"nova-cell0-8958-account-create-626c2\" (UID: \"c61136a6-9398-4fec-89a0-e4d7291b99fc\") " pod="openstack/nova-cell0-8958-account-create-626c2" Nov 22 08:17:24 crc kubenswrapper[4789]: I1122 08:17:24.774001 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c61136a6-9398-4fec-89a0-e4d7291b99fc-operator-scripts\") pod \"nova-cell0-8958-account-create-626c2\" (UID: \"c61136a6-9398-4fec-89a0-e4d7291b99fc\") " pod="openstack/nova-cell0-8958-account-create-626c2" Nov 22 08:17:24 crc kubenswrapper[4789]: I1122 08:17:24.774093 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/07941151-cfeb-4356-ad4e-439fdb875687-operator-scripts\") pod \"nova-cell1-db-create-5mvvx\" (UID: \"07941151-cfeb-4356-ad4e-439fdb875687\") " pod="openstack/nova-cell1-db-create-5mvvx" Nov 22 08:17:24 crc kubenswrapper[4789]: I1122 08:17:24.775343 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/07941151-cfeb-4356-ad4e-439fdb875687-operator-scripts\") pod \"nova-cell1-db-create-5mvvx\" (UID: \"07941151-cfeb-4356-ad4e-439fdb875687\") " pod="openstack/nova-cell1-db-create-5mvvx" Nov 22 08:17:24 crc kubenswrapper[4789]: I1122 08:17:24.803625 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xqjbz\" (UniqueName: \"kubernetes.io/projected/07941151-cfeb-4356-ad4e-439fdb875687-kube-api-access-xqjbz\") pod \"nova-cell1-db-create-5mvvx\" (UID: \"07941151-cfeb-4356-ad4e-439fdb875687\") " pod="openstack/nova-cell1-db-create-5mvvx" Nov 22 08:17:24 crc kubenswrapper[4789]: I1122 08:17:24.849156 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-5mvvx" Nov 22 08:17:24 crc kubenswrapper[4789]: I1122 08:17:24.872370 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-9a6c-account-create-2cwpt"] Nov 22 08:17:24 crc kubenswrapper[4789]: I1122 08:17:24.874082 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-9a6c-account-create-2cwpt" Nov 22 08:17:24 crc kubenswrapper[4789]: I1122 08:17:24.875277 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vpgtl\" (UniqueName: \"kubernetes.io/projected/c61136a6-9398-4fec-89a0-e4d7291b99fc-kube-api-access-vpgtl\") pod \"nova-cell0-8958-account-create-626c2\" (UID: \"c61136a6-9398-4fec-89a0-e4d7291b99fc\") " pod="openstack/nova-cell0-8958-account-create-626c2" Nov 22 08:17:24 crc kubenswrapper[4789]: I1122 08:17:24.875361 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c61136a6-9398-4fec-89a0-e4d7291b99fc-operator-scripts\") pod \"nova-cell0-8958-account-create-626c2\" (UID: \"c61136a6-9398-4fec-89a0-e4d7291b99fc\") " pod="openstack/nova-cell0-8958-account-create-626c2" Nov 22 08:17:24 crc kubenswrapper[4789]: I1122 08:17:24.876085 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c61136a6-9398-4fec-89a0-e4d7291b99fc-operator-scripts\") pod \"nova-cell0-8958-account-create-626c2\" (UID: \"c61136a6-9398-4fec-89a0-e4d7291b99fc\") " pod="openstack/nova-cell0-8958-account-create-626c2" Nov 22 08:17:24 crc kubenswrapper[4789]: I1122 08:17:24.877517 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Nov 22 08:17:24 crc kubenswrapper[4789]: I1122 08:17:24.904099 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vpgtl\" (UniqueName: \"kubernetes.io/projected/c61136a6-9398-4fec-89a0-e4d7291b99fc-kube-api-access-vpgtl\") pod \"nova-cell0-8958-account-create-626c2\" (UID: \"c61136a6-9398-4fec-89a0-e4d7291b99fc\") " pod="openstack/nova-cell0-8958-account-create-626c2" Nov 22 08:17:24 crc kubenswrapper[4789]: I1122 08:17:24.918535 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-9a6c-account-create-2cwpt"] Nov 22 08:17:24 crc kubenswrapper[4789]: I1122 08:17:24.956534 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-8958-account-create-626c2" Nov 22 08:17:25 crc kubenswrapper[4789]: I1122 08:17:25.093273 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sxhw9\" (UniqueName: \"kubernetes.io/projected/d315bf00-e54d-40cb-a261-ff8bd068e535-kube-api-access-sxhw9\") pod \"nova-cell1-9a6c-account-create-2cwpt\" (UID: \"d315bf00-e54d-40cb-a261-ff8bd068e535\") " pod="openstack/nova-cell1-9a6c-account-create-2cwpt" Nov 22 08:17:25 crc kubenswrapper[4789]: I1122 08:17:25.093354 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d315bf00-e54d-40cb-a261-ff8bd068e535-operator-scripts\") pod \"nova-cell1-9a6c-account-create-2cwpt\" (UID: \"d315bf00-e54d-40cb-a261-ff8bd068e535\") " pod="openstack/nova-cell1-9a6c-account-create-2cwpt" Nov 22 08:17:25 crc kubenswrapper[4789]: I1122 08:17:25.118149 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-5fc55b98d9-vqxcz" Nov 22 08:17:25 crc kubenswrapper[4789]: I1122 08:17:25.195360 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sxhw9\" (UniqueName: \"kubernetes.io/projected/d315bf00-e54d-40cb-a261-ff8bd068e535-kube-api-access-sxhw9\") pod \"nova-cell1-9a6c-account-create-2cwpt\" (UID: \"d315bf00-e54d-40cb-a261-ff8bd068e535\") " pod="openstack/nova-cell1-9a6c-account-create-2cwpt" Nov 22 08:17:25 crc kubenswrapper[4789]: I1122 08:17:25.196166 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d315bf00-e54d-40cb-a261-ff8bd068e535-operator-scripts\") pod \"nova-cell1-9a6c-account-create-2cwpt\" (UID: \"d315bf00-e54d-40cb-a261-ff8bd068e535\") " pod="openstack/nova-cell1-9a6c-account-create-2cwpt" Nov 22 08:17:25 crc kubenswrapper[4789]: I1122 08:17:25.197030 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d315bf00-e54d-40cb-a261-ff8bd068e535-operator-scripts\") pod \"nova-cell1-9a6c-account-create-2cwpt\" (UID: \"d315bf00-e54d-40cb-a261-ff8bd068e535\") " pod="openstack/nova-cell1-9a6c-account-create-2cwpt" Nov 22 08:17:25 crc kubenswrapper[4789]: I1122 08:17:25.206865 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-568cd49f4d-ms9nh"] Nov 22 08:17:25 crc kubenswrapper[4789]: I1122 08:17:25.207114 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-568cd49f4d-ms9nh" podUID="2949ebe1-7541-42a2-894c-2588c2bd3396" containerName="neutron-api" containerID="cri-o://77b48a1183833a07016416adb6c59b89b3874df9589e6f5b1c3dba3285e81de8" gracePeriod=30 Nov 22 08:17:25 crc kubenswrapper[4789]: I1122 08:17:25.207519 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-568cd49f4d-ms9nh" podUID="2949ebe1-7541-42a2-894c-2588c2bd3396" containerName="neutron-httpd" containerID="cri-o://9aec8c709e300ab397822f5fff34aba93e0915d5c8707aac0584b703bc60df1a" gracePeriod=30 Nov 22 08:17:25 crc kubenswrapper[4789]: I1122 08:17:25.219673 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-vcmgb"] Nov 22 08:17:25 crc kubenswrapper[4789]: I1122 08:17:25.242498 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sxhw9\" (UniqueName: \"kubernetes.io/projected/d315bf00-e54d-40cb-a261-ff8bd068e535-kube-api-access-sxhw9\") pod \"nova-cell1-9a6c-account-create-2cwpt\" (UID: \"d315bf00-e54d-40cb-a261-ff8bd068e535\") " pod="openstack/nova-cell1-9a6c-account-create-2cwpt" Nov 22 08:17:25 crc kubenswrapper[4789]: I1122 08:17:25.268710 4789 generic.go:334] "Generic (PLEG): container finished" podID="56fd3074-a15c-4024-baf9-f784a50e14a5" containerID="93ff7a096adcf0bb8be87de88e1df3ddfb208ee02cbae6c682c5da91f25f03f9" exitCode=137 Nov 22 08:17:25 crc kubenswrapper[4789]: I1122 08:17:25.268787 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5b56df9788-wsmhq" event={"ID":"56fd3074-a15c-4024-baf9-f784a50e14a5","Type":"ContainerDied","Data":"93ff7a096adcf0bb8be87de88e1df3ddfb208ee02cbae6c682c5da91f25f03f9"} Nov 22 08:17:25 crc kubenswrapper[4789]: I1122 08:17:25.270472 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-vcmgb" event={"ID":"aa7f415d-4528-41d4-996f-57c061b72885","Type":"ContainerStarted","Data":"58a4d35171fcbbdb2d527cadd57dc241e05138388bc58b1376eb5c551ccd8c22"} Nov 22 08:17:25 crc kubenswrapper[4789]: I1122 08:17:25.319106 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-frb4x"] Nov 22 08:17:25 crc kubenswrapper[4789]: I1122 08:17:25.320103 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-9a6c-account-create-2cwpt" Nov 22 08:17:25 crc kubenswrapper[4789]: I1122 08:17:25.394419 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-8cdf-account-create-hzl44"] Nov 22 08:17:25 crc kubenswrapper[4789]: I1122 08:17:25.539710 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5b56df9788-wsmhq" Nov 22 08:17:25 crc kubenswrapper[4789]: I1122 08:17:25.617668 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/56fd3074-a15c-4024-baf9-f784a50e14a5-logs\") pod \"56fd3074-a15c-4024-baf9-f784a50e14a5\" (UID: \"56fd3074-a15c-4024-baf9-f784a50e14a5\") " Nov 22 08:17:25 crc kubenswrapper[4789]: I1122 08:17:25.617745 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v9rvp\" (UniqueName: \"kubernetes.io/projected/56fd3074-a15c-4024-baf9-f784a50e14a5-kube-api-access-v9rvp\") pod \"56fd3074-a15c-4024-baf9-f784a50e14a5\" (UID: \"56fd3074-a15c-4024-baf9-f784a50e14a5\") " Nov 22 08:17:25 crc kubenswrapper[4789]: I1122 08:17:25.617851 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56fd3074-a15c-4024-baf9-f784a50e14a5-combined-ca-bundle\") pod \"56fd3074-a15c-4024-baf9-f784a50e14a5\" (UID: \"56fd3074-a15c-4024-baf9-f784a50e14a5\") " Nov 22 08:17:25 crc kubenswrapper[4789]: I1122 08:17:25.617918 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/56fd3074-a15c-4024-baf9-f784a50e14a5-horizon-secret-key\") pod \"56fd3074-a15c-4024-baf9-f784a50e14a5\" (UID: \"56fd3074-a15c-4024-baf9-f784a50e14a5\") " Nov 22 08:17:25 crc kubenswrapper[4789]: I1122 08:17:25.620338 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/56fd3074-a15c-4024-baf9-f784a50e14a5-config-data\") pod \"56fd3074-a15c-4024-baf9-f784a50e14a5\" (UID: \"56fd3074-a15c-4024-baf9-f784a50e14a5\") " Nov 22 08:17:25 crc kubenswrapper[4789]: I1122 08:17:25.620445 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/56fd3074-a15c-4024-baf9-f784a50e14a5-horizon-tls-certs\") pod \"56fd3074-a15c-4024-baf9-f784a50e14a5\" (UID: \"56fd3074-a15c-4024-baf9-f784a50e14a5\") " Nov 22 08:17:25 crc kubenswrapper[4789]: I1122 08:17:25.620494 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/56fd3074-a15c-4024-baf9-f784a50e14a5-scripts\") pod \"56fd3074-a15c-4024-baf9-f784a50e14a5\" (UID: \"56fd3074-a15c-4024-baf9-f784a50e14a5\") " Nov 22 08:17:25 crc kubenswrapper[4789]: I1122 08:17:25.620527 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/56fd3074-a15c-4024-baf9-f784a50e14a5-logs" (OuterVolumeSpecName: "logs") pod "56fd3074-a15c-4024-baf9-f784a50e14a5" (UID: "56fd3074-a15c-4024-baf9-f784a50e14a5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:17:25 crc kubenswrapper[4789]: I1122 08:17:25.621278 4789 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/56fd3074-a15c-4024-baf9-f784a50e14a5-logs\") on node \"crc\" DevicePath \"\"" Nov 22 08:17:25 crc kubenswrapper[4789]: I1122 08:17:25.628396 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56fd3074-a15c-4024-baf9-f784a50e14a5-kube-api-access-v9rvp" (OuterVolumeSpecName: "kube-api-access-v9rvp") pod "56fd3074-a15c-4024-baf9-f784a50e14a5" (UID: "56fd3074-a15c-4024-baf9-f784a50e14a5"). InnerVolumeSpecName "kube-api-access-v9rvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:17:25 crc kubenswrapper[4789]: I1122 08:17:25.632066 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56fd3074-a15c-4024-baf9-f784a50e14a5-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "56fd3074-a15c-4024-baf9-f784a50e14a5" (UID: "56fd3074-a15c-4024-baf9-f784a50e14a5"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:17:25 crc kubenswrapper[4789]: I1122 08:17:25.634146 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-5mvvx"] Nov 22 08:17:25 crc kubenswrapper[4789]: W1122 08:17:25.639385 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod07941151_cfeb_4356_ad4e_439fdb875687.slice/crio-ff00e592fbc1c20b184c645de8d3cffffeb606cf42703ba799db83d17bca13fd WatchSource:0}: Error finding container ff00e592fbc1c20b184c645de8d3cffffeb606cf42703ba799db83d17bca13fd: Status 404 returned error can't find the container with id ff00e592fbc1c20b184c645de8d3cffffeb606cf42703ba799db83d17bca13fd Nov 22 08:17:25 crc kubenswrapper[4789]: I1122 08:17:25.671182 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/56fd3074-a15c-4024-baf9-f784a50e14a5-config-data" (OuterVolumeSpecName: "config-data") pod "56fd3074-a15c-4024-baf9-f784a50e14a5" (UID: "56fd3074-a15c-4024-baf9-f784a50e14a5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:17:25 crc kubenswrapper[4789]: I1122 08:17:25.678727 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56fd3074-a15c-4024-baf9-f784a50e14a5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "56fd3074-a15c-4024-baf9-f784a50e14a5" (UID: "56fd3074-a15c-4024-baf9-f784a50e14a5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:17:25 crc kubenswrapper[4789]: I1122 08:17:25.699265 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/56fd3074-a15c-4024-baf9-f784a50e14a5-scripts" (OuterVolumeSpecName: "scripts") pod "56fd3074-a15c-4024-baf9-f784a50e14a5" (UID: "56fd3074-a15c-4024-baf9-f784a50e14a5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:17:25 crc kubenswrapper[4789]: I1122 08:17:25.714363 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56fd3074-a15c-4024-baf9-f784a50e14a5-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "56fd3074-a15c-4024-baf9-f784a50e14a5" (UID: "56fd3074-a15c-4024-baf9-f784a50e14a5"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:17:25 crc kubenswrapper[4789]: I1122 08:17:25.726659 4789 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56fd3074-a15c-4024-baf9-f784a50e14a5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:17:25 crc kubenswrapper[4789]: I1122 08:17:25.726696 4789 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/56fd3074-a15c-4024-baf9-f784a50e14a5-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Nov 22 08:17:25 crc kubenswrapper[4789]: I1122 08:17:25.726710 4789 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/56fd3074-a15c-4024-baf9-f784a50e14a5-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:17:25 crc kubenswrapper[4789]: I1122 08:17:25.726721 4789 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/56fd3074-a15c-4024-baf9-f784a50e14a5-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 22 08:17:25 crc kubenswrapper[4789]: I1122 08:17:25.726734 4789 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/56fd3074-a15c-4024-baf9-f784a50e14a5-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:17:25 crc kubenswrapper[4789]: I1122 08:17:25.726786 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v9rvp\" (UniqueName: \"kubernetes.io/projected/56fd3074-a15c-4024-baf9-f784a50e14a5-kube-api-access-v9rvp\") on node \"crc\" DevicePath \"\"" Nov 22 08:17:25 crc kubenswrapper[4789]: I1122 08:17:25.785654 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-8958-account-create-626c2"] Nov 22 08:17:25 crc kubenswrapper[4789]: I1122 08:17:25.917985 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-9a6c-account-create-2cwpt"] Nov 22 08:17:25 crc kubenswrapper[4789]: W1122 08:17:25.919394 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd315bf00_e54d_40cb_a261_ff8bd068e535.slice/crio-441ab1a55c0b99d631a57d777c3da00c561cd44d242e70638b828eafbc1a2284 WatchSource:0}: Error finding container 441ab1a55c0b99d631a57d777c3da00c561cd44d242e70638b828eafbc1a2284: Status 404 returned error can't find the container with id 441ab1a55c0b99d631a57d777c3da00c561cd44d242e70638b828eafbc1a2284 Nov 22 08:17:26 crc kubenswrapper[4789]: E1122 08:17:26.247414 4789 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod83e51032_30aa_47cb_b856_36b4280c2775.slice/crio-conmon-8c25a7cf637bf7ef49c67f9e27a1ebd73539326339bf61342245cf3d89013cce.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9e886f6c_babc_4118_b532_50b72dacb370.slice/crio-conmon-d969b22b590e1304aa90136e93d1646335c192a7ee87707af43ab394c65b5768.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaa7f415d_4528_41d4_996f_57c061b72885.slice/crio-conmon-4ccac87ec671fab5f21281811992e6a8edb929190b35525ba420234d05dff837.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod83e51032_30aa_47cb_b856_36b4280c2775.slice/crio-8c25a7cf637bf7ef49c67f9e27a1ebd73539326339bf61342245cf3d89013cce.scope\": RecentStats: unable to find data in memory cache]" Nov 22 08:17:26 crc kubenswrapper[4789]: I1122 08:17:26.286584 4789 generic.go:334] "Generic (PLEG): container finished" podID="83e51032-30aa-47cb-b856-36b4280c2775" containerID="8c25a7cf637bf7ef49c67f9e27a1ebd73539326339bf61342245cf3d89013cce" exitCode=0 Nov 22 08:17:26 crc kubenswrapper[4789]: I1122 08:17:26.286679 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-frb4x" event={"ID":"83e51032-30aa-47cb-b856-36b4280c2775","Type":"ContainerDied","Data":"8c25a7cf637bf7ef49c67f9e27a1ebd73539326339bf61342245cf3d89013cce"} Nov 22 08:17:26 crc kubenswrapper[4789]: I1122 08:17:26.286705 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-frb4x" event={"ID":"83e51032-30aa-47cb-b856-36b4280c2775","Type":"ContainerStarted","Data":"a957e281864c137f0975c5bf49cda635e544e9c9212baa2987aaad1c3af050e5"} Nov 22 08:17:26 crc kubenswrapper[4789]: I1122 08:17:26.290700 4789 generic.go:334] "Generic (PLEG): container finished" podID="9e886f6c-babc-4118-b532-50b72dacb370" containerID="d969b22b590e1304aa90136e93d1646335c192a7ee87707af43ab394c65b5768" exitCode=0 Nov 22 08:17:26 crc kubenswrapper[4789]: I1122 08:17:26.291049 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-8cdf-account-create-hzl44" event={"ID":"9e886f6c-babc-4118-b532-50b72dacb370","Type":"ContainerDied","Data":"d969b22b590e1304aa90136e93d1646335c192a7ee87707af43ab394c65b5768"} Nov 22 08:17:26 crc kubenswrapper[4789]: I1122 08:17:26.291081 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-8cdf-account-create-hzl44" event={"ID":"9e886f6c-babc-4118-b532-50b72dacb370","Type":"ContainerStarted","Data":"e57ced5703dedf939d1ea98ec6641a85b0c2f0fec619262907f6a43e2a04690d"} Nov 22 08:17:26 crc kubenswrapper[4789]: I1122 08:17:26.300807 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5b56df9788-wsmhq" event={"ID":"56fd3074-a15c-4024-baf9-f784a50e14a5","Type":"ContainerDied","Data":"3b3a905c1c5482e887bf3be488fe68c9df53059d564d5f9051d737b756870978"} Nov 22 08:17:26 crc kubenswrapper[4789]: I1122 08:17:26.300870 4789 scope.go:117] "RemoveContainer" containerID="2e99dc56ab31ddc560842a39626551056b3b1db94ece09fb54c781cd0d891832" Nov 22 08:17:26 crc kubenswrapper[4789]: I1122 08:17:26.301071 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5b56df9788-wsmhq" Nov 22 08:17:26 crc kubenswrapper[4789]: I1122 08:17:26.319464 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-8958-account-create-626c2" event={"ID":"c61136a6-9398-4fec-89a0-e4d7291b99fc","Type":"ContainerStarted","Data":"d0d43f2bd491b2aa0fa3a787614d3055caba5ed14b6ced394e05243a85efd573"} Nov 22 08:17:26 crc kubenswrapper[4789]: I1122 08:17:26.336996 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-9a6c-account-create-2cwpt" event={"ID":"d315bf00-e54d-40cb-a261-ff8bd068e535","Type":"ContainerStarted","Data":"441ab1a55c0b99d631a57d777c3da00c561cd44d242e70638b828eafbc1a2284"} Nov 22 08:17:26 crc kubenswrapper[4789]: I1122 08:17:26.354909 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5b56df9788-wsmhq"] Nov 22 08:17:26 crc kubenswrapper[4789]: I1122 08:17:26.355578 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-5mvvx" event={"ID":"07941151-cfeb-4356-ad4e-439fdb875687","Type":"ContainerStarted","Data":"ccc2c4d2207917c6e4fa2291159a5b1d29db8d298af4a1d1ef3d948cf7e498b5"} Nov 22 08:17:26 crc kubenswrapper[4789]: I1122 08:17:26.355609 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-5mvvx" event={"ID":"07941151-cfeb-4356-ad4e-439fdb875687","Type":"ContainerStarted","Data":"ff00e592fbc1c20b184c645de8d3cffffeb606cf42703ba799db83d17bca13fd"} Nov 22 08:17:26 crc kubenswrapper[4789]: I1122 08:17:26.375514 4789 generic.go:334] "Generic (PLEG): container finished" podID="aa7f415d-4528-41d4-996f-57c061b72885" containerID="4ccac87ec671fab5f21281811992e6a8edb929190b35525ba420234d05dff837" exitCode=0 Nov 22 08:17:26 crc kubenswrapper[4789]: I1122 08:17:26.375584 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-vcmgb" event={"ID":"aa7f415d-4528-41d4-996f-57c061b72885","Type":"ContainerDied","Data":"4ccac87ec671fab5f21281811992e6a8edb929190b35525ba420234d05dff837"} Nov 22 08:17:26 crc kubenswrapper[4789]: I1122 08:17:26.380065 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-5b56df9788-wsmhq"] Nov 22 08:17:26 crc kubenswrapper[4789]: I1122 08:17:26.384036 4789 generic.go:334] "Generic (PLEG): container finished" podID="2949ebe1-7541-42a2-894c-2588c2bd3396" containerID="9aec8c709e300ab397822f5fff34aba93e0915d5c8707aac0584b703bc60df1a" exitCode=0 Nov 22 08:17:26 crc kubenswrapper[4789]: I1122 08:17:26.384095 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-568cd49f4d-ms9nh" event={"ID":"2949ebe1-7541-42a2-894c-2588c2bd3396","Type":"ContainerDied","Data":"9aec8c709e300ab397822f5fff34aba93e0915d5c8707aac0584b703bc60df1a"} Nov 22 08:17:26 crc kubenswrapper[4789]: I1122 08:17:26.538954 4789 scope.go:117] "RemoveContainer" containerID="93ff7a096adcf0bb8be87de88e1df3ddfb208ee02cbae6c682c5da91f25f03f9" Nov 22 08:17:26 crc kubenswrapper[4789]: I1122 08:17:26.753273 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:17:26 crc kubenswrapper[4789]: I1122 08:17:26.753521 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="de991bf6-8f96-4c99-bf7a-b8909c0c80d2" containerName="ceilometer-central-agent" containerID="cri-o://60c80dac437908b708a2f24653527c9e713047e6113c781463535e492f681d00" gracePeriod=30 Nov 22 08:17:26 crc kubenswrapper[4789]: I1122 08:17:26.753609 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="de991bf6-8f96-4c99-bf7a-b8909c0c80d2" containerName="sg-core" containerID="cri-o://6168d229a82e8b6c67cebc486d47d83616de74ed1dcc8f74ca5a8473c64386f3" gracePeriod=30 Nov 22 08:17:26 crc kubenswrapper[4789]: I1122 08:17:26.753622 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="de991bf6-8f96-4c99-bf7a-b8909c0c80d2" containerName="ceilometer-notification-agent" containerID="cri-o://63d2e8a684de408f5cfa4728b23b07803ae5286c76ec7a0b812bf69230a7922f" gracePeriod=30 Nov 22 08:17:26 crc kubenswrapper[4789]: I1122 08:17:26.753880 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="de991bf6-8f96-4c99-bf7a-b8909c0c80d2" containerName="proxy-httpd" containerID="cri-o://4b4d49b0d5e06a9d2af814c8237d522b72d09606a55f523f315c307cfa7163bf" gracePeriod=30 Nov 22 08:17:27 crc kubenswrapper[4789]: I1122 08:17:27.395442 4789 generic.go:334] "Generic (PLEG): container finished" podID="d315bf00-e54d-40cb-a261-ff8bd068e535" containerID="d2be3b96f1f09315958202746fe2b6ff79bc62d91dcf7feef03096e72661d2f5" exitCode=0 Nov 22 08:17:27 crc kubenswrapper[4789]: I1122 08:17:27.395526 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-9a6c-account-create-2cwpt" event={"ID":"d315bf00-e54d-40cb-a261-ff8bd068e535","Type":"ContainerDied","Data":"d2be3b96f1f09315958202746fe2b6ff79bc62d91dcf7feef03096e72661d2f5"} Nov 22 08:17:27 crc kubenswrapper[4789]: I1122 08:17:27.398050 4789 generic.go:334] "Generic (PLEG): container finished" podID="c61136a6-9398-4fec-89a0-e4d7291b99fc" containerID="9b84e92c2e789ca0e69d366aec32c9b0a04e8eb06bc59b5b87e9d586cef870a1" exitCode=0 Nov 22 08:17:27 crc kubenswrapper[4789]: I1122 08:17:27.398118 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-8958-account-create-626c2" event={"ID":"c61136a6-9398-4fec-89a0-e4d7291b99fc","Type":"ContainerDied","Data":"9b84e92c2e789ca0e69d366aec32c9b0a04e8eb06bc59b5b87e9d586cef870a1"} Nov 22 08:17:27 crc kubenswrapper[4789]: I1122 08:17:27.399720 4789 generic.go:334] "Generic (PLEG): container finished" podID="07941151-cfeb-4356-ad4e-439fdb875687" containerID="ccc2c4d2207917c6e4fa2291159a5b1d29db8d298af4a1d1ef3d948cf7e498b5" exitCode=0 Nov 22 08:17:27 crc kubenswrapper[4789]: I1122 08:17:27.399812 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-5mvvx" event={"ID":"07941151-cfeb-4356-ad4e-439fdb875687","Type":"ContainerDied","Data":"ccc2c4d2207917c6e4fa2291159a5b1d29db8d298af4a1d1ef3d948cf7e498b5"} Nov 22 08:17:27 crc kubenswrapper[4789]: I1122 08:17:27.405306 4789 generic.go:334] "Generic (PLEG): container finished" podID="de991bf6-8f96-4c99-bf7a-b8909c0c80d2" containerID="4b4d49b0d5e06a9d2af814c8237d522b72d09606a55f523f315c307cfa7163bf" exitCode=0 Nov 22 08:17:27 crc kubenswrapper[4789]: I1122 08:17:27.405332 4789 generic.go:334] "Generic (PLEG): container finished" podID="de991bf6-8f96-4c99-bf7a-b8909c0c80d2" containerID="6168d229a82e8b6c67cebc486d47d83616de74ed1dcc8f74ca5a8473c64386f3" exitCode=2 Nov 22 08:17:27 crc kubenswrapper[4789]: I1122 08:17:27.405343 4789 generic.go:334] "Generic (PLEG): container finished" podID="de991bf6-8f96-4c99-bf7a-b8909c0c80d2" containerID="60c80dac437908b708a2f24653527c9e713047e6113c781463535e492f681d00" exitCode=0 Nov 22 08:17:27 crc kubenswrapper[4789]: I1122 08:17:27.405495 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"de991bf6-8f96-4c99-bf7a-b8909c0c80d2","Type":"ContainerDied","Data":"4b4d49b0d5e06a9d2af814c8237d522b72d09606a55f523f315c307cfa7163bf"} Nov 22 08:17:27 crc kubenswrapper[4789]: I1122 08:17:27.405882 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"de991bf6-8f96-4c99-bf7a-b8909c0c80d2","Type":"ContainerDied","Data":"6168d229a82e8b6c67cebc486d47d83616de74ed1dcc8f74ca5a8473c64386f3"} Nov 22 08:17:27 crc kubenswrapper[4789]: I1122 08:17:27.405900 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"de991bf6-8f96-4c99-bf7a-b8909c0c80d2","Type":"ContainerDied","Data":"60c80dac437908b708a2f24653527c9e713047e6113c781463535e492f681d00"} Nov 22 08:17:27 crc kubenswrapper[4789]: I1122 08:17:27.867181 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-frb4x" Nov 22 08:17:27 crc kubenswrapper[4789]: I1122 08:17:27.965214 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/83e51032-30aa-47cb-b856-36b4280c2775-operator-scripts\") pod \"83e51032-30aa-47cb-b856-36b4280c2775\" (UID: \"83e51032-30aa-47cb-b856-36b4280c2775\") " Nov 22 08:17:27 crc kubenswrapper[4789]: I1122 08:17:27.965715 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2xc5d\" (UniqueName: \"kubernetes.io/projected/83e51032-30aa-47cb-b856-36b4280c2775-kube-api-access-2xc5d\") pod \"83e51032-30aa-47cb-b856-36b4280c2775\" (UID: \"83e51032-30aa-47cb-b856-36b4280c2775\") " Nov 22 08:17:27 crc kubenswrapper[4789]: I1122 08:17:27.966692 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/83e51032-30aa-47cb-b856-36b4280c2775-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "83e51032-30aa-47cb-b856-36b4280c2775" (UID: "83e51032-30aa-47cb-b856-36b4280c2775"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:17:27 crc kubenswrapper[4789]: I1122 08:17:27.966857 4789 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/83e51032-30aa-47cb-b856-36b4280c2775-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:17:27 crc kubenswrapper[4789]: I1122 08:17:27.974302 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83e51032-30aa-47cb-b856-36b4280c2775-kube-api-access-2xc5d" (OuterVolumeSpecName: "kube-api-access-2xc5d") pod "83e51032-30aa-47cb-b856-36b4280c2775" (UID: "83e51032-30aa-47cb-b856-36b4280c2775"). InnerVolumeSpecName "kube-api-access-2xc5d". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:17:27 crc kubenswrapper[4789]: I1122 08:17:27.983150 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="56fd3074-a15c-4024-baf9-f784a50e14a5" path="/var/lib/kubelet/pods/56fd3074-a15c-4024-baf9-f784a50e14a5/volumes" Nov 22 08:17:28 crc kubenswrapper[4789]: I1122 08:17:28.038215 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-8cdf-account-create-hzl44" Nov 22 08:17:28 crc kubenswrapper[4789]: I1122 08:17:28.044767 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-vcmgb" Nov 22 08:17:28 crc kubenswrapper[4789]: I1122 08:17:28.053511 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-5mvvx" Nov 22 08:17:28 crc kubenswrapper[4789]: I1122 08:17:28.067827 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9e886f6c-babc-4118-b532-50b72dacb370-operator-scripts\") pod \"9e886f6c-babc-4118-b532-50b72dacb370\" (UID: \"9e886f6c-babc-4118-b532-50b72dacb370\") " Nov 22 08:17:28 crc kubenswrapper[4789]: I1122 08:17:28.068085 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8vfvc\" (UniqueName: \"kubernetes.io/projected/9e886f6c-babc-4118-b532-50b72dacb370-kube-api-access-8vfvc\") pod \"9e886f6c-babc-4118-b532-50b72dacb370\" (UID: \"9e886f6c-babc-4118-b532-50b72dacb370\") " Nov 22 08:17:28 crc kubenswrapper[4789]: I1122 08:17:28.068693 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2xc5d\" (UniqueName: \"kubernetes.io/projected/83e51032-30aa-47cb-b856-36b4280c2775-kube-api-access-2xc5d\") on node \"crc\" DevicePath \"\"" Nov 22 08:17:28 crc kubenswrapper[4789]: I1122 08:17:28.071742 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e886f6c-babc-4118-b532-50b72dacb370-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9e886f6c-babc-4118-b532-50b72dacb370" (UID: "9e886f6c-babc-4118-b532-50b72dacb370"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:17:28 crc kubenswrapper[4789]: I1122 08:17:28.088814 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e886f6c-babc-4118-b532-50b72dacb370-kube-api-access-8vfvc" (OuterVolumeSpecName: "kube-api-access-8vfvc") pod "9e886f6c-babc-4118-b532-50b72dacb370" (UID: "9e886f6c-babc-4118-b532-50b72dacb370"). InnerVolumeSpecName "kube-api-access-8vfvc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:17:28 crc kubenswrapper[4789]: I1122 08:17:28.170025 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/aa7f415d-4528-41d4-996f-57c061b72885-operator-scripts\") pod \"aa7f415d-4528-41d4-996f-57c061b72885\" (UID: \"aa7f415d-4528-41d4-996f-57c061b72885\") " Nov 22 08:17:28 crc kubenswrapper[4789]: I1122 08:17:28.170096 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/07941151-cfeb-4356-ad4e-439fdb875687-operator-scripts\") pod \"07941151-cfeb-4356-ad4e-439fdb875687\" (UID: \"07941151-cfeb-4356-ad4e-439fdb875687\") " Nov 22 08:17:28 crc kubenswrapper[4789]: I1122 08:17:28.170169 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ktvc4\" (UniqueName: \"kubernetes.io/projected/aa7f415d-4528-41d4-996f-57c061b72885-kube-api-access-ktvc4\") pod \"aa7f415d-4528-41d4-996f-57c061b72885\" (UID: \"aa7f415d-4528-41d4-996f-57c061b72885\") " Nov 22 08:17:28 crc kubenswrapper[4789]: I1122 08:17:28.170184 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xqjbz\" (UniqueName: \"kubernetes.io/projected/07941151-cfeb-4356-ad4e-439fdb875687-kube-api-access-xqjbz\") pod \"07941151-cfeb-4356-ad4e-439fdb875687\" (UID: \"07941151-cfeb-4356-ad4e-439fdb875687\") " Nov 22 08:17:28 crc kubenswrapper[4789]: I1122 08:17:28.170538 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8vfvc\" (UniqueName: \"kubernetes.io/projected/9e886f6c-babc-4118-b532-50b72dacb370-kube-api-access-8vfvc\") on node \"crc\" DevicePath \"\"" Nov 22 08:17:28 crc kubenswrapper[4789]: I1122 08:17:28.170556 4789 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9e886f6c-babc-4118-b532-50b72dacb370-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:17:28 crc kubenswrapper[4789]: I1122 08:17:28.170645 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aa7f415d-4528-41d4-996f-57c061b72885-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "aa7f415d-4528-41d4-996f-57c061b72885" (UID: "aa7f415d-4528-41d4-996f-57c061b72885"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:17:28 crc kubenswrapper[4789]: I1122 08:17:28.170707 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/07941151-cfeb-4356-ad4e-439fdb875687-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "07941151-cfeb-4356-ad4e-439fdb875687" (UID: "07941151-cfeb-4356-ad4e-439fdb875687"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:17:28 crc kubenswrapper[4789]: I1122 08:17:28.173567 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa7f415d-4528-41d4-996f-57c061b72885-kube-api-access-ktvc4" (OuterVolumeSpecName: "kube-api-access-ktvc4") pod "aa7f415d-4528-41d4-996f-57c061b72885" (UID: "aa7f415d-4528-41d4-996f-57c061b72885"). InnerVolumeSpecName "kube-api-access-ktvc4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:17:28 crc kubenswrapper[4789]: I1122 08:17:28.174331 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/07941151-cfeb-4356-ad4e-439fdb875687-kube-api-access-xqjbz" (OuterVolumeSpecName: "kube-api-access-xqjbz") pod "07941151-cfeb-4356-ad4e-439fdb875687" (UID: "07941151-cfeb-4356-ad4e-439fdb875687"). InnerVolumeSpecName "kube-api-access-xqjbz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:17:28 crc kubenswrapper[4789]: I1122 08:17:28.272585 4789 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/aa7f415d-4528-41d4-996f-57c061b72885-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:17:28 crc kubenswrapper[4789]: I1122 08:17:28.272634 4789 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/07941151-cfeb-4356-ad4e-439fdb875687-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:17:28 crc kubenswrapper[4789]: I1122 08:17:28.272646 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ktvc4\" (UniqueName: \"kubernetes.io/projected/aa7f415d-4528-41d4-996f-57c061b72885-kube-api-access-ktvc4\") on node \"crc\" DevicePath \"\"" Nov 22 08:17:28 crc kubenswrapper[4789]: I1122 08:17:28.272656 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xqjbz\" (UniqueName: \"kubernetes.io/projected/07941151-cfeb-4356-ad4e-439fdb875687-kube-api-access-xqjbz\") on node \"crc\" DevicePath \"\"" Nov 22 08:17:28 crc kubenswrapper[4789]: I1122 08:17:28.417635 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-frb4x" Nov 22 08:17:28 crc kubenswrapper[4789]: I1122 08:17:28.417654 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-frb4x" event={"ID":"83e51032-30aa-47cb-b856-36b4280c2775","Type":"ContainerDied","Data":"a957e281864c137f0975c5bf49cda635e544e9c9212baa2987aaad1c3af050e5"} Nov 22 08:17:28 crc kubenswrapper[4789]: I1122 08:17:28.417695 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a957e281864c137f0975c5bf49cda635e544e9c9212baa2987aaad1c3af050e5" Nov 22 08:17:28 crc kubenswrapper[4789]: I1122 08:17:28.422108 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-8cdf-account-create-hzl44" Nov 22 08:17:28 crc kubenswrapper[4789]: I1122 08:17:28.422116 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-8cdf-account-create-hzl44" event={"ID":"9e886f6c-babc-4118-b532-50b72dacb370","Type":"ContainerDied","Data":"e57ced5703dedf939d1ea98ec6641a85b0c2f0fec619262907f6a43e2a04690d"} Nov 22 08:17:28 crc kubenswrapper[4789]: I1122 08:17:28.422286 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e57ced5703dedf939d1ea98ec6641a85b0c2f0fec619262907f6a43e2a04690d" Nov 22 08:17:28 crc kubenswrapper[4789]: I1122 08:17:28.424247 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-5mvvx" event={"ID":"07941151-cfeb-4356-ad4e-439fdb875687","Type":"ContainerDied","Data":"ff00e592fbc1c20b184c645de8d3cffffeb606cf42703ba799db83d17bca13fd"} Nov 22 08:17:28 crc kubenswrapper[4789]: I1122 08:17:28.424263 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-5mvvx" Nov 22 08:17:28 crc kubenswrapper[4789]: I1122 08:17:28.424280 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ff00e592fbc1c20b184c645de8d3cffffeb606cf42703ba799db83d17bca13fd" Nov 22 08:17:28 crc kubenswrapper[4789]: I1122 08:17:28.426079 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-vcmgb" event={"ID":"aa7f415d-4528-41d4-996f-57c061b72885","Type":"ContainerDied","Data":"58a4d35171fcbbdb2d527cadd57dc241e05138388bc58b1376eb5c551ccd8c22"} Nov 22 08:17:28 crc kubenswrapper[4789]: I1122 08:17:28.426110 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="58a4d35171fcbbdb2d527cadd57dc241e05138388bc58b1376eb5c551ccd8c22" Nov 22 08:17:28 crc kubenswrapper[4789]: I1122 08:17:28.426196 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-vcmgb" Nov 22 08:17:28 crc kubenswrapper[4789]: I1122 08:17:28.661487 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-9a6c-account-create-2cwpt" Nov 22 08:17:28 crc kubenswrapper[4789]: I1122 08:17:28.782436 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sxhw9\" (UniqueName: \"kubernetes.io/projected/d315bf00-e54d-40cb-a261-ff8bd068e535-kube-api-access-sxhw9\") pod \"d315bf00-e54d-40cb-a261-ff8bd068e535\" (UID: \"d315bf00-e54d-40cb-a261-ff8bd068e535\") " Nov 22 08:17:28 crc kubenswrapper[4789]: I1122 08:17:28.782502 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d315bf00-e54d-40cb-a261-ff8bd068e535-operator-scripts\") pod \"d315bf00-e54d-40cb-a261-ff8bd068e535\" (UID: \"d315bf00-e54d-40cb-a261-ff8bd068e535\") " Nov 22 08:17:28 crc kubenswrapper[4789]: I1122 08:17:28.783512 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d315bf00-e54d-40cb-a261-ff8bd068e535-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d315bf00-e54d-40cb-a261-ff8bd068e535" (UID: "d315bf00-e54d-40cb-a261-ff8bd068e535"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:17:28 crc kubenswrapper[4789]: I1122 08:17:28.789611 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d315bf00-e54d-40cb-a261-ff8bd068e535-kube-api-access-sxhw9" (OuterVolumeSpecName: "kube-api-access-sxhw9") pod "d315bf00-e54d-40cb-a261-ff8bd068e535" (UID: "d315bf00-e54d-40cb-a261-ff8bd068e535"). InnerVolumeSpecName "kube-api-access-sxhw9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:17:28 crc kubenswrapper[4789]: I1122 08:17:28.832710 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-8958-account-create-626c2" Nov 22 08:17:28 crc kubenswrapper[4789]: I1122 08:17:28.884719 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vpgtl\" (UniqueName: \"kubernetes.io/projected/c61136a6-9398-4fec-89a0-e4d7291b99fc-kube-api-access-vpgtl\") pod \"c61136a6-9398-4fec-89a0-e4d7291b99fc\" (UID: \"c61136a6-9398-4fec-89a0-e4d7291b99fc\") " Nov 22 08:17:28 crc kubenswrapper[4789]: I1122 08:17:28.884902 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c61136a6-9398-4fec-89a0-e4d7291b99fc-operator-scripts\") pod \"c61136a6-9398-4fec-89a0-e4d7291b99fc\" (UID: \"c61136a6-9398-4fec-89a0-e4d7291b99fc\") " Nov 22 08:17:28 crc kubenswrapper[4789]: I1122 08:17:28.885270 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sxhw9\" (UniqueName: \"kubernetes.io/projected/d315bf00-e54d-40cb-a261-ff8bd068e535-kube-api-access-sxhw9\") on node \"crc\" DevicePath \"\"" Nov 22 08:17:28 crc kubenswrapper[4789]: I1122 08:17:28.885285 4789 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d315bf00-e54d-40cb-a261-ff8bd068e535-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:17:28 crc kubenswrapper[4789]: I1122 08:17:28.885555 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c61136a6-9398-4fec-89a0-e4d7291b99fc-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c61136a6-9398-4fec-89a0-e4d7291b99fc" (UID: "c61136a6-9398-4fec-89a0-e4d7291b99fc"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:17:28 crc kubenswrapper[4789]: I1122 08:17:28.887489 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c61136a6-9398-4fec-89a0-e4d7291b99fc-kube-api-access-vpgtl" (OuterVolumeSpecName: "kube-api-access-vpgtl") pod "c61136a6-9398-4fec-89a0-e4d7291b99fc" (UID: "c61136a6-9398-4fec-89a0-e4d7291b99fc"). InnerVolumeSpecName "kube-api-access-vpgtl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:17:28 crc kubenswrapper[4789]: I1122 08:17:28.987191 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vpgtl\" (UniqueName: \"kubernetes.io/projected/c61136a6-9398-4fec-89a0-e4d7291b99fc-kube-api-access-vpgtl\") on node \"crc\" DevicePath \"\"" Nov 22 08:17:28 crc kubenswrapper[4789]: I1122 08:17:28.987229 4789 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c61136a6-9398-4fec-89a0-e4d7291b99fc-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:17:29 crc kubenswrapper[4789]: I1122 08:17:29.433472 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-9a6c-account-create-2cwpt" event={"ID":"d315bf00-e54d-40cb-a261-ff8bd068e535","Type":"ContainerDied","Data":"441ab1a55c0b99d631a57d777c3da00c561cd44d242e70638b828eafbc1a2284"} Nov 22 08:17:29 crc kubenswrapper[4789]: I1122 08:17:29.433820 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="441ab1a55c0b99d631a57d777c3da00c561cd44d242e70638b828eafbc1a2284" Nov 22 08:17:29 crc kubenswrapper[4789]: I1122 08:17:29.433511 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-9a6c-account-create-2cwpt" Nov 22 08:17:29 crc kubenswrapper[4789]: I1122 08:17:29.436484 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-8958-account-create-626c2" event={"ID":"c61136a6-9398-4fec-89a0-e4d7291b99fc","Type":"ContainerDied","Data":"d0d43f2bd491b2aa0fa3a787614d3055caba5ed14b6ced394e05243a85efd573"} Nov 22 08:17:29 crc kubenswrapper[4789]: I1122 08:17:29.436527 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d0d43f2bd491b2aa0fa3a787614d3055caba5ed14b6ced394e05243a85efd573" Nov 22 08:17:29 crc kubenswrapper[4789]: I1122 08:17:29.436669 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-8958-account-create-626c2" Nov 22 08:17:30 crc kubenswrapper[4789]: I1122 08:17:30.448560 4789 generic.go:334] "Generic (PLEG): container finished" podID="2949ebe1-7541-42a2-894c-2588c2bd3396" containerID="77b48a1183833a07016416adb6c59b89b3874df9589e6f5b1c3dba3285e81de8" exitCode=0 Nov 22 08:17:30 crc kubenswrapper[4789]: I1122 08:17:30.448615 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-568cd49f4d-ms9nh" event={"ID":"2949ebe1-7541-42a2-894c-2588c2bd3396","Type":"ContainerDied","Data":"77b48a1183833a07016416adb6c59b89b3874df9589e6f5b1c3dba3285e81de8"} Nov 22 08:17:30 crc kubenswrapper[4789]: I1122 08:17:30.451553 4789 generic.go:334] "Generic (PLEG): container finished" podID="de991bf6-8f96-4c99-bf7a-b8909c0c80d2" containerID="63d2e8a684de408f5cfa4728b23b07803ae5286c76ec7a0b812bf69230a7922f" exitCode=0 Nov 22 08:17:30 crc kubenswrapper[4789]: I1122 08:17:30.451586 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"de991bf6-8f96-4c99-bf7a-b8909c0c80d2","Type":"ContainerDied","Data":"63d2e8a684de408f5cfa4728b23b07803ae5286c76ec7a0b812bf69230a7922f"} Nov 22 08:17:30 crc kubenswrapper[4789]: I1122 08:17:30.546941 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 08:17:30 crc kubenswrapper[4789]: I1122 08:17:30.618472 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/de991bf6-8f96-4c99-bf7a-b8909c0c80d2-log-httpd\") pod \"de991bf6-8f96-4c99-bf7a-b8909c0c80d2\" (UID: \"de991bf6-8f96-4c99-bf7a-b8909c0c80d2\") " Nov 22 08:17:30 crc kubenswrapper[4789]: I1122 08:17:30.618585 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/de991bf6-8f96-4c99-bf7a-b8909c0c80d2-sg-core-conf-yaml\") pod \"de991bf6-8f96-4c99-bf7a-b8909c0c80d2\" (UID: \"de991bf6-8f96-4c99-bf7a-b8909c0c80d2\") " Nov 22 08:17:30 crc kubenswrapper[4789]: I1122 08:17:30.618644 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/de991bf6-8f96-4c99-bf7a-b8909c0c80d2-scripts\") pod \"de991bf6-8f96-4c99-bf7a-b8909c0c80d2\" (UID: \"de991bf6-8f96-4c99-bf7a-b8909c0c80d2\") " Nov 22 08:17:30 crc kubenswrapper[4789]: I1122 08:17:30.618696 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ppk69\" (UniqueName: \"kubernetes.io/projected/de991bf6-8f96-4c99-bf7a-b8909c0c80d2-kube-api-access-ppk69\") pod \"de991bf6-8f96-4c99-bf7a-b8909c0c80d2\" (UID: \"de991bf6-8f96-4c99-bf7a-b8909c0c80d2\") " Nov 22 08:17:30 crc kubenswrapper[4789]: I1122 08:17:30.618728 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/de991bf6-8f96-4c99-bf7a-b8909c0c80d2-run-httpd\") pod \"de991bf6-8f96-4c99-bf7a-b8909c0c80d2\" (UID: \"de991bf6-8f96-4c99-bf7a-b8909c0c80d2\") " Nov 22 08:17:30 crc kubenswrapper[4789]: I1122 08:17:30.618793 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de991bf6-8f96-4c99-bf7a-b8909c0c80d2-combined-ca-bundle\") pod \"de991bf6-8f96-4c99-bf7a-b8909c0c80d2\" (UID: \"de991bf6-8f96-4c99-bf7a-b8909c0c80d2\") " Nov 22 08:17:30 crc kubenswrapper[4789]: I1122 08:17:30.618847 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de991bf6-8f96-4c99-bf7a-b8909c0c80d2-config-data\") pod \"de991bf6-8f96-4c99-bf7a-b8909c0c80d2\" (UID: \"de991bf6-8f96-4c99-bf7a-b8909c0c80d2\") " Nov 22 08:17:30 crc kubenswrapper[4789]: I1122 08:17:30.619263 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/de991bf6-8f96-4c99-bf7a-b8909c0c80d2-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "de991bf6-8f96-4c99-bf7a-b8909c0c80d2" (UID: "de991bf6-8f96-4c99-bf7a-b8909c0c80d2"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:17:30 crc kubenswrapper[4789]: I1122 08:17:30.620045 4789 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/de991bf6-8f96-4c99-bf7a-b8909c0c80d2-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 22 08:17:30 crc kubenswrapper[4789]: I1122 08:17:30.620227 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/de991bf6-8f96-4c99-bf7a-b8909c0c80d2-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "de991bf6-8f96-4c99-bf7a-b8909c0c80d2" (UID: "de991bf6-8f96-4c99-bf7a-b8909c0c80d2"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:17:30 crc kubenswrapper[4789]: I1122 08:17:30.635152 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de991bf6-8f96-4c99-bf7a-b8909c0c80d2-kube-api-access-ppk69" (OuterVolumeSpecName: "kube-api-access-ppk69") pod "de991bf6-8f96-4c99-bf7a-b8909c0c80d2" (UID: "de991bf6-8f96-4c99-bf7a-b8909c0c80d2"). InnerVolumeSpecName "kube-api-access-ppk69". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:17:30 crc kubenswrapper[4789]: I1122 08:17:30.661629 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de991bf6-8f96-4c99-bf7a-b8909c0c80d2-scripts" (OuterVolumeSpecName: "scripts") pod "de991bf6-8f96-4c99-bf7a-b8909c0c80d2" (UID: "de991bf6-8f96-4c99-bf7a-b8909c0c80d2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:17:30 crc kubenswrapper[4789]: I1122 08:17:30.720285 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de991bf6-8f96-4c99-bf7a-b8909c0c80d2-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "de991bf6-8f96-4c99-bf7a-b8909c0c80d2" (UID: "de991bf6-8f96-4c99-bf7a-b8909c0c80d2"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:17:30 crc kubenswrapper[4789]: I1122 08:17:30.724230 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/de991bf6-8f96-4c99-bf7a-b8909c0c80d2-sg-core-conf-yaml\") pod \"de991bf6-8f96-4c99-bf7a-b8909c0c80d2\" (UID: \"de991bf6-8f96-4c99-bf7a-b8909c0c80d2\") " Nov 22 08:17:30 crc kubenswrapper[4789]: I1122 08:17:30.724943 4789 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/de991bf6-8f96-4c99-bf7a-b8909c0c80d2-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:17:30 crc kubenswrapper[4789]: I1122 08:17:30.724973 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ppk69\" (UniqueName: \"kubernetes.io/projected/de991bf6-8f96-4c99-bf7a-b8909c0c80d2-kube-api-access-ppk69\") on node \"crc\" DevicePath \"\"" Nov 22 08:17:30 crc kubenswrapper[4789]: I1122 08:17:30.724987 4789 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/de991bf6-8f96-4c99-bf7a-b8909c0c80d2-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 22 08:17:30 crc kubenswrapper[4789]: W1122 08:17:30.725111 4789 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/de991bf6-8f96-4c99-bf7a-b8909c0c80d2/volumes/kubernetes.io~secret/sg-core-conf-yaml Nov 22 08:17:30 crc kubenswrapper[4789]: I1122 08:17:30.725126 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de991bf6-8f96-4c99-bf7a-b8909c0c80d2-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "de991bf6-8f96-4c99-bf7a-b8909c0c80d2" (UID: "de991bf6-8f96-4c99-bf7a-b8909c0c80d2"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:17:30 crc kubenswrapper[4789]: I1122 08:17:30.777012 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-568cd49f4d-ms9nh" Nov 22 08:17:30 crc kubenswrapper[4789]: I1122 08:17:30.791846 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de991bf6-8f96-4c99-bf7a-b8909c0c80d2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "de991bf6-8f96-4c99-bf7a-b8909c0c80d2" (UID: "de991bf6-8f96-4c99-bf7a-b8909c0c80d2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:17:30 crc kubenswrapper[4789]: I1122 08:17:30.825620 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/2949ebe1-7541-42a2-894c-2588c2bd3396-httpd-config\") pod \"2949ebe1-7541-42a2-894c-2588c2bd3396\" (UID: \"2949ebe1-7541-42a2-894c-2588c2bd3396\") " Nov 22 08:17:30 crc kubenswrapper[4789]: I1122 08:17:30.825791 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/2949ebe1-7541-42a2-894c-2588c2bd3396-config\") pod \"2949ebe1-7541-42a2-894c-2588c2bd3396\" (UID: \"2949ebe1-7541-42a2-894c-2588c2bd3396\") " Nov 22 08:17:30 crc kubenswrapper[4789]: I1122 08:17:30.825854 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4jjl7\" (UniqueName: \"kubernetes.io/projected/2949ebe1-7541-42a2-894c-2588c2bd3396-kube-api-access-4jjl7\") pod \"2949ebe1-7541-42a2-894c-2588c2bd3396\" (UID: \"2949ebe1-7541-42a2-894c-2588c2bd3396\") " Nov 22 08:17:30 crc kubenswrapper[4789]: I1122 08:17:30.825913 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/2949ebe1-7541-42a2-894c-2588c2bd3396-ovndb-tls-certs\") pod \"2949ebe1-7541-42a2-894c-2588c2bd3396\" (UID: \"2949ebe1-7541-42a2-894c-2588c2bd3396\") " Nov 22 08:17:30 crc kubenswrapper[4789]: I1122 08:17:30.826037 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2949ebe1-7541-42a2-894c-2588c2bd3396-combined-ca-bundle\") pod \"2949ebe1-7541-42a2-894c-2588c2bd3396\" (UID: \"2949ebe1-7541-42a2-894c-2588c2bd3396\") " Nov 22 08:17:30 crc kubenswrapper[4789]: I1122 08:17:30.826509 4789 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de991bf6-8f96-4c99-bf7a-b8909c0c80d2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:17:30 crc kubenswrapper[4789]: I1122 08:17:30.826527 4789 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/de991bf6-8f96-4c99-bf7a-b8909c0c80d2-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 22 08:17:30 crc kubenswrapper[4789]: I1122 08:17:30.833509 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2949ebe1-7541-42a2-894c-2588c2bd3396-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "2949ebe1-7541-42a2-894c-2588c2bd3396" (UID: "2949ebe1-7541-42a2-894c-2588c2bd3396"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:17:30 crc kubenswrapper[4789]: I1122 08:17:30.837069 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2949ebe1-7541-42a2-894c-2588c2bd3396-kube-api-access-4jjl7" (OuterVolumeSpecName: "kube-api-access-4jjl7") pod "2949ebe1-7541-42a2-894c-2588c2bd3396" (UID: "2949ebe1-7541-42a2-894c-2588c2bd3396"). InnerVolumeSpecName "kube-api-access-4jjl7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:17:30 crc kubenswrapper[4789]: I1122 08:17:30.850662 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de991bf6-8f96-4c99-bf7a-b8909c0c80d2-config-data" (OuterVolumeSpecName: "config-data") pod "de991bf6-8f96-4c99-bf7a-b8909c0c80d2" (UID: "de991bf6-8f96-4c99-bf7a-b8909c0c80d2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:17:30 crc kubenswrapper[4789]: I1122 08:17:30.895143 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2949ebe1-7541-42a2-894c-2588c2bd3396-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2949ebe1-7541-42a2-894c-2588c2bd3396" (UID: "2949ebe1-7541-42a2-894c-2588c2bd3396"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:17:30 crc kubenswrapper[4789]: I1122 08:17:30.905012 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2949ebe1-7541-42a2-894c-2588c2bd3396-config" (OuterVolumeSpecName: "config") pod "2949ebe1-7541-42a2-894c-2588c2bd3396" (UID: "2949ebe1-7541-42a2-894c-2588c2bd3396"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:17:30 crc kubenswrapper[4789]: I1122 08:17:30.929459 4789 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2949ebe1-7541-42a2-894c-2588c2bd3396-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:17:30 crc kubenswrapper[4789]: I1122 08:17:30.929501 4789 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/2949ebe1-7541-42a2-894c-2588c2bd3396-httpd-config\") on node \"crc\" DevicePath \"\"" Nov 22 08:17:30 crc kubenswrapper[4789]: I1122 08:17:30.929512 4789 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/2949ebe1-7541-42a2-894c-2588c2bd3396-config\") on node \"crc\" DevicePath \"\"" Nov 22 08:17:30 crc kubenswrapper[4789]: I1122 08:17:30.929520 4789 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de991bf6-8f96-4c99-bf7a-b8909c0c80d2-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:17:30 crc kubenswrapper[4789]: I1122 08:17:30.929529 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4jjl7\" (UniqueName: \"kubernetes.io/projected/2949ebe1-7541-42a2-894c-2588c2bd3396-kube-api-access-4jjl7\") on node \"crc\" DevicePath \"\"" Nov 22 08:17:30 crc kubenswrapper[4789]: I1122 08:17:30.934457 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2949ebe1-7541-42a2-894c-2588c2bd3396-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "2949ebe1-7541-42a2-894c-2588c2bd3396" (UID: "2949ebe1-7541-42a2-894c-2588c2bd3396"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:17:31 crc kubenswrapper[4789]: I1122 08:17:31.030950 4789 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/2949ebe1-7541-42a2-894c-2588c2bd3396-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 22 08:17:31 crc kubenswrapper[4789]: I1122 08:17:31.474239 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"de991bf6-8f96-4c99-bf7a-b8909c0c80d2","Type":"ContainerDied","Data":"6aa40ca2ed1346bdebd7b6c732bc0d934c01d47edfac917c83aafa706fec5731"} Nov 22 08:17:31 crc kubenswrapper[4789]: I1122 08:17:31.474300 4789 scope.go:117] "RemoveContainer" containerID="4b4d49b0d5e06a9d2af814c8237d522b72d09606a55f523f315c307cfa7163bf" Nov 22 08:17:31 crc kubenswrapper[4789]: I1122 08:17:31.479214 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 08:17:31 crc kubenswrapper[4789]: I1122 08:17:31.485080 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-568cd49f4d-ms9nh" event={"ID":"2949ebe1-7541-42a2-894c-2588c2bd3396","Type":"ContainerDied","Data":"fde8262fa183d29bdeda97160e0de154acc0b30e8d06355d1c99fb047c5f1ba0"} Nov 22 08:17:31 crc kubenswrapper[4789]: I1122 08:17:31.485150 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-568cd49f4d-ms9nh" Nov 22 08:17:31 crc kubenswrapper[4789]: I1122 08:17:31.513584 4789 scope.go:117] "RemoveContainer" containerID="6168d229a82e8b6c67cebc486d47d83616de74ed1dcc8f74ca5a8473c64386f3" Nov 22 08:17:31 crc kubenswrapper[4789]: I1122 08:17:31.519818 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:17:31 crc kubenswrapper[4789]: I1122 08:17:31.531999 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:17:31 crc kubenswrapper[4789]: I1122 08:17:31.541900 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-568cd49f4d-ms9nh"] Nov 22 08:17:31 crc kubenswrapper[4789]: I1122 08:17:31.559034 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-568cd49f4d-ms9nh"] Nov 22 08:17:31 crc kubenswrapper[4789]: I1122 08:17:31.564321 4789 scope.go:117] "RemoveContainer" containerID="63d2e8a684de408f5cfa4728b23b07803ae5286c76ec7a0b812bf69230a7922f" Nov 22 08:17:31 crc kubenswrapper[4789]: I1122 08:17:31.575661 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:17:31 crc kubenswrapper[4789]: E1122 08:17:31.576140 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07941151-cfeb-4356-ad4e-439fdb875687" containerName="mariadb-database-create" Nov 22 08:17:31 crc kubenswrapper[4789]: I1122 08:17:31.576164 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="07941151-cfeb-4356-ad4e-439fdb875687" containerName="mariadb-database-create" Nov 22 08:17:31 crc kubenswrapper[4789]: E1122 08:17:31.576179 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56fd3074-a15c-4024-baf9-f784a50e14a5" containerName="horizon" Nov 22 08:17:31 crc kubenswrapper[4789]: I1122 08:17:31.576187 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="56fd3074-a15c-4024-baf9-f784a50e14a5" containerName="horizon" Nov 22 08:17:31 crc kubenswrapper[4789]: E1122 08:17:31.576205 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa7f415d-4528-41d4-996f-57c061b72885" containerName="mariadb-database-create" Nov 22 08:17:31 crc kubenswrapper[4789]: I1122 08:17:31.576213 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa7f415d-4528-41d4-996f-57c061b72885" containerName="mariadb-database-create" Nov 22 08:17:31 crc kubenswrapper[4789]: E1122 08:17:31.576228 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de991bf6-8f96-4c99-bf7a-b8909c0c80d2" containerName="sg-core" Nov 22 08:17:31 crc kubenswrapper[4789]: I1122 08:17:31.576234 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="de991bf6-8f96-4c99-bf7a-b8909c0c80d2" containerName="sg-core" Nov 22 08:17:31 crc kubenswrapper[4789]: E1122 08:17:31.576243 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e886f6c-babc-4118-b532-50b72dacb370" containerName="mariadb-account-create" Nov 22 08:17:31 crc kubenswrapper[4789]: I1122 08:17:31.576249 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e886f6c-babc-4118-b532-50b72dacb370" containerName="mariadb-account-create" Nov 22 08:17:31 crc kubenswrapper[4789]: E1122 08:17:31.576255 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c61136a6-9398-4fec-89a0-e4d7291b99fc" containerName="mariadb-account-create" Nov 22 08:17:31 crc kubenswrapper[4789]: I1122 08:17:31.576263 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="c61136a6-9398-4fec-89a0-e4d7291b99fc" containerName="mariadb-account-create" Nov 22 08:17:31 crc kubenswrapper[4789]: E1122 08:17:31.576272 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d315bf00-e54d-40cb-a261-ff8bd068e535" containerName="mariadb-account-create" Nov 22 08:17:31 crc kubenswrapper[4789]: I1122 08:17:31.576280 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="d315bf00-e54d-40cb-a261-ff8bd068e535" containerName="mariadb-account-create" Nov 22 08:17:31 crc kubenswrapper[4789]: E1122 08:17:31.576291 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2949ebe1-7541-42a2-894c-2588c2bd3396" containerName="neutron-api" Nov 22 08:17:31 crc kubenswrapper[4789]: I1122 08:17:31.576299 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="2949ebe1-7541-42a2-894c-2588c2bd3396" containerName="neutron-api" Nov 22 08:17:31 crc kubenswrapper[4789]: E1122 08:17:31.576312 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2949ebe1-7541-42a2-894c-2588c2bd3396" containerName="neutron-httpd" Nov 22 08:17:31 crc kubenswrapper[4789]: I1122 08:17:31.576318 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="2949ebe1-7541-42a2-894c-2588c2bd3396" containerName="neutron-httpd" Nov 22 08:17:31 crc kubenswrapper[4789]: E1122 08:17:31.576328 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de991bf6-8f96-4c99-bf7a-b8909c0c80d2" containerName="ceilometer-central-agent" Nov 22 08:17:31 crc kubenswrapper[4789]: I1122 08:17:31.576334 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="de991bf6-8f96-4c99-bf7a-b8909c0c80d2" containerName="ceilometer-central-agent" Nov 22 08:17:31 crc kubenswrapper[4789]: E1122 08:17:31.576346 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83e51032-30aa-47cb-b856-36b4280c2775" containerName="mariadb-database-create" Nov 22 08:17:31 crc kubenswrapper[4789]: I1122 08:17:31.576351 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="83e51032-30aa-47cb-b856-36b4280c2775" containerName="mariadb-database-create" Nov 22 08:17:31 crc kubenswrapper[4789]: E1122 08:17:31.576364 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56fd3074-a15c-4024-baf9-f784a50e14a5" containerName="horizon-log" Nov 22 08:17:31 crc kubenswrapper[4789]: I1122 08:17:31.576369 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="56fd3074-a15c-4024-baf9-f784a50e14a5" containerName="horizon-log" Nov 22 08:17:31 crc kubenswrapper[4789]: E1122 08:17:31.576380 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de991bf6-8f96-4c99-bf7a-b8909c0c80d2" containerName="proxy-httpd" Nov 22 08:17:31 crc kubenswrapper[4789]: I1122 08:17:31.576385 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="de991bf6-8f96-4c99-bf7a-b8909c0c80d2" containerName="proxy-httpd" Nov 22 08:17:31 crc kubenswrapper[4789]: E1122 08:17:31.576400 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de991bf6-8f96-4c99-bf7a-b8909c0c80d2" containerName="ceilometer-notification-agent" Nov 22 08:17:31 crc kubenswrapper[4789]: I1122 08:17:31.576406 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="de991bf6-8f96-4c99-bf7a-b8909c0c80d2" containerName="ceilometer-notification-agent" Nov 22 08:17:31 crc kubenswrapper[4789]: I1122 08:17:31.576557 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="de991bf6-8f96-4c99-bf7a-b8909c0c80d2" containerName="ceilometer-central-agent" Nov 22 08:17:31 crc kubenswrapper[4789]: I1122 08:17:31.576570 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="de991bf6-8f96-4c99-bf7a-b8909c0c80d2" containerName="proxy-httpd" Nov 22 08:17:31 crc kubenswrapper[4789]: I1122 08:17:31.576581 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="56fd3074-a15c-4024-baf9-f784a50e14a5" containerName="horizon" Nov 22 08:17:31 crc kubenswrapper[4789]: I1122 08:17:31.576592 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="07941151-cfeb-4356-ad4e-439fdb875687" containerName="mariadb-database-create" Nov 22 08:17:31 crc kubenswrapper[4789]: I1122 08:17:31.576602 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="de991bf6-8f96-4c99-bf7a-b8909c0c80d2" containerName="ceilometer-notification-agent" Nov 22 08:17:31 crc kubenswrapper[4789]: I1122 08:17:31.576612 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="83e51032-30aa-47cb-b856-36b4280c2775" containerName="mariadb-database-create" Nov 22 08:17:31 crc kubenswrapper[4789]: I1122 08:17:31.576622 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="2949ebe1-7541-42a2-894c-2588c2bd3396" containerName="neutron-httpd" Nov 22 08:17:31 crc kubenswrapper[4789]: I1122 08:17:31.576630 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="2949ebe1-7541-42a2-894c-2588c2bd3396" containerName="neutron-api" Nov 22 08:17:31 crc kubenswrapper[4789]: I1122 08:17:31.576638 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa7f415d-4528-41d4-996f-57c061b72885" containerName="mariadb-database-create" Nov 22 08:17:31 crc kubenswrapper[4789]: I1122 08:17:31.576651 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="d315bf00-e54d-40cb-a261-ff8bd068e535" containerName="mariadb-account-create" Nov 22 08:17:31 crc kubenswrapper[4789]: I1122 08:17:31.576658 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="56fd3074-a15c-4024-baf9-f784a50e14a5" containerName="horizon-log" Nov 22 08:17:31 crc kubenswrapper[4789]: I1122 08:17:31.576669 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e886f6c-babc-4118-b532-50b72dacb370" containerName="mariadb-account-create" Nov 22 08:17:31 crc kubenswrapper[4789]: I1122 08:17:31.576678 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="c61136a6-9398-4fec-89a0-e4d7291b99fc" containerName="mariadb-account-create" Nov 22 08:17:31 crc kubenswrapper[4789]: I1122 08:17:31.576686 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="de991bf6-8f96-4c99-bf7a-b8909c0c80d2" containerName="sg-core" Nov 22 08:17:31 crc kubenswrapper[4789]: I1122 08:17:31.578409 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 08:17:31 crc kubenswrapper[4789]: I1122 08:17:31.583866 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 22 08:17:31 crc kubenswrapper[4789]: I1122 08:17:31.583897 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 22 08:17:31 crc kubenswrapper[4789]: I1122 08:17:31.584021 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 22 08:17:31 crc kubenswrapper[4789]: I1122 08:17:31.584310 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="20a309c0-4fac-480e-b19d-e2ca480a6b6c" containerName="kube-state-metrics" containerID="cri-o://39f28765bd714106b4fd5010156d2c44ca1777ed061f27f9555f21c3d40d7898" gracePeriod=30 Nov 22 08:17:31 crc kubenswrapper[4789]: I1122 08:17:31.594780 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:17:31 crc kubenswrapper[4789]: I1122 08:17:31.604990 4789 scope.go:117] "RemoveContainer" containerID="60c80dac437908b708a2f24653527c9e713047e6113c781463535e492f681d00" Nov 22 08:17:31 crc kubenswrapper[4789]: I1122 08:17:31.641220 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e740aa5b-0619-4092-b893-1ea056025be8-run-httpd\") pod \"ceilometer-0\" (UID: \"e740aa5b-0619-4092-b893-1ea056025be8\") " pod="openstack/ceilometer-0" Nov 22 08:17:31 crc kubenswrapper[4789]: I1122 08:17:31.641292 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e740aa5b-0619-4092-b893-1ea056025be8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e740aa5b-0619-4092-b893-1ea056025be8\") " pod="openstack/ceilometer-0" Nov 22 08:17:31 crc kubenswrapper[4789]: I1122 08:17:31.641309 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e740aa5b-0619-4092-b893-1ea056025be8-scripts\") pod \"ceilometer-0\" (UID: \"e740aa5b-0619-4092-b893-1ea056025be8\") " pod="openstack/ceilometer-0" Nov 22 08:17:31 crc kubenswrapper[4789]: I1122 08:17:31.641666 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e740aa5b-0619-4092-b893-1ea056025be8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e740aa5b-0619-4092-b893-1ea056025be8\") " pod="openstack/ceilometer-0" Nov 22 08:17:31 crc kubenswrapper[4789]: I1122 08:17:31.641727 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rc7rz\" (UniqueName: \"kubernetes.io/projected/e740aa5b-0619-4092-b893-1ea056025be8-kube-api-access-rc7rz\") pod \"ceilometer-0\" (UID: \"e740aa5b-0619-4092-b893-1ea056025be8\") " pod="openstack/ceilometer-0" Nov 22 08:17:31 crc kubenswrapper[4789]: I1122 08:17:31.641893 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e740aa5b-0619-4092-b893-1ea056025be8-log-httpd\") pod \"ceilometer-0\" (UID: \"e740aa5b-0619-4092-b893-1ea056025be8\") " pod="openstack/ceilometer-0" Nov 22 08:17:31 crc kubenswrapper[4789]: I1122 08:17:31.642004 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e740aa5b-0619-4092-b893-1ea056025be8-config-data\") pod \"ceilometer-0\" (UID: \"e740aa5b-0619-4092-b893-1ea056025be8\") " pod="openstack/ceilometer-0" Nov 22 08:17:31 crc kubenswrapper[4789]: I1122 08:17:31.668002 4789 scope.go:117] "RemoveContainer" containerID="9aec8c709e300ab397822f5fff34aba93e0915d5c8707aac0584b703bc60df1a" Nov 22 08:17:31 crc kubenswrapper[4789]: I1122 08:17:31.692997 4789 scope.go:117] "RemoveContainer" containerID="77b48a1183833a07016416adb6c59b89b3874df9589e6f5b1c3dba3285e81de8" Nov 22 08:17:31 crc kubenswrapper[4789]: I1122 08:17:31.746102 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e740aa5b-0619-4092-b893-1ea056025be8-config-data\") pod \"ceilometer-0\" (UID: \"e740aa5b-0619-4092-b893-1ea056025be8\") " pod="openstack/ceilometer-0" Nov 22 08:17:31 crc kubenswrapper[4789]: I1122 08:17:31.746191 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e740aa5b-0619-4092-b893-1ea056025be8-run-httpd\") pod \"ceilometer-0\" (UID: \"e740aa5b-0619-4092-b893-1ea056025be8\") " pod="openstack/ceilometer-0" Nov 22 08:17:31 crc kubenswrapper[4789]: I1122 08:17:31.746283 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e740aa5b-0619-4092-b893-1ea056025be8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e740aa5b-0619-4092-b893-1ea056025be8\") " pod="openstack/ceilometer-0" Nov 22 08:17:31 crc kubenswrapper[4789]: I1122 08:17:31.746304 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e740aa5b-0619-4092-b893-1ea056025be8-scripts\") pod \"ceilometer-0\" (UID: \"e740aa5b-0619-4092-b893-1ea056025be8\") " pod="openstack/ceilometer-0" Nov 22 08:17:31 crc kubenswrapper[4789]: I1122 08:17:31.746406 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e740aa5b-0619-4092-b893-1ea056025be8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e740aa5b-0619-4092-b893-1ea056025be8\") " pod="openstack/ceilometer-0" Nov 22 08:17:31 crc kubenswrapper[4789]: I1122 08:17:31.746430 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rc7rz\" (UniqueName: \"kubernetes.io/projected/e740aa5b-0619-4092-b893-1ea056025be8-kube-api-access-rc7rz\") pod \"ceilometer-0\" (UID: \"e740aa5b-0619-4092-b893-1ea056025be8\") " pod="openstack/ceilometer-0" Nov 22 08:17:31 crc kubenswrapper[4789]: I1122 08:17:31.746488 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e740aa5b-0619-4092-b893-1ea056025be8-log-httpd\") pod \"ceilometer-0\" (UID: \"e740aa5b-0619-4092-b893-1ea056025be8\") " pod="openstack/ceilometer-0" Nov 22 08:17:31 crc kubenswrapper[4789]: I1122 08:17:31.747036 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e740aa5b-0619-4092-b893-1ea056025be8-log-httpd\") pod \"ceilometer-0\" (UID: \"e740aa5b-0619-4092-b893-1ea056025be8\") " pod="openstack/ceilometer-0" Nov 22 08:17:31 crc kubenswrapper[4789]: I1122 08:17:31.748131 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e740aa5b-0619-4092-b893-1ea056025be8-run-httpd\") pod \"ceilometer-0\" (UID: \"e740aa5b-0619-4092-b893-1ea056025be8\") " pod="openstack/ceilometer-0" Nov 22 08:17:31 crc kubenswrapper[4789]: I1122 08:17:31.752971 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e740aa5b-0619-4092-b893-1ea056025be8-config-data\") pod \"ceilometer-0\" (UID: \"e740aa5b-0619-4092-b893-1ea056025be8\") " pod="openstack/ceilometer-0" Nov 22 08:17:31 crc kubenswrapper[4789]: I1122 08:17:31.755899 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e740aa5b-0619-4092-b893-1ea056025be8-scripts\") pod \"ceilometer-0\" (UID: \"e740aa5b-0619-4092-b893-1ea056025be8\") " pod="openstack/ceilometer-0" Nov 22 08:17:31 crc kubenswrapper[4789]: I1122 08:17:31.758294 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e740aa5b-0619-4092-b893-1ea056025be8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e740aa5b-0619-4092-b893-1ea056025be8\") " pod="openstack/ceilometer-0" Nov 22 08:17:31 crc kubenswrapper[4789]: I1122 08:17:31.760530 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e740aa5b-0619-4092-b893-1ea056025be8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e740aa5b-0619-4092-b893-1ea056025be8\") " pod="openstack/ceilometer-0" Nov 22 08:17:31 crc kubenswrapper[4789]: I1122 08:17:31.766651 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rc7rz\" (UniqueName: \"kubernetes.io/projected/e740aa5b-0619-4092-b893-1ea056025be8-kube-api-access-rc7rz\") pod \"ceilometer-0\" (UID: \"e740aa5b-0619-4092-b893-1ea056025be8\") " pod="openstack/ceilometer-0" Nov 22 08:17:31 crc kubenswrapper[4789]: I1122 08:17:31.904472 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 08:17:31 crc kubenswrapper[4789]: I1122 08:17:31.981617 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2949ebe1-7541-42a2-894c-2588c2bd3396" path="/var/lib/kubelet/pods/2949ebe1-7541-42a2-894c-2588c2bd3396/volumes" Nov 22 08:17:31 crc kubenswrapper[4789]: I1122 08:17:31.982241 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="de991bf6-8f96-4c99-bf7a-b8909c0c80d2" path="/var/lib/kubelet/pods/de991bf6-8f96-4c99-bf7a-b8909c0c80d2/volumes" Nov 22 08:17:32 crc kubenswrapper[4789]: I1122 08:17:32.058668 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 22 08:17:32 crc kubenswrapper[4789]: I1122 08:17:32.154768 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6t4wv\" (UniqueName: \"kubernetes.io/projected/20a309c0-4fac-480e-b19d-e2ca480a6b6c-kube-api-access-6t4wv\") pod \"20a309c0-4fac-480e-b19d-e2ca480a6b6c\" (UID: \"20a309c0-4fac-480e-b19d-e2ca480a6b6c\") " Nov 22 08:17:32 crc kubenswrapper[4789]: I1122 08:17:32.166469 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20a309c0-4fac-480e-b19d-e2ca480a6b6c-kube-api-access-6t4wv" (OuterVolumeSpecName: "kube-api-access-6t4wv") pod "20a309c0-4fac-480e-b19d-e2ca480a6b6c" (UID: "20a309c0-4fac-480e-b19d-e2ca480a6b6c"). InnerVolumeSpecName "kube-api-access-6t4wv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:17:32 crc kubenswrapper[4789]: I1122 08:17:32.259866 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6t4wv\" (UniqueName: \"kubernetes.io/projected/20a309c0-4fac-480e-b19d-e2ca480a6b6c-kube-api-access-6t4wv\") on node \"crc\" DevicePath \"\"" Nov 22 08:17:32 crc kubenswrapper[4789]: I1122 08:17:32.430394 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:17:32 crc kubenswrapper[4789]: W1122 08:17:32.441203 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode740aa5b_0619_4092_b893_1ea056025be8.slice/crio-c59ade7ac1759ffdb35338f898e6ff95cc3b450a99ac1304f280952ee5d44272 WatchSource:0}: Error finding container c59ade7ac1759ffdb35338f898e6ff95cc3b450a99ac1304f280952ee5d44272: Status 404 returned error can't find the container with id c59ade7ac1759ffdb35338f898e6ff95cc3b450a99ac1304f280952ee5d44272 Nov 22 08:17:32 crc kubenswrapper[4789]: I1122 08:17:32.500207 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e740aa5b-0619-4092-b893-1ea056025be8","Type":"ContainerStarted","Data":"c59ade7ac1759ffdb35338f898e6ff95cc3b450a99ac1304f280952ee5d44272"} Nov 22 08:17:32 crc kubenswrapper[4789]: I1122 08:17:32.502420 4789 generic.go:334] "Generic (PLEG): container finished" podID="20a309c0-4fac-480e-b19d-e2ca480a6b6c" containerID="39f28765bd714106b4fd5010156d2c44ca1777ed061f27f9555f21c3d40d7898" exitCode=2 Nov 22 08:17:32 crc kubenswrapper[4789]: I1122 08:17:32.502467 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"20a309c0-4fac-480e-b19d-e2ca480a6b6c","Type":"ContainerDied","Data":"39f28765bd714106b4fd5010156d2c44ca1777ed061f27f9555f21c3d40d7898"} Nov 22 08:17:32 crc kubenswrapper[4789]: I1122 08:17:32.502492 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"20a309c0-4fac-480e-b19d-e2ca480a6b6c","Type":"ContainerDied","Data":"35ebde5382595cf8e01a5138ab3b0185954a6bcd09892ad0565fa7abbcf20083"} Nov 22 08:17:32 crc kubenswrapper[4789]: I1122 08:17:32.502490 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 22 08:17:32 crc kubenswrapper[4789]: I1122 08:17:32.502580 4789 scope.go:117] "RemoveContainer" containerID="39f28765bd714106b4fd5010156d2c44ca1777ed061f27f9555f21c3d40d7898" Nov 22 08:17:32 crc kubenswrapper[4789]: I1122 08:17:32.539314 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 22 08:17:32 crc kubenswrapper[4789]: I1122 08:17:32.549796 4789 scope.go:117] "RemoveContainer" containerID="39f28765bd714106b4fd5010156d2c44ca1777ed061f27f9555f21c3d40d7898" Nov 22 08:17:32 crc kubenswrapper[4789]: E1122 08:17:32.551664 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"39f28765bd714106b4fd5010156d2c44ca1777ed061f27f9555f21c3d40d7898\": container with ID starting with 39f28765bd714106b4fd5010156d2c44ca1777ed061f27f9555f21c3d40d7898 not found: ID does not exist" containerID="39f28765bd714106b4fd5010156d2c44ca1777ed061f27f9555f21c3d40d7898" Nov 22 08:17:32 crc kubenswrapper[4789]: I1122 08:17:32.551729 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39f28765bd714106b4fd5010156d2c44ca1777ed061f27f9555f21c3d40d7898"} err="failed to get container status \"39f28765bd714106b4fd5010156d2c44ca1777ed061f27f9555f21c3d40d7898\": rpc error: code = NotFound desc = could not find container \"39f28765bd714106b4fd5010156d2c44ca1777ed061f27f9555f21c3d40d7898\": container with ID starting with 39f28765bd714106b4fd5010156d2c44ca1777ed061f27f9555f21c3d40d7898 not found: ID does not exist" Nov 22 08:17:32 crc kubenswrapper[4789]: I1122 08:17:32.559861 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 22 08:17:32 crc kubenswrapper[4789]: I1122 08:17:32.570526 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Nov 22 08:17:32 crc kubenswrapper[4789]: E1122 08:17:32.571032 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20a309c0-4fac-480e-b19d-e2ca480a6b6c" containerName="kube-state-metrics" Nov 22 08:17:32 crc kubenswrapper[4789]: I1122 08:17:32.571054 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="20a309c0-4fac-480e-b19d-e2ca480a6b6c" containerName="kube-state-metrics" Nov 22 08:17:32 crc kubenswrapper[4789]: I1122 08:17:32.571248 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="20a309c0-4fac-480e-b19d-e2ca480a6b6c" containerName="kube-state-metrics" Nov 22 08:17:32 crc kubenswrapper[4789]: I1122 08:17:32.572643 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 22 08:17:32 crc kubenswrapper[4789]: I1122 08:17:32.579883 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Nov 22 08:17:32 crc kubenswrapper[4789]: I1122 08:17:32.580358 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Nov 22 08:17:32 crc kubenswrapper[4789]: I1122 08:17:32.587415 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 22 08:17:32 crc kubenswrapper[4789]: I1122 08:17:32.666551 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/a9945dfe-b305-48fc-9a1e-f5559b19a8b3-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"a9945dfe-b305-48fc-9a1e-f5559b19a8b3\") " pod="openstack/kube-state-metrics-0" Nov 22 08:17:32 crc kubenswrapper[4789]: I1122 08:17:32.666653 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z2t82\" (UniqueName: \"kubernetes.io/projected/a9945dfe-b305-48fc-9a1e-f5559b19a8b3-kube-api-access-z2t82\") pod \"kube-state-metrics-0\" (UID: \"a9945dfe-b305-48fc-9a1e-f5559b19a8b3\") " pod="openstack/kube-state-metrics-0" Nov 22 08:17:32 crc kubenswrapper[4789]: I1122 08:17:32.666678 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9945dfe-b305-48fc-9a1e-f5559b19a8b3-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"a9945dfe-b305-48fc-9a1e-f5559b19a8b3\") " pod="openstack/kube-state-metrics-0" Nov 22 08:17:32 crc kubenswrapper[4789]: I1122 08:17:32.666736 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/a9945dfe-b305-48fc-9a1e-f5559b19a8b3-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"a9945dfe-b305-48fc-9a1e-f5559b19a8b3\") " pod="openstack/kube-state-metrics-0" Nov 22 08:17:32 crc kubenswrapper[4789]: I1122 08:17:32.744678 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:17:32 crc kubenswrapper[4789]: I1122 08:17:32.768529 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/a9945dfe-b305-48fc-9a1e-f5559b19a8b3-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"a9945dfe-b305-48fc-9a1e-f5559b19a8b3\") " pod="openstack/kube-state-metrics-0" Nov 22 08:17:32 crc kubenswrapper[4789]: I1122 08:17:32.768695 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/a9945dfe-b305-48fc-9a1e-f5559b19a8b3-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"a9945dfe-b305-48fc-9a1e-f5559b19a8b3\") " pod="openstack/kube-state-metrics-0" Nov 22 08:17:32 crc kubenswrapper[4789]: I1122 08:17:32.768786 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z2t82\" (UniqueName: \"kubernetes.io/projected/a9945dfe-b305-48fc-9a1e-f5559b19a8b3-kube-api-access-z2t82\") pod \"kube-state-metrics-0\" (UID: \"a9945dfe-b305-48fc-9a1e-f5559b19a8b3\") " pod="openstack/kube-state-metrics-0" Nov 22 08:17:32 crc kubenswrapper[4789]: I1122 08:17:32.768808 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9945dfe-b305-48fc-9a1e-f5559b19a8b3-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"a9945dfe-b305-48fc-9a1e-f5559b19a8b3\") " pod="openstack/kube-state-metrics-0" Nov 22 08:17:32 crc kubenswrapper[4789]: I1122 08:17:32.773582 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9945dfe-b305-48fc-9a1e-f5559b19a8b3-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"a9945dfe-b305-48fc-9a1e-f5559b19a8b3\") " pod="openstack/kube-state-metrics-0" Nov 22 08:17:32 crc kubenswrapper[4789]: I1122 08:17:32.774299 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/a9945dfe-b305-48fc-9a1e-f5559b19a8b3-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"a9945dfe-b305-48fc-9a1e-f5559b19a8b3\") " pod="openstack/kube-state-metrics-0" Nov 22 08:17:32 crc kubenswrapper[4789]: I1122 08:17:32.774400 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/a9945dfe-b305-48fc-9a1e-f5559b19a8b3-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"a9945dfe-b305-48fc-9a1e-f5559b19a8b3\") " pod="openstack/kube-state-metrics-0" Nov 22 08:17:32 crc kubenswrapper[4789]: I1122 08:17:32.790490 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z2t82\" (UniqueName: \"kubernetes.io/projected/a9945dfe-b305-48fc-9a1e-f5559b19a8b3-kube-api-access-z2t82\") pod \"kube-state-metrics-0\" (UID: \"a9945dfe-b305-48fc-9a1e-f5559b19a8b3\") " pod="openstack/kube-state-metrics-0" Nov 22 08:17:32 crc kubenswrapper[4789]: I1122 08:17:32.897168 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 22 08:17:33 crc kubenswrapper[4789]: I1122 08:17:33.364149 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 22 08:17:33 crc kubenswrapper[4789]: I1122 08:17:33.517265 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e740aa5b-0619-4092-b893-1ea056025be8","Type":"ContainerStarted","Data":"8b87e836beb293dd213296102b968d3365d3eb9e8aabdf9e251c07c34633e8c3"} Nov 22 08:17:33 crc kubenswrapper[4789]: I1122 08:17:33.519190 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"a9945dfe-b305-48fc-9a1e-f5559b19a8b3","Type":"ContainerStarted","Data":"e8bc358043119ccf69511e616c18dc581e594dfd65b964ae0109ce0a52e0560b"} Nov 22 08:17:33 crc kubenswrapper[4789]: I1122 08:17:33.979386 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20a309c0-4fac-480e-b19d-e2ca480a6b6c" path="/var/lib/kubelet/pods/20a309c0-4fac-480e-b19d-e2ca480a6b6c/volumes" Nov 22 08:17:34 crc kubenswrapper[4789]: I1122 08:17:34.534466 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"a9945dfe-b305-48fc-9a1e-f5559b19a8b3","Type":"ContainerStarted","Data":"93a86f752d6c120837f10671a39043ec5d0dc8267210859b7277ca8dfbd22610"} Nov 22 08:17:34 crc kubenswrapper[4789]: I1122 08:17:34.534564 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Nov 22 08:17:34 crc kubenswrapper[4789]: I1122 08:17:34.537844 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e740aa5b-0619-4092-b893-1ea056025be8","Type":"ContainerStarted","Data":"8f09a2b27810ec2d6f5da622a3a69e262eca1c5ad63d813fe986625e1517ebed"} Nov 22 08:17:34 crc kubenswrapper[4789]: I1122 08:17:34.538177 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e740aa5b-0619-4092-b893-1ea056025be8","Type":"ContainerStarted","Data":"93e753c8079d6becb5107b39136c7eb0d3588a057c96760bb3c0bc5535e6c7ac"} Nov 22 08:17:34 crc kubenswrapper[4789]: I1122 08:17:34.559820 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.192782049 podStartE2EDuration="2.55979522s" podCreationTimestamp="2025-11-22 08:17:32 +0000 UTC" firstStartedPulling="2025-11-22 08:17:33.377616669 +0000 UTC m=+1267.612016942" lastFinishedPulling="2025-11-22 08:17:33.74462984 +0000 UTC m=+1267.979030113" observedRunningTime="2025-11-22 08:17:34.551951396 +0000 UTC m=+1268.786351669" watchObservedRunningTime="2025-11-22 08:17:34.55979522 +0000 UTC m=+1268.794195493" Nov 22 08:17:35 crc kubenswrapper[4789]: I1122 08:17:35.137984 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-mg9lj"] Nov 22 08:17:35 crc kubenswrapper[4789]: I1122 08:17:35.139042 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-mg9lj" Nov 22 08:17:35 crc kubenswrapper[4789]: I1122 08:17:35.141268 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Nov 22 08:17:35 crc kubenswrapper[4789]: I1122 08:17:35.141637 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Nov 22 08:17:35 crc kubenswrapper[4789]: I1122 08:17:35.141805 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-mckw6" Nov 22 08:17:35 crc kubenswrapper[4789]: I1122 08:17:35.148974 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-mg9lj"] Nov 22 08:17:35 crc kubenswrapper[4789]: I1122 08:17:35.208800 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ea41118-064f-4760-a1cc-0ebe3fc81a02-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-mg9lj\" (UID: \"8ea41118-064f-4760-a1cc-0ebe3fc81a02\") " pod="openstack/nova-cell0-conductor-db-sync-mg9lj" Nov 22 08:17:35 crc kubenswrapper[4789]: I1122 08:17:35.209043 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ea41118-064f-4760-a1cc-0ebe3fc81a02-config-data\") pod \"nova-cell0-conductor-db-sync-mg9lj\" (UID: \"8ea41118-064f-4760-a1cc-0ebe3fc81a02\") " pod="openstack/nova-cell0-conductor-db-sync-mg9lj" Nov 22 08:17:35 crc kubenswrapper[4789]: I1122 08:17:35.209123 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ea41118-064f-4760-a1cc-0ebe3fc81a02-scripts\") pod \"nova-cell0-conductor-db-sync-mg9lj\" (UID: \"8ea41118-064f-4760-a1cc-0ebe3fc81a02\") " pod="openstack/nova-cell0-conductor-db-sync-mg9lj" Nov 22 08:17:35 crc kubenswrapper[4789]: I1122 08:17:35.209461 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vlcqh\" (UniqueName: \"kubernetes.io/projected/8ea41118-064f-4760-a1cc-0ebe3fc81a02-kube-api-access-vlcqh\") pod \"nova-cell0-conductor-db-sync-mg9lj\" (UID: \"8ea41118-064f-4760-a1cc-0ebe3fc81a02\") " pod="openstack/nova-cell0-conductor-db-sync-mg9lj" Nov 22 08:17:35 crc kubenswrapper[4789]: I1122 08:17:35.311154 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ea41118-064f-4760-a1cc-0ebe3fc81a02-config-data\") pod \"nova-cell0-conductor-db-sync-mg9lj\" (UID: \"8ea41118-064f-4760-a1cc-0ebe3fc81a02\") " pod="openstack/nova-cell0-conductor-db-sync-mg9lj" Nov 22 08:17:35 crc kubenswrapper[4789]: I1122 08:17:35.311229 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ea41118-064f-4760-a1cc-0ebe3fc81a02-scripts\") pod \"nova-cell0-conductor-db-sync-mg9lj\" (UID: \"8ea41118-064f-4760-a1cc-0ebe3fc81a02\") " pod="openstack/nova-cell0-conductor-db-sync-mg9lj" Nov 22 08:17:35 crc kubenswrapper[4789]: I1122 08:17:35.311275 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vlcqh\" (UniqueName: \"kubernetes.io/projected/8ea41118-064f-4760-a1cc-0ebe3fc81a02-kube-api-access-vlcqh\") pod \"nova-cell0-conductor-db-sync-mg9lj\" (UID: \"8ea41118-064f-4760-a1cc-0ebe3fc81a02\") " pod="openstack/nova-cell0-conductor-db-sync-mg9lj" Nov 22 08:17:35 crc kubenswrapper[4789]: I1122 08:17:35.311311 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ea41118-064f-4760-a1cc-0ebe3fc81a02-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-mg9lj\" (UID: \"8ea41118-064f-4760-a1cc-0ebe3fc81a02\") " pod="openstack/nova-cell0-conductor-db-sync-mg9lj" Nov 22 08:17:35 crc kubenswrapper[4789]: I1122 08:17:35.316927 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ea41118-064f-4760-a1cc-0ebe3fc81a02-config-data\") pod \"nova-cell0-conductor-db-sync-mg9lj\" (UID: \"8ea41118-064f-4760-a1cc-0ebe3fc81a02\") " pod="openstack/nova-cell0-conductor-db-sync-mg9lj" Nov 22 08:17:35 crc kubenswrapper[4789]: I1122 08:17:35.317980 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ea41118-064f-4760-a1cc-0ebe3fc81a02-scripts\") pod \"nova-cell0-conductor-db-sync-mg9lj\" (UID: \"8ea41118-064f-4760-a1cc-0ebe3fc81a02\") " pod="openstack/nova-cell0-conductor-db-sync-mg9lj" Nov 22 08:17:35 crc kubenswrapper[4789]: I1122 08:17:35.318459 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ea41118-064f-4760-a1cc-0ebe3fc81a02-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-mg9lj\" (UID: \"8ea41118-064f-4760-a1cc-0ebe3fc81a02\") " pod="openstack/nova-cell0-conductor-db-sync-mg9lj" Nov 22 08:17:35 crc kubenswrapper[4789]: I1122 08:17:35.337244 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vlcqh\" (UniqueName: \"kubernetes.io/projected/8ea41118-064f-4760-a1cc-0ebe3fc81a02-kube-api-access-vlcqh\") pod \"nova-cell0-conductor-db-sync-mg9lj\" (UID: \"8ea41118-064f-4760-a1cc-0ebe3fc81a02\") " pod="openstack/nova-cell0-conductor-db-sync-mg9lj" Nov 22 08:17:35 crc kubenswrapper[4789]: I1122 08:17:35.468839 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-mg9lj" Nov 22 08:17:35 crc kubenswrapper[4789]: I1122 08:17:35.903236 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-mg9lj"] Nov 22 08:17:35 crc kubenswrapper[4789]: W1122 08:17:35.905179 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8ea41118_064f_4760_a1cc_0ebe3fc81a02.slice/crio-2c15d8971b3f53e1ba28fd992ea1a87344d646ffefac0c97ea5be78c52f84efe WatchSource:0}: Error finding container 2c15d8971b3f53e1ba28fd992ea1a87344d646ffefac0c97ea5be78c52f84efe: Status 404 returned error can't find the container with id 2c15d8971b3f53e1ba28fd992ea1a87344d646ffefac0c97ea5be78c52f84efe Nov 22 08:17:36 crc kubenswrapper[4789]: I1122 08:17:36.558335 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-mg9lj" event={"ID":"8ea41118-064f-4760-a1cc-0ebe3fc81a02","Type":"ContainerStarted","Data":"2c15d8971b3f53e1ba28fd992ea1a87344d646ffefac0c97ea5be78c52f84efe"} Nov 22 08:17:39 crc kubenswrapper[4789]: I1122 08:17:39.589347 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e740aa5b-0619-4092-b893-1ea056025be8","Type":"ContainerStarted","Data":"9bbc4011d0613b64eb20a140cd0e48276cda3a646ae9e1856274c2800bebca0c"} Nov 22 08:17:39 crc kubenswrapper[4789]: I1122 08:17:39.590152 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 22 08:17:39 crc kubenswrapper[4789]: I1122 08:17:39.590275 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e740aa5b-0619-4092-b893-1ea056025be8" containerName="proxy-httpd" containerID="cri-o://9bbc4011d0613b64eb20a140cd0e48276cda3a646ae9e1856274c2800bebca0c" gracePeriod=30 Nov 22 08:17:39 crc kubenswrapper[4789]: I1122 08:17:39.590237 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e740aa5b-0619-4092-b893-1ea056025be8" containerName="ceilometer-central-agent" containerID="cri-o://8b87e836beb293dd213296102b968d3365d3eb9e8aabdf9e251c07c34633e8c3" gracePeriod=30 Nov 22 08:17:39 crc kubenswrapper[4789]: I1122 08:17:39.594794 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e740aa5b-0619-4092-b893-1ea056025be8" containerName="sg-core" containerID="cri-o://8f09a2b27810ec2d6f5da622a3a69e262eca1c5ad63d813fe986625e1517ebed" gracePeriod=30 Nov 22 08:17:39 crc kubenswrapper[4789]: I1122 08:17:39.590276 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e740aa5b-0619-4092-b893-1ea056025be8" containerName="ceilometer-notification-agent" containerID="cri-o://93e753c8079d6becb5107b39136c7eb0d3588a057c96760bb3c0bc5535e6c7ac" gracePeriod=30 Nov 22 08:17:39 crc kubenswrapper[4789]: I1122 08:17:39.624835 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.359432722 podStartE2EDuration="8.624809626s" podCreationTimestamp="2025-11-22 08:17:31 +0000 UTC" firstStartedPulling="2025-11-22 08:17:32.449885246 +0000 UTC m=+1266.684294660" lastFinishedPulling="2025-11-22 08:17:38.715271291 +0000 UTC m=+1272.949671564" observedRunningTime="2025-11-22 08:17:39.616310924 +0000 UTC m=+1273.850711227" watchObservedRunningTime="2025-11-22 08:17:39.624809626 +0000 UTC m=+1273.859209909" Nov 22 08:17:40 crc kubenswrapper[4789]: I1122 08:17:40.528056 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 08:17:40 crc kubenswrapper[4789]: I1122 08:17:40.601375 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e740aa5b-0619-4092-b893-1ea056025be8-combined-ca-bundle\") pod \"e740aa5b-0619-4092-b893-1ea056025be8\" (UID: \"e740aa5b-0619-4092-b893-1ea056025be8\") " Nov 22 08:17:40 crc kubenswrapper[4789]: I1122 08:17:40.601698 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e740aa5b-0619-4092-b893-1ea056025be8-run-httpd\") pod \"e740aa5b-0619-4092-b893-1ea056025be8\" (UID: \"e740aa5b-0619-4092-b893-1ea056025be8\") " Nov 22 08:17:40 crc kubenswrapper[4789]: I1122 08:17:40.601771 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e740aa5b-0619-4092-b893-1ea056025be8-config-data\") pod \"e740aa5b-0619-4092-b893-1ea056025be8\" (UID: \"e740aa5b-0619-4092-b893-1ea056025be8\") " Nov 22 08:17:40 crc kubenswrapper[4789]: I1122 08:17:40.601866 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e740aa5b-0619-4092-b893-1ea056025be8-scripts\") pod \"e740aa5b-0619-4092-b893-1ea056025be8\" (UID: \"e740aa5b-0619-4092-b893-1ea056025be8\") " Nov 22 08:17:40 crc kubenswrapper[4789]: I1122 08:17:40.601922 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e740aa5b-0619-4092-b893-1ea056025be8-sg-core-conf-yaml\") pod \"e740aa5b-0619-4092-b893-1ea056025be8\" (UID: \"e740aa5b-0619-4092-b893-1ea056025be8\") " Nov 22 08:17:40 crc kubenswrapper[4789]: I1122 08:17:40.602064 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rc7rz\" (UniqueName: \"kubernetes.io/projected/e740aa5b-0619-4092-b893-1ea056025be8-kube-api-access-rc7rz\") pod \"e740aa5b-0619-4092-b893-1ea056025be8\" (UID: \"e740aa5b-0619-4092-b893-1ea056025be8\") " Nov 22 08:17:40 crc kubenswrapper[4789]: I1122 08:17:40.602502 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e740aa5b-0619-4092-b893-1ea056025be8-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "e740aa5b-0619-4092-b893-1ea056025be8" (UID: "e740aa5b-0619-4092-b893-1ea056025be8"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:17:40 crc kubenswrapper[4789]: I1122 08:17:40.602541 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e740aa5b-0619-4092-b893-1ea056025be8-log-httpd\") pod \"e740aa5b-0619-4092-b893-1ea056025be8\" (UID: \"e740aa5b-0619-4092-b893-1ea056025be8\") " Nov 22 08:17:40 crc kubenswrapper[4789]: I1122 08:17:40.602903 4789 generic.go:334] "Generic (PLEG): container finished" podID="e740aa5b-0619-4092-b893-1ea056025be8" containerID="9bbc4011d0613b64eb20a140cd0e48276cda3a646ae9e1856274c2800bebca0c" exitCode=0 Nov 22 08:17:40 crc kubenswrapper[4789]: I1122 08:17:40.602944 4789 generic.go:334] "Generic (PLEG): container finished" podID="e740aa5b-0619-4092-b893-1ea056025be8" containerID="8f09a2b27810ec2d6f5da622a3a69e262eca1c5ad63d813fe986625e1517ebed" exitCode=2 Nov 22 08:17:40 crc kubenswrapper[4789]: I1122 08:17:40.602959 4789 generic.go:334] "Generic (PLEG): container finished" podID="e740aa5b-0619-4092-b893-1ea056025be8" containerID="93e753c8079d6becb5107b39136c7eb0d3588a057c96760bb3c0bc5535e6c7ac" exitCode=0 Nov 22 08:17:40 crc kubenswrapper[4789]: I1122 08:17:40.602971 4789 generic.go:334] "Generic (PLEG): container finished" podID="e740aa5b-0619-4092-b893-1ea056025be8" containerID="8b87e836beb293dd213296102b968d3365d3eb9e8aabdf9e251c07c34633e8c3" exitCode=0 Nov 22 08:17:40 crc kubenswrapper[4789]: I1122 08:17:40.602985 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e740aa5b-0619-4092-b893-1ea056025be8-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "e740aa5b-0619-4092-b893-1ea056025be8" (UID: "e740aa5b-0619-4092-b893-1ea056025be8"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:17:40 crc kubenswrapper[4789]: I1122 08:17:40.602998 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e740aa5b-0619-4092-b893-1ea056025be8","Type":"ContainerDied","Data":"9bbc4011d0613b64eb20a140cd0e48276cda3a646ae9e1856274c2800bebca0c"} Nov 22 08:17:40 crc kubenswrapper[4789]: I1122 08:17:40.603010 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 08:17:40 crc kubenswrapper[4789]: I1122 08:17:40.603041 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e740aa5b-0619-4092-b893-1ea056025be8","Type":"ContainerDied","Data":"8f09a2b27810ec2d6f5da622a3a69e262eca1c5ad63d813fe986625e1517ebed"} Nov 22 08:17:40 crc kubenswrapper[4789]: I1122 08:17:40.603060 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e740aa5b-0619-4092-b893-1ea056025be8","Type":"ContainerDied","Data":"93e753c8079d6becb5107b39136c7eb0d3588a057c96760bb3c0bc5535e6c7ac"} Nov 22 08:17:40 crc kubenswrapper[4789]: I1122 08:17:40.603076 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e740aa5b-0619-4092-b893-1ea056025be8","Type":"ContainerDied","Data":"8b87e836beb293dd213296102b968d3365d3eb9e8aabdf9e251c07c34633e8c3"} Nov 22 08:17:40 crc kubenswrapper[4789]: I1122 08:17:40.603091 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e740aa5b-0619-4092-b893-1ea056025be8","Type":"ContainerDied","Data":"c59ade7ac1759ffdb35338f898e6ff95cc3b450a99ac1304f280952ee5d44272"} Nov 22 08:17:40 crc kubenswrapper[4789]: I1122 08:17:40.603115 4789 scope.go:117] "RemoveContainer" containerID="9bbc4011d0613b64eb20a140cd0e48276cda3a646ae9e1856274c2800bebca0c" Nov 22 08:17:40 crc kubenswrapper[4789]: I1122 08:17:40.603657 4789 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e740aa5b-0619-4092-b893-1ea056025be8-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 22 08:17:40 crc kubenswrapper[4789]: I1122 08:17:40.603685 4789 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e740aa5b-0619-4092-b893-1ea056025be8-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 22 08:17:40 crc kubenswrapper[4789]: I1122 08:17:40.608301 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e740aa5b-0619-4092-b893-1ea056025be8-kube-api-access-rc7rz" (OuterVolumeSpecName: "kube-api-access-rc7rz") pod "e740aa5b-0619-4092-b893-1ea056025be8" (UID: "e740aa5b-0619-4092-b893-1ea056025be8"). InnerVolumeSpecName "kube-api-access-rc7rz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:17:40 crc kubenswrapper[4789]: I1122 08:17:40.610291 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e740aa5b-0619-4092-b893-1ea056025be8-scripts" (OuterVolumeSpecName: "scripts") pod "e740aa5b-0619-4092-b893-1ea056025be8" (UID: "e740aa5b-0619-4092-b893-1ea056025be8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:17:40 crc kubenswrapper[4789]: I1122 08:17:40.632493 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e740aa5b-0619-4092-b893-1ea056025be8-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "e740aa5b-0619-4092-b893-1ea056025be8" (UID: "e740aa5b-0619-4092-b893-1ea056025be8"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:17:40 crc kubenswrapper[4789]: I1122 08:17:40.639604 4789 scope.go:117] "RemoveContainer" containerID="8f09a2b27810ec2d6f5da622a3a69e262eca1c5ad63d813fe986625e1517ebed" Nov 22 08:17:40 crc kubenswrapper[4789]: I1122 08:17:40.679295 4789 scope.go:117] "RemoveContainer" containerID="93e753c8079d6becb5107b39136c7eb0d3588a057c96760bb3c0bc5535e6c7ac" Nov 22 08:17:40 crc kubenswrapper[4789]: I1122 08:17:40.679319 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e740aa5b-0619-4092-b893-1ea056025be8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e740aa5b-0619-4092-b893-1ea056025be8" (UID: "e740aa5b-0619-4092-b893-1ea056025be8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:17:40 crc kubenswrapper[4789]: I1122 08:17:40.703228 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e740aa5b-0619-4092-b893-1ea056025be8-config-data" (OuterVolumeSpecName: "config-data") pod "e740aa5b-0619-4092-b893-1ea056025be8" (UID: "e740aa5b-0619-4092-b893-1ea056025be8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:17:40 crc kubenswrapper[4789]: I1122 08:17:40.703491 4789 scope.go:117] "RemoveContainer" containerID="8b87e836beb293dd213296102b968d3365d3eb9e8aabdf9e251c07c34633e8c3" Nov 22 08:17:40 crc kubenswrapper[4789]: I1122 08:17:40.705642 4789 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e740aa5b-0619-4092-b893-1ea056025be8-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:17:40 crc kubenswrapper[4789]: I1122 08:17:40.705673 4789 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e740aa5b-0619-4092-b893-1ea056025be8-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:17:40 crc kubenswrapper[4789]: I1122 08:17:40.705686 4789 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e740aa5b-0619-4092-b893-1ea056025be8-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 22 08:17:40 crc kubenswrapper[4789]: I1122 08:17:40.705699 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rc7rz\" (UniqueName: \"kubernetes.io/projected/e740aa5b-0619-4092-b893-1ea056025be8-kube-api-access-rc7rz\") on node \"crc\" DevicePath \"\"" Nov 22 08:17:40 crc kubenswrapper[4789]: I1122 08:17:40.705711 4789 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e740aa5b-0619-4092-b893-1ea056025be8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:17:40 crc kubenswrapper[4789]: I1122 08:17:40.735363 4789 scope.go:117] "RemoveContainer" containerID="9bbc4011d0613b64eb20a140cd0e48276cda3a646ae9e1856274c2800bebca0c" Nov 22 08:17:40 crc kubenswrapper[4789]: E1122 08:17:40.736181 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9bbc4011d0613b64eb20a140cd0e48276cda3a646ae9e1856274c2800bebca0c\": container with ID starting with 9bbc4011d0613b64eb20a140cd0e48276cda3a646ae9e1856274c2800bebca0c not found: ID does not exist" containerID="9bbc4011d0613b64eb20a140cd0e48276cda3a646ae9e1856274c2800bebca0c" Nov 22 08:17:40 crc kubenswrapper[4789]: I1122 08:17:40.736257 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9bbc4011d0613b64eb20a140cd0e48276cda3a646ae9e1856274c2800bebca0c"} err="failed to get container status \"9bbc4011d0613b64eb20a140cd0e48276cda3a646ae9e1856274c2800bebca0c\": rpc error: code = NotFound desc = could not find container \"9bbc4011d0613b64eb20a140cd0e48276cda3a646ae9e1856274c2800bebca0c\": container with ID starting with 9bbc4011d0613b64eb20a140cd0e48276cda3a646ae9e1856274c2800bebca0c not found: ID does not exist" Nov 22 08:17:40 crc kubenswrapper[4789]: I1122 08:17:40.736307 4789 scope.go:117] "RemoveContainer" containerID="8f09a2b27810ec2d6f5da622a3a69e262eca1c5ad63d813fe986625e1517ebed" Nov 22 08:17:40 crc kubenswrapper[4789]: E1122 08:17:40.736847 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8f09a2b27810ec2d6f5da622a3a69e262eca1c5ad63d813fe986625e1517ebed\": container with ID starting with 8f09a2b27810ec2d6f5da622a3a69e262eca1c5ad63d813fe986625e1517ebed not found: ID does not exist" containerID="8f09a2b27810ec2d6f5da622a3a69e262eca1c5ad63d813fe986625e1517ebed" Nov 22 08:17:40 crc kubenswrapper[4789]: I1122 08:17:40.736893 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f09a2b27810ec2d6f5da622a3a69e262eca1c5ad63d813fe986625e1517ebed"} err="failed to get container status \"8f09a2b27810ec2d6f5da622a3a69e262eca1c5ad63d813fe986625e1517ebed\": rpc error: code = NotFound desc = could not find container \"8f09a2b27810ec2d6f5da622a3a69e262eca1c5ad63d813fe986625e1517ebed\": container with ID starting with 8f09a2b27810ec2d6f5da622a3a69e262eca1c5ad63d813fe986625e1517ebed not found: ID does not exist" Nov 22 08:17:40 crc kubenswrapper[4789]: I1122 08:17:40.736926 4789 scope.go:117] "RemoveContainer" containerID="93e753c8079d6becb5107b39136c7eb0d3588a057c96760bb3c0bc5535e6c7ac" Nov 22 08:17:40 crc kubenswrapper[4789]: E1122 08:17:40.737385 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"93e753c8079d6becb5107b39136c7eb0d3588a057c96760bb3c0bc5535e6c7ac\": container with ID starting with 93e753c8079d6becb5107b39136c7eb0d3588a057c96760bb3c0bc5535e6c7ac not found: ID does not exist" containerID="93e753c8079d6becb5107b39136c7eb0d3588a057c96760bb3c0bc5535e6c7ac" Nov 22 08:17:40 crc kubenswrapper[4789]: I1122 08:17:40.737410 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93e753c8079d6becb5107b39136c7eb0d3588a057c96760bb3c0bc5535e6c7ac"} err="failed to get container status \"93e753c8079d6becb5107b39136c7eb0d3588a057c96760bb3c0bc5535e6c7ac\": rpc error: code = NotFound desc = could not find container \"93e753c8079d6becb5107b39136c7eb0d3588a057c96760bb3c0bc5535e6c7ac\": container with ID starting with 93e753c8079d6becb5107b39136c7eb0d3588a057c96760bb3c0bc5535e6c7ac not found: ID does not exist" Nov 22 08:17:40 crc kubenswrapper[4789]: I1122 08:17:40.737426 4789 scope.go:117] "RemoveContainer" containerID="8b87e836beb293dd213296102b968d3365d3eb9e8aabdf9e251c07c34633e8c3" Nov 22 08:17:40 crc kubenswrapper[4789]: E1122 08:17:40.737649 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8b87e836beb293dd213296102b968d3365d3eb9e8aabdf9e251c07c34633e8c3\": container with ID starting with 8b87e836beb293dd213296102b968d3365d3eb9e8aabdf9e251c07c34633e8c3 not found: ID does not exist" containerID="8b87e836beb293dd213296102b968d3365d3eb9e8aabdf9e251c07c34633e8c3" Nov 22 08:17:40 crc kubenswrapper[4789]: I1122 08:17:40.737673 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b87e836beb293dd213296102b968d3365d3eb9e8aabdf9e251c07c34633e8c3"} err="failed to get container status \"8b87e836beb293dd213296102b968d3365d3eb9e8aabdf9e251c07c34633e8c3\": rpc error: code = NotFound desc = could not find container \"8b87e836beb293dd213296102b968d3365d3eb9e8aabdf9e251c07c34633e8c3\": container with ID starting with 8b87e836beb293dd213296102b968d3365d3eb9e8aabdf9e251c07c34633e8c3 not found: ID does not exist" Nov 22 08:17:40 crc kubenswrapper[4789]: I1122 08:17:40.737687 4789 scope.go:117] "RemoveContainer" containerID="9bbc4011d0613b64eb20a140cd0e48276cda3a646ae9e1856274c2800bebca0c" Nov 22 08:17:40 crc kubenswrapper[4789]: I1122 08:17:40.738065 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9bbc4011d0613b64eb20a140cd0e48276cda3a646ae9e1856274c2800bebca0c"} err="failed to get container status \"9bbc4011d0613b64eb20a140cd0e48276cda3a646ae9e1856274c2800bebca0c\": rpc error: code = NotFound desc = could not find container \"9bbc4011d0613b64eb20a140cd0e48276cda3a646ae9e1856274c2800bebca0c\": container with ID starting with 9bbc4011d0613b64eb20a140cd0e48276cda3a646ae9e1856274c2800bebca0c not found: ID does not exist" Nov 22 08:17:40 crc kubenswrapper[4789]: I1122 08:17:40.738086 4789 scope.go:117] "RemoveContainer" containerID="8f09a2b27810ec2d6f5da622a3a69e262eca1c5ad63d813fe986625e1517ebed" Nov 22 08:17:40 crc kubenswrapper[4789]: I1122 08:17:40.738475 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f09a2b27810ec2d6f5da622a3a69e262eca1c5ad63d813fe986625e1517ebed"} err="failed to get container status \"8f09a2b27810ec2d6f5da622a3a69e262eca1c5ad63d813fe986625e1517ebed\": rpc error: code = NotFound desc = could not find container \"8f09a2b27810ec2d6f5da622a3a69e262eca1c5ad63d813fe986625e1517ebed\": container with ID starting with 8f09a2b27810ec2d6f5da622a3a69e262eca1c5ad63d813fe986625e1517ebed not found: ID does not exist" Nov 22 08:17:40 crc kubenswrapper[4789]: I1122 08:17:40.738498 4789 scope.go:117] "RemoveContainer" containerID="93e753c8079d6becb5107b39136c7eb0d3588a057c96760bb3c0bc5535e6c7ac" Nov 22 08:17:40 crc kubenswrapper[4789]: I1122 08:17:40.738770 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93e753c8079d6becb5107b39136c7eb0d3588a057c96760bb3c0bc5535e6c7ac"} err="failed to get container status \"93e753c8079d6becb5107b39136c7eb0d3588a057c96760bb3c0bc5535e6c7ac\": rpc error: code = NotFound desc = could not find container \"93e753c8079d6becb5107b39136c7eb0d3588a057c96760bb3c0bc5535e6c7ac\": container with ID starting with 93e753c8079d6becb5107b39136c7eb0d3588a057c96760bb3c0bc5535e6c7ac not found: ID does not exist" Nov 22 08:17:40 crc kubenswrapper[4789]: I1122 08:17:40.738796 4789 scope.go:117] "RemoveContainer" containerID="8b87e836beb293dd213296102b968d3365d3eb9e8aabdf9e251c07c34633e8c3" Nov 22 08:17:40 crc kubenswrapper[4789]: I1122 08:17:40.738977 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b87e836beb293dd213296102b968d3365d3eb9e8aabdf9e251c07c34633e8c3"} err="failed to get container status \"8b87e836beb293dd213296102b968d3365d3eb9e8aabdf9e251c07c34633e8c3\": rpc error: code = NotFound desc = could not find container \"8b87e836beb293dd213296102b968d3365d3eb9e8aabdf9e251c07c34633e8c3\": container with ID starting with 8b87e836beb293dd213296102b968d3365d3eb9e8aabdf9e251c07c34633e8c3 not found: ID does not exist" Nov 22 08:17:40 crc kubenswrapper[4789]: I1122 08:17:40.739000 4789 scope.go:117] "RemoveContainer" containerID="9bbc4011d0613b64eb20a140cd0e48276cda3a646ae9e1856274c2800bebca0c" Nov 22 08:17:40 crc kubenswrapper[4789]: I1122 08:17:40.739245 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9bbc4011d0613b64eb20a140cd0e48276cda3a646ae9e1856274c2800bebca0c"} err="failed to get container status \"9bbc4011d0613b64eb20a140cd0e48276cda3a646ae9e1856274c2800bebca0c\": rpc error: code = NotFound desc = could not find container \"9bbc4011d0613b64eb20a140cd0e48276cda3a646ae9e1856274c2800bebca0c\": container with ID starting with 9bbc4011d0613b64eb20a140cd0e48276cda3a646ae9e1856274c2800bebca0c not found: ID does not exist" Nov 22 08:17:40 crc kubenswrapper[4789]: I1122 08:17:40.739270 4789 scope.go:117] "RemoveContainer" containerID="8f09a2b27810ec2d6f5da622a3a69e262eca1c5ad63d813fe986625e1517ebed" Nov 22 08:17:40 crc kubenswrapper[4789]: I1122 08:17:40.739500 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f09a2b27810ec2d6f5da622a3a69e262eca1c5ad63d813fe986625e1517ebed"} err="failed to get container status \"8f09a2b27810ec2d6f5da622a3a69e262eca1c5ad63d813fe986625e1517ebed\": rpc error: code = NotFound desc = could not find container \"8f09a2b27810ec2d6f5da622a3a69e262eca1c5ad63d813fe986625e1517ebed\": container with ID starting with 8f09a2b27810ec2d6f5da622a3a69e262eca1c5ad63d813fe986625e1517ebed not found: ID does not exist" Nov 22 08:17:40 crc kubenswrapper[4789]: I1122 08:17:40.739523 4789 scope.go:117] "RemoveContainer" containerID="93e753c8079d6becb5107b39136c7eb0d3588a057c96760bb3c0bc5535e6c7ac" Nov 22 08:17:40 crc kubenswrapper[4789]: I1122 08:17:40.739890 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93e753c8079d6becb5107b39136c7eb0d3588a057c96760bb3c0bc5535e6c7ac"} err="failed to get container status \"93e753c8079d6becb5107b39136c7eb0d3588a057c96760bb3c0bc5535e6c7ac\": rpc error: code = NotFound desc = could not find container \"93e753c8079d6becb5107b39136c7eb0d3588a057c96760bb3c0bc5535e6c7ac\": container with ID starting with 93e753c8079d6becb5107b39136c7eb0d3588a057c96760bb3c0bc5535e6c7ac not found: ID does not exist" Nov 22 08:17:40 crc kubenswrapper[4789]: I1122 08:17:40.739924 4789 scope.go:117] "RemoveContainer" containerID="8b87e836beb293dd213296102b968d3365d3eb9e8aabdf9e251c07c34633e8c3" Nov 22 08:17:40 crc kubenswrapper[4789]: I1122 08:17:40.740167 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b87e836beb293dd213296102b968d3365d3eb9e8aabdf9e251c07c34633e8c3"} err="failed to get container status \"8b87e836beb293dd213296102b968d3365d3eb9e8aabdf9e251c07c34633e8c3\": rpc error: code = NotFound desc = could not find container \"8b87e836beb293dd213296102b968d3365d3eb9e8aabdf9e251c07c34633e8c3\": container with ID starting with 8b87e836beb293dd213296102b968d3365d3eb9e8aabdf9e251c07c34633e8c3 not found: ID does not exist" Nov 22 08:17:40 crc kubenswrapper[4789]: I1122 08:17:40.740192 4789 scope.go:117] "RemoveContainer" containerID="9bbc4011d0613b64eb20a140cd0e48276cda3a646ae9e1856274c2800bebca0c" Nov 22 08:17:40 crc kubenswrapper[4789]: I1122 08:17:40.740396 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9bbc4011d0613b64eb20a140cd0e48276cda3a646ae9e1856274c2800bebca0c"} err="failed to get container status \"9bbc4011d0613b64eb20a140cd0e48276cda3a646ae9e1856274c2800bebca0c\": rpc error: code = NotFound desc = could not find container \"9bbc4011d0613b64eb20a140cd0e48276cda3a646ae9e1856274c2800bebca0c\": container with ID starting with 9bbc4011d0613b64eb20a140cd0e48276cda3a646ae9e1856274c2800bebca0c not found: ID does not exist" Nov 22 08:17:40 crc kubenswrapper[4789]: I1122 08:17:40.740420 4789 scope.go:117] "RemoveContainer" containerID="8f09a2b27810ec2d6f5da622a3a69e262eca1c5ad63d813fe986625e1517ebed" Nov 22 08:17:40 crc kubenswrapper[4789]: I1122 08:17:40.740670 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f09a2b27810ec2d6f5da622a3a69e262eca1c5ad63d813fe986625e1517ebed"} err="failed to get container status \"8f09a2b27810ec2d6f5da622a3a69e262eca1c5ad63d813fe986625e1517ebed\": rpc error: code = NotFound desc = could not find container \"8f09a2b27810ec2d6f5da622a3a69e262eca1c5ad63d813fe986625e1517ebed\": container with ID starting with 8f09a2b27810ec2d6f5da622a3a69e262eca1c5ad63d813fe986625e1517ebed not found: ID does not exist" Nov 22 08:17:40 crc kubenswrapper[4789]: I1122 08:17:40.740697 4789 scope.go:117] "RemoveContainer" containerID="93e753c8079d6becb5107b39136c7eb0d3588a057c96760bb3c0bc5535e6c7ac" Nov 22 08:17:40 crc kubenswrapper[4789]: I1122 08:17:40.740938 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93e753c8079d6becb5107b39136c7eb0d3588a057c96760bb3c0bc5535e6c7ac"} err="failed to get container status \"93e753c8079d6becb5107b39136c7eb0d3588a057c96760bb3c0bc5535e6c7ac\": rpc error: code = NotFound desc = could not find container \"93e753c8079d6becb5107b39136c7eb0d3588a057c96760bb3c0bc5535e6c7ac\": container with ID starting with 93e753c8079d6becb5107b39136c7eb0d3588a057c96760bb3c0bc5535e6c7ac not found: ID does not exist" Nov 22 08:17:40 crc kubenswrapper[4789]: I1122 08:17:40.740967 4789 scope.go:117] "RemoveContainer" containerID="8b87e836beb293dd213296102b968d3365d3eb9e8aabdf9e251c07c34633e8c3" Nov 22 08:17:40 crc kubenswrapper[4789]: I1122 08:17:40.741208 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b87e836beb293dd213296102b968d3365d3eb9e8aabdf9e251c07c34633e8c3"} err="failed to get container status \"8b87e836beb293dd213296102b968d3365d3eb9e8aabdf9e251c07c34633e8c3\": rpc error: code = NotFound desc = could not find container \"8b87e836beb293dd213296102b968d3365d3eb9e8aabdf9e251c07c34633e8c3\": container with ID starting with 8b87e836beb293dd213296102b968d3365d3eb9e8aabdf9e251c07c34633e8c3 not found: ID does not exist" Nov 22 08:17:40 crc kubenswrapper[4789]: I1122 08:17:40.940672 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:17:40 crc kubenswrapper[4789]: I1122 08:17:40.948614 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:17:40 crc kubenswrapper[4789]: I1122 08:17:40.957112 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:17:40 crc kubenswrapper[4789]: E1122 08:17:40.957463 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e740aa5b-0619-4092-b893-1ea056025be8" containerName="ceilometer-central-agent" Nov 22 08:17:40 crc kubenswrapper[4789]: I1122 08:17:40.957475 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="e740aa5b-0619-4092-b893-1ea056025be8" containerName="ceilometer-central-agent" Nov 22 08:17:40 crc kubenswrapper[4789]: E1122 08:17:40.957497 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e740aa5b-0619-4092-b893-1ea056025be8" containerName="ceilometer-notification-agent" Nov 22 08:17:40 crc kubenswrapper[4789]: I1122 08:17:40.957503 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="e740aa5b-0619-4092-b893-1ea056025be8" containerName="ceilometer-notification-agent" Nov 22 08:17:40 crc kubenswrapper[4789]: E1122 08:17:40.957510 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e740aa5b-0619-4092-b893-1ea056025be8" containerName="proxy-httpd" Nov 22 08:17:40 crc kubenswrapper[4789]: I1122 08:17:40.957515 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="e740aa5b-0619-4092-b893-1ea056025be8" containerName="proxy-httpd" Nov 22 08:17:40 crc kubenswrapper[4789]: E1122 08:17:40.957531 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e740aa5b-0619-4092-b893-1ea056025be8" containerName="sg-core" Nov 22 08:17:40 crc kubenswrapper[4789]: I1122 08:17:40.957536 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="e740aa5b-0619-4092-b893-1ea056025be8" containerName="sg-core" Nov 22 08:17:40 crc kubenswrapper[4789]: I1122 08:17:40.957684 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="e740aa5b-0619-4092-b893-1ea056025be8" containerName="sg-core" Nov 22 08:17:40 crc kubenswrapper[4789]: I1122 08:17:40.957700 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="e740aa5b-0619-4092-b893-1ea056025be8" containerName="proxy-httpd" Nov 22 08:17:40 crc kubenswrapper[4789]: I1122 08:17:40.957715 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="e740aa5b-0619-4092-b893-1ea056025be8" containerName="ceilometer-central-agent" Nov 22 08:17:40 crc kubenswrapper[4789]: I1122 08:17:40.957726 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="e740aa5b-0619-4092-b893-1ea056025be8" containerName="ceilometer-notification-agent" Nov 22 08:17:40 crc kubenswrapper[4789]: I1122 08:17:40.959221 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 08:17:40 crc kubenswrapper[4789]: I1122 08:17:40.961185 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 22 08:17:40 crc kubenswrapper[4789]: I1122 08:17:40.961400 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 22 08:17:40 crc kubenswrapper[4789]: I1122 08:17:40.961482 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 22 08:17:40 crc kubenswrapper[4789]: I1122 08:17:40.975364 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:17:41 crc kubenswrapper[4789]: I1122 08:17:41.009659 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/856539fc-0f5e-4ca1-8549-ad867a8c9f9d-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"856539fc-0f5e-4ca1-8549-ad867a8c9f9d\") " pod="openstack/ceilometer-0" Nov 22 08:17:41 crc kubenswrapper[4789]: I1122 08:17:41.009703 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/856539fc-0f5e-4ca1-8549-ad867a8c9f9d-config-data\") pod \"ceilometer-0\" (UID: \"856539fc-0f5e-4ca1-8549-ad867a8c9f9d\") " pod="openstack/ceilometer-0" Nov 22 08:17:41 crc kubenswrapper[4789]: I1122 08:17:41.009742 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g9kt7\" (UniqueName: \"kubernetes.io/projected/856539fc-0f5e-4ca1-8549-ad867a8c9f9d-kube-api-access-g9kt7\") pod \"ceilometer-0\" (UID: \"856539fc-0f5e-4ca1-8549-ad867a8c9f9d\") " pod="openstack/ceilometer-0" Nov 22 08:17:41 crc kubenswrapper[4789]: I1122 08:17:41.009782 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/856539fc-0f5e-4ca1-8549-ad867a8c9f9d-scripts\") pod \"ceilometer-0\" (UID: \"856539fc-0f5e-4ca1-8549-ad867a8c9f9d\") " pod="openstack/ceilometer-0" Nov 22 08:17:41 crc kubenswrapper[4789]: I1122 08:17:41.009815 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/856539fc-0f5e-4ca1-8549-ad867a8c9f9d-log-httpd\") pod \"ceilometer-0\" (UID: \"856539fc-0f5e-4ca1-8549-ad867a8c9f9d\") " pod="openstack/ceilometer-0" Nov 22 08:17:41 crc kubenswrapper[4789]: I1122 08:17:41.009900 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/856539fc-0f5e-4ca1-8549-ad867a8c9f9d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"856539fc-0f5e-4ca1-8549-ad867a8c9f9d\") " pod="openstack/ceilometer-0" Nov 22 08:17:41 crc kubenswrapper[4789]: I1122 08:17:41.009947 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/856539fc-0f5e-4ca1-8549-ad867a8c9f9d-run-httpd\") pod \"ceilometer-0\" (UID: \"856539fc-0f5e-4ca1-8549-ad867a8c9f9d\") " pod="openstack/ceilometer-0" Nov 22 08:17:41 crc kubenswrapper[4789]: I1122 08:17:41.009969 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/856539fc-0f5e-4ca1-8549-ad867a8c9f9d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"856539fc-0f5e-4ca1-8549-ad867a8c9f9d\") " pod="openstack/ceilometer-0" Nov 22 08:17:41 crc kubenswrapper[4789]: I1122 08:17:41.111915 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/856539fc-0f5e-4ca1-8549-ad867a8c9f9d-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"856539fc-0f5e-4ca1-8549-ad867a8c9f9d\") " pod="openstack/ceilometer-0" Nov 22 08:17:41 crc kubenswrapper[4789]: I1122 08:17:41.111958 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/856539fc-0f5e-4ca1-8549-ad867a8c9f9d-config-data\") pod \"ceilometer-0\" (UID: \"856539fc-0f5e-4ca1-8549-ad867a8c9f9d\") " pod="openstack/ceilometer-0" Nov 22 08:17:41 crc kubenswrapper[4789]: I1122 08:17:41.111993 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g9kt7\" (UniqueName: \"kubernetes.io/projected/856539fc-0f5e-4ca1-8549-ad867a8c9f9d-kube-api-access-g9kt7\") pod \"ceilometer-0\" (UID: \"856539fc-0f5e-4ca1-8549-ad867a8c9f9d\") " pod="openstack/ceilometer-0" Nov 22 08:17:41 crc kubenswrapper[4789]: I1122 08:17:41.112011 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/856539fc-0f5e-4ca1-8549-ad867a8c9f9d-scripts\") pod \"ceilometer-0\" (UID: \"856539fc-0f5e-4ca1-8549-ad867a8c9f9d\") " pod="openstack/ceilometer-0" Nov 22 08:17:41 crc kubenswrapper[4789]: I1122 08:17:41.112037 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/856539fc-0f5e-4ca1-8549-ad867a8c9f9d-log-httpd\") pod \"ceilometer-0\" (UID: \"856539fc-0f5e-4ca1-8549-ad867a8c9f9d\") " pod="openstack/ceilometer-0" Nov 22 08:17:41 crc kubenswrapper[4789]: I1122 08:17:41.112100 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/856539fc-0f5e-4ca1-8549-ad867a8c9f9d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"856539fc-0f5e-4ca1-8549-ad867a8c9f9d\") " pod="openstack/ceilometer-0" Nov 22 08:17:41 crc kubenswrapper[4789]: I1122 08:17:41.112142 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/856539fc-0f5e-4ca1-8549-ad867a8c9f9d-run-httpd\") pod \"ceilometer-0\" (UID: \"856539fc-0f5e-4ca1-8549-ad867a8c9f9d\") " pod="openstack/ceilometer-0" Nov 22 08:17:41 crc kubenswrapper[4789]: I1122 08:17:41.112163 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/856539fc-0f5e-4ca1-8549-ad867a8c9f9d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"856539fc-0f5e-4ca1-8549-ad867a8c9f9d\") " pod="openstack/ceilometer-0" Nov 22 08:17:41 crc kubenswrapper[4789]: I1122 08:17:41.112966 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/856539fc-0f5e-4ca1-8549-ad867a8c9f9d-log-httpd\") pod \"ceilometer-0\" (UID: \"856539fc-0f5e-4ca1-8549-ad867a8c9f9d\") " pod="openstack/ceilometer-0" Nov 22 08:17:41 crc kubenswrapper[4789]: I1122 08:17:41.113156 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/856539fc-0f5e-4ca1-8549-ad867a8c9f9d-run-httpd\") pod \"ceilometer-0\" (UID: \"856539fc-0f5e-4ca1-8549-ad867a8c9f9d\") " pod="openstack/ceilometer-0" Nov 22 08:17:41 crc kubenswrapper[4789]: I1122 08:17:41.116188 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/856539fc-0f5e-4ca1-8549-ad867a8c9f9d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"856539fc-0f5e-4ca1-8549-ad867a8c9f9d\") " pod="openstack/ceilometer-0" Nov 22 08:17:41 crc kubenswrapper[4789]: I1122 08:17:41.117006 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/856539fc-0f5e-4ca1-8549-ad867a8c9f9d-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"856539fc-0f5e-4ca1-8549-ad867a8c9f9d\") " pod="openstack/ceilometer-0" Nov 22 08:17:41 crc kubenswrapper[4789]: I1122 08:17:41.117185 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/856539fc-0f5e-4ca1-8549-ad867a8c9f9d-scripts\") pod \"ceilometer-0\" (UID: \"856539fc-0f5e-4ca1-8549-ad867a8c9f9d\") " pod="openstack/ceilometer-0" Nov 22 08:17:41 crc kubenswrapper[4789]: I1122 08:17:41.117882 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/856539fc-0f5e-4ca1-8549-ad867a8c9f9d-config-data\") pod \"ceilometer-0\" (UID: \"856539fc-0f5e-4ca1-8549-ad867a8c9f9d\") " pod="openstack/ceilometer-0" Nov 22 08:17:41 crc kubenswrapper[4789]: I1122 08:17:41.121241 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/856539fc-0f5e-4ca1-8549-ad867a8c9f9d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"856539fc-0f5e-4ca1-8549-ad867a8c9f9d\") " pod="openstack/ceilometer-0" Nov 22 08:17:41 crc kubenswrapper[4789]: I1122 08:17:41.131809 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g9kt7\" (UniqueName: \"kubernetes.io/projected/856539fc-0f5e-4ca1-8549-ad867a8c9f9d-kube-api-access-g9kt7\") pod \"ceilometer-0\" (UID: \"856539fc-0f5e-4ca1-8549-ad867a8c9f9d\") " pod="openstack/ceilometer-0" Nov 22 08:17:41 crc kubenswrapper[4789]: I1122 08:17:41.278799 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 08:17:41 crc kubenswrapper[4789]: I1122 08:17:41.832857 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:17:41 crc kubenswrapper[4789]: I1122 08:17:41.974421 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e740aa5b-0619-4092-b893-1ea056025be8" path="/var/lib/kubelet/pods/e740aa5b-0619-4092-b893-1ea056025be8/volumes" Nov 22 08:17:42 crc kubenswrapper[4789]: I1122 08:17:42.934487 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Nov 22 08:17:48 crc kubenswrapper[4789]: I1122 08:17:48.093935 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:17:48 crc kubenswrapper[4789]: I1122 08:17:48.673175 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"856539fc-0f5e-4ca1-8549-ad867a8c9f9d","Type":"ContainerStarted","Data":"504975174c5cdfc95f33065c2d75184032ca2130312974c233328e2c5ea60ef2"} Nov 22 08:17:50 crc kubenswrapper[4789]: I1122 08:17:50.690468 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-mg9lj" event={"ID":"8ea41118-064f-4760-a1cc-0ebe3fc81a02","Type":"ContainerStarted","Data":"8f13b0868aa5253a8993b673578603567f018491fca6b9f2caa8e7ea94d33934"} Nov 22 08:17:50 crc kubenswrapper[4789]: I1122 08:17:50.692409 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"856539fc-0f5e-4ca1-8549-ad867a8c9f9d","Type":"ContainerStarted","Data":"acc4d97cf4d1b79b9e4b7d195cba33808b9bcd8ef2955aec35f1a347e7d32c63"} Nov 22 08:17:50 crc kubenswrapper[4789]: I1122 08:17:50.720707 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-mg9lj" podStartSLOduration=2.231401122 podStartE2EDuration="15.720682374s" podCreationTimestamp="2025-11-22 08:17:35 +0000 UTC" firstStartedPulling="2025-11-22 08:17:35.90818459 +0000 UTC m=+1270.142584863" lastFinishedPulling="2025-11-22 08:17:49.397465852 +0000 UTC m=+1283.631866115" observedRunningTime="2025-11-22 08:17:50.707782172 +0000 UTC m=+1284.942182495" watchObservedRunningTime="2025-11-22 08:17:50.720682374 +0000 UTC m=+1284.955082677" Nov 22 08:17:52 crc kubenswrapper[4789]: I1122 08:17:52.736957 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"856539fc-0f5e-4ca1-8549-ad867a8c9f9d","Type":"ContainerStarted","Data":"f4cd6ed6644f823ba0812cebc2689b9d59a3a59af61900e4c24defb560d8721b"} Nov 22 08:17:53 crc kubenswrapper[4789]: I1122 08:17:53.755641 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"856539fc-0f5e-4ca1-8549-ad867a8c9f9d","Type":"ContainerStarted","Data":"011275144e4f64abc449f62b890778c065865977d16719107ebb53ea2c97a59f"} Nov 22 08:18:01 crc kubenswrapper[4789]: I1122 08:18:01.822261 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"856539fc-0f5e-4ca1-8549-ad867a8c9f9d","Type":"ContainerStarted","Data":"4cd28ed7b40e96e8db44506eb470515c7d474bc326d7a6c52dd52041a24beeda"} Nov 22 08:18:01 crc kubenswrapper[4789]: I1122 08:18:01.823377 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 22 08:18:01 crc kubenswrapper[4789]: I1122 08:18:01.822483 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="856539fc-0f5e-4ca1-8549-ad867a8c9f9d" containerName="sg-core" containerID="cri-o://011275144e4f64abc449f62b890778c065865977d16719107ebb53ea2c97a59f" gracePeriod=30 Nov 22 08:18:01 crc kubenswrapper[4789]: I1122 08:18:01.822429 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="856539fc-0f5e-4ca1-8549-ad867a8c9f9d" containerName="ceilometer-central-agent" containerID="cri-o://acc4d97cf4d1b79b9e4b7d195cba33808b9bcd8ef2955aec35f1a347e7d32c63" gracePeriod=30 Nov 22 08:18:01 crc kubenswrapper[4789]: I1122 08:18:01.822518 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="856539fc-0f5e-4ca1-8549-ad867a8c9f9d" containerName="ceilometer-notification-agent" containerID="cri-o://f4cd6ed6644f823ba0812cebc2689b9d59a3a59af61900e4c24defb560d8721b" gracePeriod=30 Nov 22 08:18:01 crc kubenswrapper[4789]: I1122 08:18:01.822530 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="856539fc-0f5e-4ca1-8549-ad867a8c9f9d" containerName="proxy-httpd" containerID="cri-o://4cd28ed7b40e96e8db44506eb470515c7d474bc326d7a6c52dd52041a24beeda" gracePeriod=30 Nov 22 08:18:01 crc kubenswrapper[4789]: I1122 08:18:01.849870 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=8.583190507 podStartE2EDuration="21.84985379s" podCreationTimestamp="2025-11-22 08:17:40 +0000 UTC" firstStartedPulling="2025-11-22 08:17:48.099655213 +0000 UTC m=+1282.334055486" lastFinishedPulling="2025-11-22 08:18:01.366318496 +0000 UTC m=+1295.600718769" observedRunningTime="2025-11-22 08:18:01.84472447 +0000 UTC m=+1296.079124773" watchObservedRunningTime="2025-11-22 08:18:01.84985379 +0000 UTC m=+1296.084254063" Nov 22 08:18:02 crc kubenswrapper[4789]: I1122 08:18:02.833884 4789 generic.go:334] "Generic (PLEG): container finished" podID="856539fc-0f5e-4ca1-8549-ad867a8c9f9d" containerID="4cd28ed7b40e96e8db44506eb470515c7d474bc326d7a6c52dd52041a24beeda" exitCode=0 Nov 22 08:18:02 crc kubenswrapper[4789]: I1122 08:18:02.833926 4789 generic.go:334] "Generic (PLEG): container finished" podID="856539fc-0f5e-4ca1-8549-ad867a8c9f9d" containerID="011275144e4f64abc449f62b890778c065865977d16719107ebb53ea2c97a59f" exitCode=2 Nov 22 08:18:02 crc kubenswrapper[4789]: I1122 08:18:02.833947 4789 generic.go:334] "Generic (PLEG): container finished" podID="856539fc-0f5e-4ca1-8549-ad867a8c9f9d" containerID="acc4d97cf4d1b79b9e4b7d195cba33808b9bcd8ef2955aec35f1a347e7d32c63" exitCode=0 Nov 22 08:18:02 crc kubenswrapper[4789]: I1122 08:18:02.833896 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"856539fc-0f5e-4ca1-8549-ad867a8c9f9d","Type":"ContainerDied","Data":"4cd28ed7b40e96e8db44506eb470515c7d474bc326d7a6c52dd52041a24beeda"} Nov 22 08:18:02 crc kubenswrapper[4789]: I1122 08:18:02.833999 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"856539fc-0f5e-4ca1-8549-ad867a8c9f9d","Type":"ContainerDied","Data":"011275144e4f64abc449f62b890778c065865977d16719107ebb53ea2c97a59f"} Nov 22 08:18:02 crc kubenswrapper[4789]: I1122 08:18:02.834018 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"856539fc-0f5e-4ca1-8549-ad867a8c9f9d","Type":"ContainerDied","Data":"acc4d97cf4d1b79b9e4b7d195cba33808b9bcd8ef2955aec35f1a347e7d32c63"} Nov 22 08:18:04 crc kubenswrapper[4789]: I1122 08:18:04.853548 4789 generic.go:334] "Generic (PLEG): container finished" podID="856539fc-0f5e-4ca1-8549-ad867a8c9f9d" containerID="f4cd6ed6644f823ba0812cebc2689b9d59a3a59af61900e4c24defb560d8721b" exitCode=0 Nov 22 08:18:04 crc kubenswrapper[4789]: I1122 08:18:04.853644 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"856539fc-0f5e-4ca1-8549-ad867a8c9f9d","Type":"ContainerDied","Data":"f4cd6ed6644f823ba0812cebc2689b9d59a3a59af61900e4c24defb560d8721b"} Nov 22 08:18:05 crc kubenswrapper[4789]: I1122 08:18:05.095968 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 08:18:05 crc kubenswrapper[4789]: I1122 08:18:05.239652 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/856539fc-0f5e-4ca1-8549-ad867a8c9f9d-config-data\") pod \"856539fc-0f5e-4ca1-8549-ad867a8c9f9d\" (UID: \"856539fc-0f5e-4ca1-8549-ad867a8c9f9d\") " Nov 22 08:18:05 crc kubenswrapper[4789]: I1122 08:18:05.239783 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/856539fc-0f5e-4ca1-8549-ad867a8c9f9d-sg-core-conf-yaml\") pod \"856539fc-0f5e-4ca1-8549-ad867a8c9f9d\" (UID: \"856539fc-0f5e-4ca1-8549-ad867a8c9f9d\") " Nov 22 08:18:05 crc kubenswrapper[4789]: I1122 08:18:05.239804 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/856539fc-0f5e-4ca1-8549-ad867a8c9f9d-scripts\") pod \"856539fc-0f5e-4ca1-8549-ad867a8c9f9d\" (UID: \"856539fc-0f5e-4ca1-8549-ad867a8c9f9d\") " Nov 22 08:18:05 crc kubenswrapper[4789]: I1122 08:18:05.239839 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/856539fc-0f5e-4ca1-8549-ad867a8c9f9d-combined-ca-bundle\") pod \"856539fc-0f5e-4ca1-8549-ad867a8c9f9d\" (UID: \"856539fc-0f5e-4ca1-8549-ad867a8c9f9d\") " Nov 22 08:18:05 crc kubenswrapper[4789]: I1122 08:18:05.239898 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g9kt7\" (UniqueName: \"kubernetes.io/projected/856539fc-0f5e-4ca1-8549-ad867a8c9f9d-kube-api-access-g9kt7\") pod \"856539fc-0f5e-4ca1-8549-ad867a8c9f9d\" (UID: \"856539fc-0f5e-4ca1-8549-ad867a8c9f9d\") " Nov 22 08:18:05 crc kubenswrapper[4789]: I1122 08:18:05.239926 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/856539fc-0f5e-4ca1-8549-ad867a8c9f9d-log-httpd\") pod \"856539fc-0f5e-4ca1-8549-ad867a8c9f9d\" (UID: \"856539fc-0f5e-4ca1-8549-ad867a8c9f9d\") " Nov 22 08:18:05 crc kubenswrapper[4789]: I1122 08:18:05.239983 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/856539fc-0f5e-4ca1-8549-ad867a8c9f9d-run-httpd\") pod \"856539fc-0f5e-4ca1-8549-ad867a8c9f9d\" (UID: \"856539fc-0f5e-4ca1-8549-ad867a8c9f9d\") " Nov 22 08:18:05 crc kubenswrapper[4789]: I1122 08:18:05.240010 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/856539fc-0f5e-4ca1-8549-ad867a8c9f9d-ceilometer-tls-certs\") pod \"856539fc-0f5e-4ca1-8549-ad867a8c9f9d\" (UID: \"856539fc-0f5e-4ca1-8549-ad867a8c9f9d\") " Nov 22 08:18:05 crc kubenswrapper[4789]: I1122 08:18:05.240552 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/856539fc-0f5e-4ca1-8549-ad867a8c9f9d-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "856539fc-0f5e-4ca1-8549-ad867a8c9f9d" (UID: "856539fc-0f5e-4ca1-8549-ad867a8c9f9d"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:18:05 crc kubenswrapper[4789]: I1122 08:18:05.240818 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/856539fc-0f5e-4ca1-8549-ad867a8c9f9d-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "856539fc-0f5e-4ca1-8549-ad867a8c9f9d" (UID: "856539fc-0f5e-4ca1-8549-ad867a8c9f9d"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:18:05 crc kubenswrapper[4789]: I1122 08:18:05.244890 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/856539fc-0f5e-4ca1-8549-ad867a8c9f9d-scripts" (OuterVolumeSpecName: "scripts") pod "856539fc-0f5e-4ca1-8549-ad867a8c9f9d" (UID: "856539fc-0f5e-4ca1-8549-ad867a8c9f9d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:18:05 crc kubenswrapper[4789]: I1122 08:18:05.245834 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/856539fc-0f5e-4ca1-8549-ad867a8c9f9d-kube-api-access-g9kt7" (OuterVolumeSpecName: "kube-api-access-g9kt7") pod "856539fc-0f5e-4ca1-8549-ad867a8c9f9d" (UID: "856539fc-0f5e-4ca1-8549-ad867a8c9f9d"). InnerVolumeSpecName "kube-api-access-g9kt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:18:05 crc kubenswrapper[4789]: I1122 08:18:05.274782 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/856539fc-0f5e-4ca1-8549-ad867a8c9f9d-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "856539fc-0f5e-4ca1-8549-ad867a8c9f9d" (UID: "856539fc-0f5e-4ca1-8549-ad867a8c9f9d"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:18:05 crc kubenswrapper[4789]: I1122 08:18:05.285791 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/856539fc-0f5e-4ca1-8549-ad867a8c9f9d-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "856539fc-0f5e-4ca1-8549-ad867a8c9f9d" (UID: "856539fc-0f5e-4ca1-8549-ad867a8c9f9d"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:18:05 crc kubenswrapper[4789]: I1122 08:18:05.318249 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/856539fc-0f5e-4ca1-8549-ad867a8c9f9d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "856539fc-0f5e-4ca1-8549-ad867a8c9f9d" (UID: "856539fc-0f5e-4ca1-8549-ad867a8c9f9d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:18:05 crc kubenswrapper[4789]: I1122 08:18:05.334089 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/856539fc-0f5e-4ca1-8549-ad867a8c9f9d-config-data" (OuterVolumeSpecName: "config-data") pod "856539fc-0f5e-4ca1-8549-ad867a8c9f9d" (UID: "856539fc-0f5e-4ca1-8549-ad867a8c9f9d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:18:05 crc kubenswrapper[4789]: I1122 08:18:05.342508 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g9kt7\" (UniqueName: \"kubernetes.io/projected/856539fc-0f5e-4ca1-8549-ad867a8c9f9d-kube-api-access-g9kt7\") on node \"crc\" DevicePath \"\"" Nov 22 08:18:05 crc kubenswrapper[4789]: I1122 08:18:05.342559 4789 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/856539fc-0f5e-4ca1-8549-ad867a8c9f9d-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 22 08:18:05 crc kubenswrapper[4789]: I1122 08:18:05.342571 4789 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/856539fc-0f5e-4ca1-8549-ad867a8c9f9d-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 22 08:18:05 crc kubenswrapper[4789]: I1122 08:18:05.342582 4789 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/856539fc-0f5e-4ca1-8549-ad867a8c9f9d-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 22 08:18:05 crc kubenswrapper[4789]: I1122 08:18:05.342593 4789 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/856539fc-0f5e-4ca1-8549-ad867a8c9f9d-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:18:05 crc kubenswrapper[4789]: I1122 08:18:05.342605 4789 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/856539fc-0f5e-4ca1-8549-ad867a8c9f9d-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 22 08:18:05 crc kubenswrapper[4789]: I1122 08:18:05.342621 4789 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/856539fc-0f5e-4ca1-8549-ad867a8c9f9d-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:18:05 crc kubenswrapper[4789]: I1122 08:18:05.342632 4789 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/856539fc-0f5e-4ca1-8549-ad867a8c9f9d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:18:05 crc kubenswrapper[4789]: I1122 08:18:05.865425 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"856539fc-0f5e-4ca1-8549-ad867a8c9f9d","Type":"ContainerDied","Data":"504975174c5cdfc95f33065c2d75184032ca2130312974c233328e2c5ea60ef2"} Nov 22 08:18:05 crc kubenswrapper[4789]: I1122 08:18:05.865478 4789 scope.go:117] "RemoveContainer" containerID="4cd28ed7b40e96e8db44506eb470515c7d474bc326d7a6c52dd52041a24beeda" Nov 22 08:18:05 crc kubenswrapper[4789]: I1122 08:18:05.865515 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 08:18:05 crc kubenswrapper[4789]: I1122 08:18:05.885055 4789 scope.go:117] "RemoveContainer" containerID="011275144e4f64abc449f62b890778c065865977d16719107ebb53ea2c97a59f" Nov 22 08:18:05 crc kubenswrapper[4789]: I1122 08:18:05.899275 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:18:05 crc kubenswrapper[4789]: I1122 08:18:05.905636 4789 scope.go:117] "RemoveContainer" containerID="f4cd6ed6644f823ba0812cebc2689b9d59a3a59af61900e4c24defb560d8721b" Nov 22 08:18:05 crc kubenswrapper[4789]: I1122 08:18:05.905724 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:18:05 crc kubenswrapper[4789]: I1122 08:18:05.924677 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:18:05 crc kubenswrapper[4789]: E1122 08:18:05.925001 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="856539fc-0f5e-4ca1-8549-ad867a8c9f9d" containerName="sg-core" Nov 22 08:18:05 crc kubenswrapper[4789]: I1122 08:18:05.925018 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="856539fc-0f5e-4ca1-8549-ad867a8c9f9d" containerName="sg-core" Nov 22 08:18:05 crc kubenswrapper[4789]: E1122 08:18:05.925039 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="856539fc-0f5e-4ca1-8549-ad867a8c9f9d" containerName="ceilometer-central-agent" Nov 22 08:18:05 crc kubenswrapper[4789]: I1122 08:18:05.925046 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="856539fc-0f5e-4ca1-8549-ad867a8c9f9d" containerName="ceilometer-central-agent" Nov 22 08:18:05 crc kubenswrapper[4789]: E1122 08:18:05.925059 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="856539fc-0f5e-4ca1-8549-ad867a8c9f9d" containerName="ceilometer-notification-agent" Nov 22 08:18:05 crc kubenswrapper[4789]: I1122 08:18:05.925064 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="856539fc-0f5e-4ca1-8549-ad867a8c9f9d" containerName="ceilometer-notification-agent" Nov 22 08:18:05 crc kubenswrapper[4789]: E1122 08:18:05.925087 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="856539fc-0f5e-4ca1-8549-ad867a8c9f9d" containerName="proxy-httpd" Nov 22 08:18:05 crc kubenswrapper[4789]: I1122 08:18:05.925093 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="856539fc-0f5e-4ca1-8549-ad867a8c9f9d" containerName="proxy-httpd" Nov 22 08:18:05 crc kubenswrapper[4789]: I1122 08:18:05.925243 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="856539fc-0f5e-4ca1-8549-ad867a8c9f9d" containerName="ceilometer-central-agent" Nov 22 08:18:05 crc kubenswrapper[4789]: I1122 08:18:05.925252 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="856539fc-0f5e-4ca1-8549-ad867a8c9f9d" containerName="sg-core" Nov 22 08:18:05 crc kubenswrapper[4789]: I1122 08:18:05.925262 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="856539fc-0f5e-4ca1-8549-ad867a8c9f9d" containerName="proxy-httpd" Nov 22 08:18:05 crc kubenswrapper[4789]: I1122 08:18:05.925276 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="856539fc-0f5e-4ca1-8549-ad867a8c9f9d" containerName="ceilometer-notification-agent" Nov 22 08:18:05 crc kubenswrapper[4789]: I1122 08:18:05.926764 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 08:18:05 crc kubenswrapper[4789]: I1122 08:18:05.932496 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 22 08:18:05 crc kubenswrapper[4789]: I1122 08:18:05.932607 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 22 08:18:05 crc kubenswrapper[4789]: I1122 08:18:05.933053 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 22 08:18:05 crc kubenswrapper[4789]: I1122 08:18:05.934106 4789 scope.go:117] "RemoveContainer" containerID="acc4d97cf4d1b79b9e4b7d195cba33808b9bcd8ef2955aec35f1a347e7d32c63" Nov 22 08:18:05 crc kubenswrapper[4789]: I1122 08:18:05.939664 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:18:05 crc kubenswrapper[4789]: I1122 08:18:05.979459 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="856539fc-0f5e-4ca1-8549-ad867a8c9f9d" path="/var/lib/kubelet/pods/856539fc-0f5e-4ca1-8549-ad867a8c9f9d/volumes" Nov 22 08:18:06 crc kubenswrapper[4789]: I1122 08:18:06.054667 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h2gvj\" (UniqueName: \"kubernetes.io/projected/6ff6994a-950e-4aae-8126-1f55702f7eba-kube-api-access-h2gvj\") pod \"ceilometer-0\" (UID: \"6ff6994a-950e-4aae-8126-1f55702f7eba\") " pod="openstack/ceilometer-0" Nov 22 08:18:06 crc kubenswrapper[4789]: I1122 08:18:06.055154 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6ff6994a-950e-4aae-8126-1f55702f7eba-run-httpd\") pod \"ceilometer-0\" (UID: \"6ff6994a-950e-4aae-8126-1f55702f7eba\") " pod="openstack/ceilometer-0" Nov 22 08:18:06 crc kubenswrapper[4789]: I1122 08:18:06.055264 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/6ff6994a-950e-4aae-8126-1f55702f7eba-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"6ff6994a-950e-4aae-8126-1f55702f7eba\") " pod="openstack/ceilometer-0" Nov 22 08:18:06 crc kubenswrapper[4789]: I1122 08:18:06.055362 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ff6994a-950e-4aae-8126-1f55702f7eba-config-data\") pod \"ceilometer-0\" (UID: \"6ff6994a-950e-4aae-8126-1f55702f7eba\") " pod="openstack/ceilometer-0" Nov 22 08:18:06 crc kubenswrapper[4789]: I1122 08:18:06.055482 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6ff6994a-950e-4aae-8126-1f55702f7eba-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"6ff6994a-950e-4aae-8126-1f55702f7eba\") " pod="openstack/ceilometer-0" Nov 22 08:18:06 crc kubenswrapper[4789]: I1122 08:18:06.055565 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6ff6994a-950e-4aae-8126-1f55702f7eba-scripts\") pod \"ceilometer-0\" (UID: \"6ff6994a-950e-4aae-8126-1f55702f7eba\") " pod="openstack/ceilometer-0" Nov 22 08:18:06 crc kubenswrapper[4789]: I1122 08:18:06.055656 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6ff6994a-950e-4aae-8126-1f55702f7eba-log-httpd\") pod \"ceilometer-0\" (UID: \"6ff6994a-950e-4aae-8126-1f55702f7eba\") " pod="openstack/ceilometer-0" Nov 22 08:18:06 crc kubenswrapper[4789]: I1122 08:18:06.055793 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ff6994a-950e-4aae-8126-1f55702f7eba-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"6ff6994a-950e-4aae-8126-1f55702f7eba\") " pod="openstack/ceilometer-0" Nov 22 08:18:06 crc kubenswrapper[4789]: I1122 08:18:06.157707 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6ff6994a-950e-4aae-8126-1f55702f7eba-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"6ff6994a-950e-4aae-8126-1f55702f7eba\") " pod="openstack/ceilometer-0" Nov 22 08:18:06 crc kubenswrapper[4789]: I1122 08:18:06.157782 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6ff6994a-950e-4aae-8126-1f55702f7eba-scripts\") pod \"ceilometer-0\" (UID: \"6ff6994a-950e-4aae-8126-1f55702f7eba\") " pod="openstack/ceilometer-0" Nov 22 08:18:06 crc kubenswrapper[4789]: I1122 08:18:06.157809 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6ff6994a-950e-4aae-8126-1f55702f7eba-log-httpd\") pod \"ceilometer-0\" (UID: \"6ff6994a-950e-4aae-8126-1f55702f7eba\") " pod="openstack/ceilometer-0" Nov 22 08:18:06 crc kubenswrapper[4789]: I1122 08:18:06.157839 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ff6994a-950e-4aae-8126-1f55702f7eba-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"6ff6994a-950e-4aae-8126-1f55702f7eba\") " pod="openstack/ceilometer-0" Nov 22 08:18:06 crc kubenswrapper[4789]: I1122 08:18:06.157872 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h2gvj\" (UniqueName: \"kubernetes.io/projected/6ff6994a-950e-4aae-8126-1f55702f7eba-kube-api-access-h2gvj\") pod \"ceilometer-0\" (UID: \"6ff6994a-950e-4aae-8126-1f55702f7eba\") " pod="openstack/ceilometer-0" Nov 22 08:18:06 crc kubenswrapper[4789]: I1122 08:18:06.157891 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6ff6994a-950e-4aae-8126-1f55702f7eba-run-httpd\") pod \"ceilometer-0\" (UID: \"6ff6994a-950e-4aae-8126-1f55702f7eba\") " pod="openstack/ceilometer-0" Nov 22 08:18:06 crc kubenswrapper[4789]: I1122 08:18:06.157904 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/6ff6994a-950e-4aae-8126-1f55702f7eba-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"6ff6994a-950e-4aae-8126-1f55702f7eba\") " pod="openstack/ceilometer-0" Nov 22 08:18:06 crc kubenswrapper[4789]: I1122 08:18:06.157940 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ff6994a-950e-4aae-8126-1f55702f7eba-config-data\") pod \"ceilometer-0\" (UID: \"6ff6994a-950e-4aae-8126-1f55702f7eba\") " pod="openstack/ceilometer-0" Nov 22 08:18:06 crc kubenswrapper[4789]: I1122 08:18:06.158483 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6ff6994a-950e-4aae-8126-1f55702f7eba-log-httpd\") pod \"ceilometer-0\" (UID: \"6ff6994a-950e-4aae-8126-1f55702f7eba\") " pod="openstack/ceilometer-0" Nov 22 08:18:06 crc kubenswrapper[4789]: I1122 08:18:06.159980 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6ff6994a-950e-4aae-8126-1f55702f7eba-run-httpd\") pod \"ceilometer-0\" (UID: \"6ff6994a-950e-4aae-8126-1f55702f7eba\") " pod="openstack/ceilometer-0" Nov 22 08:18:06 crc kubenswrapper[4789]: I1122 08:18:06.161963 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/6ff6994a-950e-4aae-8126-1f55702f7eba-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"6ff6994a-950e-4aae-8126-1f55702f7eba\") " pod="openstack/ceilometer-0" Nov 22 08:18:06 crc kubenswrapper[4789]: I1122 08:18:06.170597 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6ff6994a-950e-4aae-8126-1f55702f7eba-scripts\") pod \"ceilometer-0\" (UID: \"6ff6994a-950e-4aae-8126-1f55702f7eba\") " pod="openstack/ceilometer-0" Nov 22 08:18:06 crc kubenswrapper[4789]: I1122 08:18:06.170636 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ff6994a-950e-4aae-8126-1f55702f7eba-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"6ff6994a-950e-4aae-8126-1f55702f7eba\") " pod="openstack/ceilometer-0" Nov 22 08:18:06 crc kubenswrapper[4789]: I1122 08:18:06.170818 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ff6994a-950e-4aae-8126-1f55702f7eba-config-data\") pod \"ceilometer-0\" (UID: \"6ff6994a-950e-4aae-8126-1f55702f7eba\") " pod="openstack/ceilometer-0" Nov 22 08:18:06 crc kubenswrapper[4789]: I1122 08:18:06.172466 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6ff6994a-950e-4aae-8126-1f55702f7eba-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"6ff6994a-950e-4aae-8126-1f55702f7eba\") " pod="openstack/ceilometer-0" Nov 22 08:18:06 crc kubenswrapper[4789]: I1122 08:18:06.172820 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h2gvj\" (UniqueName: \"kubernetes.io/projected/6ff6994a-950e-4aae-8126-1f55702f7eba-kube-api-access-h2gvj\") pod \"ceilometer-0\" (UID: \"6ff6994a-950e-4aae-8126-1f55702f7eba\") " pod="openstack/ceilometer-0" Nov 22 08:18:06 crc kubenswrapper[4789]: I1122 08:18:06.257045 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 08:18:06 crc kubenswrapper[4789]: I1122 08:18:06.684820 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:18:06 crc kubenswrapper[4789]: W1122 08:18:06.692869 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6ff6994a_950e_4aae_8126_1f55702f7eba.slice/crio-8b9efd471d7b1a57c2270c235f707634e2d025f454cdebccd3eb354c61e8027b WatchSource:0}: Error finding container 8b9efd471d7b1a57c2270c235f707634e2d025f454cdebccd3eb354c61e8027b: Status 404 returned error can't find the container with id 8b9efd471d7b1a57c2270c235f707634e2d025f454cdebccd3eb354c61e8027b Nov 22 08:18:06 crc kubenswrapper[4789]: I1122 08:18:06.891205 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6ff6994a-950e-4aae-8126-1f55702f7eba","Type":"ContainerStarted","Data":"8b9efd471d7b1a57c2270c235f707634e2d025f454cdebccd3eb354c61e8027b"} Nov 22 08:18:07 crc kubenswrapper[4789]: I1122 08:18:07.901130 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6ff6994a-950e-4aae-8126-1f55702f7eba","Type":"ContainerStarted","Data":"7211383e403d85898341931feec5b06f1e65e2c29b5288f6582db9b01610d7c9"} Nov 22 08:18:09 crc kubenswrapper[4789]: I1122 08:18:09.920037 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6ff6994a-950e-4aae-8126-1f55702f7eba","Type":"ContainerStarted","Data":"264d4c5e4620a285e31eeaf8220e3156dbbfc1893c0d8b9f1a01d6947b8a9e7b"} Nov 22 08:18:10 crc kubenswrapper[4789]: I1122 08:18:10.939459 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6ff6994a-950e-4aae-8126-1f55702f7eba","Type":"ContainerStarted","Data":"2bfdd608fd028e7cbf93dfab2e22298c8e6476fc7c4897dc406c59747d223d3a"} Nov 22 08:18:16 crc kubenswrapper[4789]: I1122 08:18:16.005201 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6ff6994a-950e-4aae-8126-1f55702f7eba","Type":"ContainerStarted","Data":"7f2db1070da806a8e0a6409cda690ba4b4cfadd5f9454f9ef89548607b6c2de9"} Nov 22 08:18:16 crc kubenswrapper[4789]: I1122 08:18:16.034658 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.420196115 podStartE2EDuration="11.034637863s" podCreationTimestamp="2025-11-22 08:18:05 +0000 UTC" firstStartedPulling="2025-11-22 08:18:06.694867839 +0000 UTC m=+1300.929268112" lastFinishedPulling="2025-11-22 08:18:15.309309587 +0000 UTC m=+1309.543709860" observedRunningTime="2025-11-22 08:18:16.029312068 +0000 UTC m=+1310.263712341" watchObservedRunningTime="2025-11-22 08:18:16.034637863 +0000 UTC m=+1310.269038136" Nov 22 08:18:17 crc kubenswrapper[4789]: I1122 08:18:17.013374 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 22 08:18:26 crc kubenswrapper[4789]: I1122 08:18:26.091730 4789 generic.go:334] "Generic (PLEG): container finished" podID="8ea41118-064f-4760-a1cc-0ebe3fc81a02" containerID="8f13b0868aa5253a8993b673578603567f018491fca6b9f2caa8e7ea94d33934" exitCode=0 Nov 22 08:18:26 crc kubenswrapper[4789]: I1122 08:18:26.091777 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-mg9lj" event={"ID":"8ea41118-064f-4760-a1cc-0ebe3fc81a02","Type":"ContainerDied","Data":"8f13b0868aa5253a8993b673578603567f018491fca6b9f2caa8e7ea94d33934"} Nov 22 08:18:27 crc kubenswrapper[4789]: I1122 08:18:27.435843 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-mg9lj" Nov 22 08:18:27 crc kubenswrapper[4789]: I1122 08:18:27.628070 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ea41118-064f-4760-a1cc-0ebe3fc81a02-combined-ca-bundle\") pod \"8ea41118-064f-4760-a1cc-0ebe3fc81a02\" (UID: \"8ea41118-064f-4760-a1cc-0ebe3fc81a02\") " Nov 22 08:18:27 crc kubenswrapper[4789]: I1122 08:18:27.628480 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ea41118-064f-4760-a1cc-0ebe3fc81a02-scripts\") pod \"8ea41118-064f-4760-a1cc-0ebe3fc81a02\" (UID: \"8ea41118-064f-4760-a1cc-0ebe3fc81a02\") " Nov 22 08:18:27 crc kubenswrapper[4789]: I1122 08:18:27.628544 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vlcqh\" (UniqueName: \"kubernetes.io/projected/8ea41118-064f-4760-a1cc-0ebe3fc81a02-kube-api-access-vlcqh\") pod \"8ea41118-064f-4760-a1cc-0ebe3fc81a02\" (UID: \"8ea41118-064f-4760-a1cc-0ebe3fc81a02\") " Nov 22 08:18:27 crc kubenswrapper[4789]: I1122 08:18:27.628571 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ea41118-064f-4760-a1cc-0ebe3fc81a02-config-data\") pod \"8ea41118-064f-4760-a1cc-0ebe3fc81a02\" (UID: \"8ea41118-064f-4760-a1cc-0ebe3fc81a02\") " Nov 22 08:18:27 crc kubenswrapper[4789]: I1122 08:18:27.635008 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ea41118-064f-4760-a1cc-0ebe3fc81a02-kube-api-access-vlcqh" (OuterVolumeSpecName: "kube-api-access-vlcqh") pod "8ea41118-064f-4760-a1cc-0ebe3fc81a02" (UID: "8ea41118-064f-4760-a1cc-0ebe3fc81a02"). InnerVolumeSpecName "kube-api-access-vlcqh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:18:27 crc kubenswrapper[4789]: I1122 08:18:27.635919 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ea41118-064f-4760-a1cc-0ebe3fc81a02-scripts" (OuterVolumeSpecName: "scripts") pod "8ea41118-064f-4760-a1cc-0ebe3fc81a02" (UID: "8ea41118-064f-4760-a1cc-0ebe3fc81a02"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:18:27 crc kubenswrapper[4789]: I1122 08:18:27.655775 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ea41118-064f-4760-a1cc-0ebe3fc81a02-config-data" (OuterVolumeSpecName: "config-data") pod "8ea41118-064f-4760-a1cc-0ebe3fc81a02" (UID: "8ea41118-064f-4760-a1cc-0ebe3fc81a02"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:18:27 crc kubenswrapper[4789]: I1122 08:18:27.656176 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ea41118-064f-4760-a1cc-0ebe3fc81a02-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8ea41118-064f-4760-a1cc-0ebe3fc81a02" (UID: "8ea41118-064f-4760-a1cc-0ebe3fc81a02"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:18:27 crc kubenswrapper[4789]: I1122 08:18:27.732629 4789 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ea41118-064f-4760-a1cc-0ebe3fc81a02-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:18:27 crc kubenswrapper[4789]: I1122 08:18:27.732701 4789 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ea41118-064f-4760-a1cc-0ebe3fc81a02-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:18:27 crc kubenswrapper[4789]: I1122 08:18:27.732724 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vlcqh\" (UniqueName: \"kubernetes.io/projected/8ea41118-064f-4760-a1cc-0ebe3fc81a02-kube-api-access-vlcqh\") on node \"crc\" DevicePath \"\"" Nov 22 08:18:27 crc kubenswrapper[4789]: I1122 08:18:27.732784 4789 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ea41118-064f-4760-a1cc-0ebe3fc81a02-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:18:28 crc kubenswrapper[4789]: I1122 08:18:28.108367 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-mg9lj" event={"ID":"8ea41118-064f-4760-a1cc-0ebe3fc81a02","Type":"ContainerDied","Data":"2c15d8971b3f53e1ba28fd992ea1a87344d646ffefac0c97ea5be78c52f84efe"} Nov 22 08:18:28 crc kubenswrapper[4789]: I1122 08:18:28.108417 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2c15d8971b3f53e1ba28fd992ea1a87344d646ffefac0c97ea5be78c52f84efe" Nov 22 08:18:28 crc kubenswrapper[4789]: I1122 08:18:28.108415 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-mg9lj" Nov 22 08:18:28 crc kubenswrapper[4789]: I1122 08:18:28.193304 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 22 08:18:28 crc kubenswrapper[4789]: E1122 08:18:28.193786 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ea41118-064f-4760-a1cc-0ebe3fc81a02" containerName="nova-cell0-conductor-db-sync" Nov 22 08:18:28 crc kubenswrapper[4789]: I1122 08:18:28.193805 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ea41118-064f-4760-a1cc-0ebe3fc81a02" containerName="nova-cell0-conductor-db-sync" Nov 22 08:18:28 crc kubenswrapper[4789]: I1122 08:18:28.194034 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ea41118-064f-4760-a1cc-0ebe3fc81a02" containerName="nova-cell0-conductor-db-sync" Nov 22 08:18:28 crc kubenswrapper[4789]: I1122 08:18:28.194736 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 22 08:18:28 crc kubenswrapper[4789]: I1122 08:18:28.200112 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-mckw6" Nov 22 08:18:28 crc kubenswrapper[4789]: I1122 08:18:28.200905 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Nov 22 08:18:28 crc kubenswrapper[4789]: I1122 08:18:28.207722 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 22 08:18:28 crc kubenswrapper[4789]: I1122 08:18:28.341966 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94983cbb-ec60-487e-8feb-42196210f1b6-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"94983cbb-ec60-487e-8feb-42196210f1b6\") " pod="openstack/nova-cell0-conductor-0" Nov 22 08:18:28 crc kubenswrapper[4789]: I1122 08:18:28.342094 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jh7vp\" (UniqueName: \"kubernetes.io/projected/94983cbb-ec60-487e-8feb-42196210f1b6-kube-api-access-jh7vp\") pod \"nova-cell0-conductor-0\" (UID: \"94983cbb-ec60-487e-8feb-42196210f1b6\") " pod="openstack/nova-cell0-conductor-0" Nov 22 08:18:28 crc kubenswrapper[4789]: I1122 08:18:28.342836 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94983cbb-ec60-487e-8feb-42196210f1b6-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"94983cbb-ec60-487e-8feb-42196210f1b6\") " pod="openstack/nova-cell0-conductor-0" Nov 22 08:18:28 crc kubenswrapper[4789]: I1122 08:18:28.445300 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jh7vp\" (UniqueName: \"kubernetes.io/projected/94983cbb-ec60-487e-8feb-42196210f1b6-kube-api-access-jh7vp\") pod \"nova-cell0-conductor-0\" (UID: \"94983cbb-ec60-487e-8feb-42196210f1b6\") " pod="openstack/nova-cell0-conductor-0" Nov 22 08:18:28 crc kubenswrapper[4789]: I1122 08:18:28.445358 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94983cbb-ec60-487e-8feb-42196210f1b6-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"94983cbb-ec60-487e-8feb-42196210f1b6\") " pod="openstack/nova-cell0-conductor-0" Nov 22 08:18:28 crc kubenswrapper[4789]: I1122 08:18:28.445444 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94983cbb-ec60-487e-8feb-42196210f1b6-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"94983cbb-ec60-487e-8feb-42196210f1b6\") " pod="openstack/nova-cell0-conductor-0" Nov 22 08:18:28 crc kubenswrapper[4789]: I1122 08:18:28.449228 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94983cbb-ec60-487e-8feb-42196210f1b6-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"94983cbb-ec60-487e-8feb-42196210f1b6\") " pod="openstack/nova-cell0-conductor-0" Nov 22 08:18:28 crc kubenswrapper[4789]: I1122 08:18:28.450831 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94983cbb-ec60-487e-8feb-42196210f1b6-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"94983cbb-ec60-487e-8feb-42196210f1b6\") " pod="openstack/nova-cell0-conductor-0" Nov 22 08:18:28 crc kubenswrapper[4789]: I1122 08:18:28.463122 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jh7vp\" (UniqueName: \"kubernetes.io/projected/94983cbb-ec60-487e-8feb-42196210f1b6-kube-api-access-jh7vp\") pod \"nova-cell0-conductor-0\" (UID: \"94983cbb-ec60-487e-8feb-42196210f1b6\") " pod="openstack/nova-cell0-conductor-0" Nov 22 08:18:28 crc kubenswrapper[4789]: I1122 08:18:28.514154 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 22 08:18:28 crc kubenswrapper[4789]: I1122 08:18:28.975310 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 22 08:18:28 crc kubenswrapper[4789]: W1122 08:18:28.978559 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod94983cbb_ec60_487e_8feb_42196210f1b6.slice/crio-fb6eb9772dfbae2eaa259a5694609f4fa4b40540c2291a384097f61b1361c1a6 WatchSource:0}: Error finding container fb6eb9772dfbae2eaa259a5694609f4fa4b40540c2291a384097f61b1361c1a6: Status 404 returned error can't find the container with id fb6eb9772dfbae2eaa259a5694609f4fa4b40540c2291a384097f61b1361c1a6 Nov 22 08:18:29 crc kubenswrapper[4789]: I1122 08:18:29.116725 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"94983cbb-ec60-487e-8feb-42196210f1b6","Type":"ContainerStarted","Data":"fb6eb9772dfbae2eaa259a5694609f4fa4b40540c2291a384097f61b1361c1a6"} Nov 22 08:18:30 crc kubenswrapper[4789]: I1122 08:18:30.127836 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"94983cbb-ec60-487e-8feb-42196210f1b6","Type":"ContainerStarted","Data":"18f79500883e4bab0869d9178367328cddf9b8fee3f88074ea37d9e73a184cb5"} Nov 22 08:18:30 crc kubenswrapper[4789]: I1122 08:18:30.128224 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Nov 22 08:18:35 crc kubenswrapper[4789]: I1122 08:18:35.372115 4789 patch_prober.go:28] interesting pod/machine-config-daemon-pxvrm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 08:18:35 crc kubenswrapper[4789]: I1122 08:18:35.372703 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 08:18:36 crc kubenswrapper[4789]: I1122 08:18:36.267719 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 22 08:18:36 crc kubenswrapper[4789]: I1122 08:18:36.309864 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=8.309842643 podStartE2EDuration="8.309842643s" podCreationTimestamp="2025-11-22 08:18:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:18:30.147575905 +0000 UTC m=+1324.381976198" watchObservedRunningTime="2025-11-22 08:18:36.309842643 +0000 UTC m=+1330.544242926" Nov 22 08:18:38 crc kubenswrapper[4789]: I1122 08:18:38.541930 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Nov 22 08:18:38 crc kubenswrapper[4789]: I1122 08:18:38.965957 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-fzs5c"] Nov 22 08:18:38 crc kubenswrapper[4789]: I1122 08:18:38.968630 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-fzs5c" Nov 22 08:18:38 crc kubenswrapper[4789]: I1122 08:18:38.970615 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Nov 22 08:18:38 crc kubenswrapper[4789]: I1122 08:18:38.970846 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Nov 22 08:18:38 crc kubenswrapper[4789]: I1122 08:18:38.976712 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-fzs5c"] Nov 22 08:18:39 crc kubenswrapper[4789]: I1122 08:18:39.134058 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79d1ad09-f02f-421e-b1e5-10860722cc77-config-data\") pod \"nova-cell0-cell-mapping-fzs5c\" (UID: \"79d1ad09-f02f-421e-b1e5-10860722cc77\") " pod="openstack/nova-cell0-cell-mapping-fzs5c" Nov 22 08:18:39 crc kubenswrapper[4789]: I1122 08:18:39.134636 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/79d1ad09-f02f-421e-b1e5-10860722cc77-scripts\") pod \"nova-cell0-cell-mapping-fzs5c\" (UID: \"79d1ad09-f02f-421e-b1e5-10860722cc77\") " pod="openstack/nova-cell0-cell-mapping-fzs5c" Nov 22 08:18:39 crc kubenswrapper[4789]: I1122 08:18:39.134785 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mb5qv\" (UniqueName: \"kubernetes.io/projected/79d1ad09-f02f-421e-b1e5-10860722cc77-kube-api-access-mb5qv\") pod \"nova-cell0-cell-mapping-fzs5c\" (UID: \"79d1ad09-f02f-421e-b1e5-10860722cc77\") " pod="openstack/nova-cell0-cell-mapping-fzs5c" Nov 22 08:18:39 crc kubenswrapper[4789]: I1122 08:18:39.134909 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79d1ad09-f02f-421e-b1e5-10860722cc77-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-fzs5c\" (UID: \"79d1ad09-f02f-421e-b1e5-10860722cc77\") " pod="openstack/nova-cell0-cell-mapping-fzs5c" Nov 22 08:18:39 crc kubenswrapper[4789]: I1122 08:18:39.237918 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79d1ad09-f02f-421e-b1e5-10860722cc77-config-data\") pod \"nova-cell0-cell-mapping-fzs5c\" (UID: \"79d1ad09-f02f-421e-b1e5-10860722cc77\") " pod="openstack/nova-cell0-cell-mapping-fzs5c" Nov 22 08:18:39 crc kubenswrapper[4789]: I1122 08:18:39.238033 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/79d1ad09-f02f-421e-b1e5-10860722cc77-scripts\") pod \"nova-cell0-cell-mapping-fzs5c\" (UID: \"79d1ad09-f02f-421e-b1e5-10860722cc77\") " pod="openstack/nova-cell0-cell-mapping-fzs5c" Nov 22 08:18:39 crc kubenswrapper[4789]: I1122 08:18:39.238069 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mb5qv\" (UniqueName: \"kubernetes.io/projected/79d1ad09-f02f-421e-b1e5-10860722cc77-kube-api-access-mb5qv\") pod \"nova-cell0-cell-mapping-fzs5c\" (UID: \"79d1ad09-f02f-421e-b1e5-10860722cc77\") " pod="openstack/nova-cell0-cell-mapping-fzs5c" Nov 22 08:18:39 crc kubenswrapper[4789]: I1122 08:18:39.238103 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79d1ad09-f02f-421e-b1e5-10860722cc77-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-fzs5c\" (UID: \"79d1ad09-f02f-421e-b1e5-10860722cc77\") " pod="openstack/nova-cell0-cell-mapping-fzs5c" Nov 22 08:18:39 crc kubenswrapper[4789]: I1122 08:18:39.255906 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79d1ad09-f02f-421e-b1e5-10860722cc77-config-data\") pod \"nova-cell0-cell-mapping-fzs5c\" (UID: \"79d1ad09-f02f-421e-b1e5-10860722cc77\") " pod="openstack/nova-cell0-cell-mapping-fzs5c" Nov 22 08:18:39 crc kubenswrapper[4789]: I1122 08:18:39.273907 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/79d1ad09-f02f-421e-b1e5-10860722cc77-scripts\") pod \"nova-cell0-cell-mapping-fzs5c\" (UID: \"79d1ad09-f02f-421e-b1e5-10860722cc77\") " pod="openstack/nova-cell0-cell-mapping-fzs5c" Nov 22 08:18:39 crc kubenswrapper[4789]: I1122 08:18:39.273947 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79d1ad09-f02f-421e-b1e5-10860722cc77-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-fzs5c\" (UID: \"79d1ad09-f02f-421e-b1e5-10860722cc77\") " pod="openstack/nova-cell0-cell-mapping-fzs5c" Nov 22 08:18:39 crc kubenswrapper[4789]: I1122 08:18:39.305689 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 22 08:18:39 crc kubenswrapper[4789]: I1122 08:18:39.326990 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mb5qv\" (UniqueName: \"kubernetes.io/projected/79d1ad09-f02f-421e-b1e5-10860722cc77-kube-api-access-mb5qv\") pod \"nova-cell0-cell-mapping-fzs5c\" (UID: \"79d1ad09-f02f-421e-b1e5-10860722cc77\") " pod="openstack/nova-cell0-cell-mapping-fzs5c" Nov 22 08:18:39 crc kubenswrapper[4789]: I1122 08:18:39.329937 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 22 08:18:39 crc kubenswrapper[4789]: I1122 08:18:39.333623 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 22 08:18:39 crc kubenswrapper[4789]: I1122 08:18:39.387385 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 22 08:18:39 crc kubenswrapper[4789]: I1122 08:18:39.390009 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 22 08:18:39 crc kubenswrapper[4789]: I1122 08:18:39.393389 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 22 08:18:39 crc kubenswrapper[4789]: I1122 08:18:39.478401 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 22 08:18:39 crc kubenswrapper[4789]: I1122 08:18:39.506929 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ebf4ac2c-3bd1-474e-94ee-e707e215a5f3-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"ebf4ac2c-3bd1-474e-94ee-e707e215a5f3\") " pod="openstack/nova-metadata-0" Nov 22 08:18:39 crc kubenswrapper[4789]: I1122 08:18:39.507043 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e4eaef7-30a8-4a62-bb67-9aafc8f52b76-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"1e4eaef7-30a8-4a62-bb67-9aafc8f52b76\") " pod="openstack/nova-api-0" Nov 22 08:18:39 crc kubenswrapper[4789]: I1122 08:18:39.507077 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e4eaef7-30a8-4a62-bb67-9aafc8f52b76-config-data\") pod \"nova-api-0\" (UID: \"1e4eaef7-30a8-4a62-bb67-9aafc8f52b76\") " pod="openstack/nova-api-0" Nov 22 08:18:39 crc kubenswrapper[4789]: I1122 08:18:39.507852 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1e4eaef7-30a8-4a62-bb67-9aafc8f52b76-logs\") pod \"nova-api-0\" (UID: \"1e4eaef7-30a8-4a62-bb67-9aafc8f52b76\") " pod="openstack/nova-api-0" Nov 22 08:18:39 crc kubenswrapper[4789]: I1122 08:18:39.507892 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ebf4ac2c-3bd1-474e-94ee-e707e215a5f3-config-data\") pod \"nova-metadata-0\" (UID: \"ebf4ac2c-3bd1-474e-94ee-e707e215a5f3\") " pod="openstack/nova-metadata-0" Nov 22 08:18:39 crc kubenswrapper[4789]: I1122 08:18:39.507937 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lgcp8\" (UniqueName: \"kubernetes.io/projected/1e4eaef7-30a8-4a62-bb67-9aafc8f52b76-kube-api-access-lgcp8\") pod \"nova-api-0\" (UID: \"1e4eaef7-30a8-4a62-bb67-9aafc8f52b76\") " pod="openstack/nova-api-0" Nov 22 08:18:39 crc kubenswrapper[4789]: I1122 08:18:39.507966 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6rqgs\" (UniqueName: \"kubernetes.io/projected/ebf4ac2c-3bd1-474e-94ee-e707e215a5f3-kube-api-access-6rqgs\") pod \"nova-metadata-0\" (UID: \"ebf4ac2c-3bd1-474e-94ee-e707e215a5f3\") " pod="openstack/nova-metadata-0" Nov 22 08:18:39 crc kubenswrapper[4789]: I1122 08:18:39.508004 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ebf4ac2c-3bd1-474e-94ee-e707e215a5f3-logs\") pod \"nova-metadata-0\" (UID: \"ebf4ac2c-3bd1-474e-94ee-e707e215a5f3\") " pod="openstack/nova-metadata-0" Nov 22 08:18:39 crc kubenswrapper[4789]: I1122 08:18:39.543831 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 22 08:18:39 crc kubenswrapper[4789]: I1122 08:18:39.588821 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 22 08:18:39 crc kubenswrapper[4789]: I1122 08:18:39.590381 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 22 08:18:39 crc kubenswrapper[4789]: I1122 08:18:39.591351 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-fzs5c" Nov 22 08:18:39 crc kubenswrapper[4789]: I1122 08:18:39.598159 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 22 08:18:39 crc kubenswrapper[4789]: I1122 08:18:39.615230 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lgcp8\" (UniqueName: \"kubernetes.io/projected/1e4eaef7-30a8-4a62-bb67-9aafc8f52b76-kube-api-access-lgcp8\") pod \"nova-api-0\" (UID: \"1e4eaef7-30a8-4a62-bb67-9aafc8f52b76\") " pod="openstack/nova-api-0" Nov 22 08:18:39 crc kubenswrapper[4789]: I1122 08:18:39.615474 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6rqgs\" (UniqueName: \"kubernetes.io/projected/ebf4ac2c-3bd1-474e-94ee-e707e215a5f3-kube-api-access-6rqgs\") pod \"nova-metadata-0\" (UID: \"ebf4ac2c-3bd1-474e-94ee-e707e215a5f3\") " pod="openstack/nova-metadata-0" Nov 22 08:18:39 crc kubenswrapper[4789]: I1122 08:18:39.615703 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ebf4ac2c-3bd1-474e-94ee-e707e215a5f3-logs\") pod \"nova-metadata-0\" (UID: \"ebf4ac2c-3bd1-474e-94ee-e707e215a5f3\") " pod="openstack/nova-metadata-0" Nov 22 08:18:39 crc kubenswrapper[4789]: I1122 08:18:39.616276 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ebf4ac2c-3bd1-474e-94ee-e707e215a5f3-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"ebf4ac2c-3bd1-474e-94ee-e707e215a5f3\") " pod="openstack/nova-metadata-0" Nov 22 08:18:39 crc kubenswrapper[4789]: I1122 08:18:39.616313 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ebf4ac2c-3bd1-474e-94ee-e707e215a5f3-logs\") pod \"nova-metadata-0\" (UID: \"ebf4ac2c-3bd1-474e-94ee-e707e215a5f3\") " pod="openstack/nova-metadata-0" Nov 22 08:18:39 crc kubenswrapper[4789]: I1122 08:18:39.620961 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e4eaef7-30a8-4a62-bb67-9aafc8f52b76-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"1e4eaef7-30a8-4a62-bb67-9aafc8f52b76\") " pod="openstack/nova-api-0" Nov 22 08:18:39 crc kubenswrapper[4789]: I1122 08:18:39.621007 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e4eaef7-30a8-4a62-bb67-9aafc8f52b76-config-data\") pod \"nova-api-0\" (UID: \"1e4eaef7-30a8-4a62-bb67-9aafc8f52b76\") " pod="openstack/nova-api-0" Nov 22 08:18:39 crc kubenswrapper[4789]: I1122 08:18:39.621069 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1e4eaef7-30a8-4a62-bb67-9aafc8f52b76-logs\") pod \"nova-api-0\" (UID: \"1e4eaef7-30a8-4a62-bb67-9aafc8f52b76\") " pod="openstack/nova-api-0" Nov 22 08:18:39 crc kubenswrapper[4789]: I1122 08:18:39.621090 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ebf4ac2c-3bd1-474e-94ee-e707e215a5f3-config-data\") pod \"nova-metadata-0\" (UID: \"ebf4ac2c-3bd1-474e-94ee-e707e215a5f3\") " pod="openstack/nova-metadata-0" Nov 22 08:18:39 crc kubenswrapper[4789]: I1122 08:18:39.630969 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1e4eaef7-30a8-4a62-bb67-9aafc8f52b76-logs\") pod \"nova-api-0\" (UID: \"1e4eaef7-30a8-4a62-bb67-9aafc8f52b76\") " pod="openstack/nova-api-0" Nov 22 08:18:39 crc kubenswrapper[4789]: I1122 08:18:39.633610 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ebf4ac2c-3bd1-474e-94ee-e707e215a5f3-config-data\") pod \"nova-metadata-0\" (UID: \"ebf4ac2c-3bd1-474e-94ee-e707e215a5f3\") " pod="openstack/nova-metadata-0" Nov 22 08:18:39 crc kubenswrapper[4789]: I1122 08:18:39.634629 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e4eaef7-30a8-4a62-bb67-9aafc8f52b76-config-data\") pod \"nova-api-0\" (UID: \"1e4eaef7-30a8-4a62-bb67-9aafc8f52b76\") " pod="openstack/nova-api-0" Nov 22 08:18:39 crc kubenswrapper[4789]: I1122 08:18:39.640198 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ebf4ac2c-3bd1-474e-94ee-e707e215a5f3-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"ebf4ac2c-3bd1-474e-94ee-e707e215a5f3\") " pod="openstack/nova-metadata-0" Nov 22 08:18:39 crc kubenswrapper[4789]: I1122 08:18:39.641391 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e4eaef7-30a8-4a62-bb67-9aafc8f52b76-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"1e4eaef7-30a8-4a62-bb67-9aafc8f52b76\") " pod="openstack/nova-api-0" Nov 22 08:18:39 crc kubenswrapper[4789]: I1122 08:18:39.641854 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6rqgs\" (UniqueName: \"kubernetes.io/projected/ebf4ac2c-3bd1-474e-94ee-e707e215a5f3-kube-api-access-6rqgs\") pod \"nova-metadata-0\" (UID: \"ebf4ac2c-3bd1-474e-94ee-e707e215a5f3\") " pod="openstack/nova-metadata-0" Nov 22 08:18:39 crc kubenswrapper[4789]: I1122 08:18:39.647250 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lgcp8\" (UniqueName: \"kubernetes.io/projected/1e4eaef7-30a8-4a62-bb67-9aafc8f52b76-kube-api-access-lgcp8\") pod \"nova-api-0\" (UID: \"1e4eaef7-30a8-4a62-bb67-9aafc8f52b76\") " pod="openstack/nova-api-0" Nov 22 08:18:39 crc kubenswrapper[4789]: I1122 08:18:39.663685 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 22 08:18:39 crc kubenswrapper[4789]: I1122 08:18:39.696273 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-69494d9f89-z9hd6"] Nov 22 08:18:39 crc kubenswrapper[4789]: I1122 08:18:39.698511 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-69494d9f89-z9hd6" Nov 22 08:18:39 crc kubenswrapper[4789]: I1122 08:18:39.706990 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-69494d9f89-z9hd6"] Nov 22 08:18:39 crc kubenswrapper[4789]: I1122 08:18:39.719911 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 22 08:18:39 crc kubenswrapper[4789]: I1122 08:18:39.721479 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 22 08:18:39 crc kubenswrapper[4789]: I1122 08:18:39.723767 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pd4g6\" (UniqueName: \"kubernetes.io/projected/e012dd92-8842-4e20-beca-54115a7c933e-kube-api-access-pd4g6\") pod \"nova-cell1-novncproxy-0\" (UID: \"e012dd92-8842-4e20-beca-54115a7c933e\") " pod="openstack/nova-cell1-novncproxy-0" Nov 22 08:18:39 crc kubenswrapper[4789]: I1122 08:18:39.724186 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/75e7c40c-ba8c-45f6-b30c-f654f48956ca-ovsdbserver-nb\") pod \"dnsmasq-dns-69494d9f89-z9hd6\" (UID: \"75e7c40c-ba8c-45f6-b30c-f654f48956ca\") " pod="openstack/dnsmasq-dns-69494d9f89-z9hd6" Nov 22 08:18:39 crc kubenswrapper[4789]: I1122 08:18:39.724341 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Nov 22 08:18:39 crc kubenswrapper[4789]: I1122 08:18:39.725361 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cb3ad54-8a01-43df-b938-b6e68bcb704c-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"9cb3ad54-8a01-43df-b938-b6e68bcb704c\") " pod="openstack/nova-scheduler-0" Nov 22 08:18:39 crc kubenswrapper[4789]: I1122 08:18:39.725520 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lcnnm\" (UniqueName: \"kubernetes.io/projected/9cb3ad54-8a01-43df-b938-b6e68bcb704c-kube-api-access-lcnnm\") pod \"nova-scheduler-0\" (UID: \"9cb3ad54-8a01-43df-b938-b6e68bcb704c\") " pod="openstack/nova-scheduler-0" Nov 22 08:18:39 crc kubenswrapper[4789]: I1122 08:18:39.725578 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-57kgk\" (UniqueName: \"kubernetes.io/projected/75e7c40c-ba8c-45f6-b30c-f654f48956ca-kube-api-access-57kgk\") pod \"dnsmasq-dns-69494d9f89-z9hd6\" (UID: \"75e7c40c-ba8c-45f6-b30c-f654f48956ca\") " pod="openstack/dnsmasq-dns-69494d9f89-z9hd6" Nov 22 08:18:39 crc kubenswrapper[4789]: I1122 08:18:39.725680 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9cb3ad54-8a01-43df-b938-b6e68bcb704c-config-data\") pod \"nova-scheduler-0\" (UID: \"9cb3ad54-8a01-43df-b938-b6e68bcb704c\") " pod="openstack/nova-scheduler-0" Nov 22 08:18:39 crc kubenswrapper[4789]: I1122 08:18:39.725760 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e012dd92-8842-4e20-beca-54115a7c933e-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"e012dd92-8842-4e20-beca-54115a7c933e\") " pod="openstack/nova-cell1-novncproxy-0" Nov 22 08:18:39 crc kubenswrapper[4789]: I1122 08:18:39.725821 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/75e7c40c-ba8c-45f6-b30c-f654f48956ca-ovsdbserver-sb\") pod \"dnsmasq-dns-69494d9f89-z9hd6\" (UID: \"75e7c40c-ba8c-45f6-b30c-f654f48956ca\") " pod="openstack/dnsmasq-dns-69494d9f89-z9hd6" Nov 22 08:18:39 crc kubenswrapper[4789]: I1122 08:18:39.725873 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/75e7c40c-ba8c-45f6-b30c-f654f48956ca-dns-svc\") pod \"dnsmasq-dns-69494d9f89-z9hd6\" (UID: \"75e7c40c-ba8c-45f6-b30c-f654f48956ca\") " pod="openstack/dnsmasq-dns-69494d9f89-z9hd6" Nov 22 08:18:39 crc kubenswrapper[4789]: I1122 08:18:39.725903 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/75e7c40c-ba8c-45f6-b30c-f654f48956ca-config\") pod \"dnsmasq-dns-69494d9f89-z9hd6\" (UID: \"75e7c40c-ba8c-45f6-b30c-f654f48956ca\") " pod="openstack/dnsmasq-dns-69494d9f89-z9hd6" Nov 22 08:18:39 crc kubenswrapper[4789]: I1122 08:18:39.725937 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e012dd92-8842-4e20-beca-54115a7c933e-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"e012dd92-8842-4e20-beca-54115a7c933e\") " pod="openstack/nova-cell1-novncproxy-0" Nov 22 08:18:39 crc kubenswrapper[4789]: I1122 08:18:39.730683 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 22 08:18:39 crc kubenswrapper[4789]: I1122 08:18:39.752302 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 22 08:18:39 crc kubenswrapper[4789]: I1122 08:18:39.780287 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 22 08:18:39 crc kubenswrapper[4789]: I1122 08:18:39.829844 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9cb3ad54-8a01-43df-b938-b6e68bcb704c-config-data\") pod \"nova-scheduler-0\" (UID: \"9cb3ad54-8a01-43df-b938-b6e68bcb704c\") " pod="openstack/nova-scheduler-0" Nov 22 08:18:39 crc kubenswrapper[4789]: I1122 08:18:39.829933 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e012dd92-8842-4e20-beca-54115a7c933e-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"e012dd92-8842-4e20-beca-54115a7c933e\") " pod="openstack/nova-cell1-novncproxy-0" Nov 22 08:18:39 crc kubenswrapper[4789]: I1122 08:18:39.829963 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/75e7c40c-ba8c-45f6-b30c-f654f48956ca-ovsdbserver-sb\") pod \"dnsmasq-dns-69494d9f89-z9hd6\" (UID: \"75e7c40c-ba8c-45f6-b30c-f654f48956ca\") " pod="openstack/dnsmasq-dns-69494d9f89-z9hd6" Nov 22 08:18:39 crc kubenswrapper[4789]: I1122 08:18:39.830002 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/75e7c40c-ba8c-45f6-b30c-f654f48956ca-dns-svc\") pod \"dnsmasq-dns-69494d9f89-z9hd6\" (UID: \"75e7c40c-ba8c-45f6-b30c-f654f48956ca\") " pod="openstack/dnsmasq-dns-69494d9f89-z9hd6" Nov 22 08:18:39 crc kubenswrapper[4789]: I1122 08:18:39.830020 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/75e7c40c-ba8c-45f6-b30c-f654f48956ca-config\") pod \"dnsmasq-dns-69494d9f89-z9hd6\" (UID: \"75e7c40c-ba8c-45f6-b30c-f654f48956ca\") " pod="openstack/dnsmasq-dns-69494d9f89-z9hd6" Nov 22 08:18:39 crc kubenswrapper[4789]: I1122 08:18:39.830047 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e012dd92-8842-4e20-beca-54115a7c933e-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"e012dd92-8842-4e20-beca-54115a7c933e\") " pod="openstack/nova-cell1-novncproxy-0" Nov 22 08:18:39 crc kubenswrapper[4789]: I1122 08:18:39.830202 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pd4g6\" (UniqueName: \"kubernetes.io/projected/e012dd92-8842-4e20-beca-54115a7c933e-kube-api-access-pd4g6\") pod \"nova-cell1-novncproxy-0\" (UID: \"e012dd92-8842-4e20-beca-54115a7c933e\") " pod="openstack/nova-cell1-novncproxy-0" Nov 22 08:18:39 crc kubenswrapper[4789]: I1122 08:18:39.830220 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/75e7c40c-ba8c-45f6-b30c-f654f48956ca-ovsdbserver-nb\") pod \"dnsmasq-dns-69494d9f89-z9hd6\" (UID: \"75e7c40c-ba8c-45f6-b30c-f654f48956ca\") " pod="openstack/dnsmasq-dns-69494d9f89-z9hd6" Nov 22 08:18:39 crc kubenswrapper[4789]: I1122 08:18:39.830264 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cb3ad54-8a01-43df-b938-b6e68bcb704c-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"9cb3ad54-8a01-43df-b938-b6e68bcb704c\") " pod="openstack/nova-scheduler-0" Nov 22 08:18:39 crc kubenswrapper[4789]: I1122 08:18:39.830345 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lcnnm\" (UniqueName: \"kubernetes.io/projected/9cb3ad54-8a01-43df-b938-b6e68bcb704c-kube-api-access-lcnnm\") pod \"nova-scheduler-0\" (UID: \"9cb3ad54-8a01-43df-b938-b6e68bcb704c\") " pod="openstack/nova-scheduler-0" Nov 22 08:18:39 crc kubenswrapper[4789]: I1122 08:18:39.830375 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-57kgk\" (UniqueName: \"kubernetes.io/projected/75e7c40c-ba8c-45f6-b30c-f654f48956ca-kube-api-access-57kgk\") pod \"dnsmasq-dns-69494d9f89-z9hd6\" (UID: \"75e7c40c-ba8c-45f6-b30c-f654f48956ca\") " pod="openstack/dnsmasq-dns-69494d9f89-z9hd6" Nov 22 08:18:39 crc kubenswrapper[4789]: I1122 08:18:39.831972 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/75e7c40c-ba8c-45f6-b30c-f654f48956ca-config\") pod \"dnsmasq-dns-69494d9f89-z9hd6\" (UID: \"75e7c40c-ba8c-45f6-b30c-f654f48956ca\") " pod="openstack/dnsmasq-dns-69494d9f89-z9hd6" Nov 22 08:18:39 crc kubenswrapper[4789]: I1122 08:18:39.833065 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/75e7c40c-ba8c-45f6-b30c-f654f48956ca-ovsdbserver-sb\") pod \"dnsmasq-dns-69494d9f89-z9hd6\" (UID: \"75e7c40c-ba8c-45f6-b30c-f654f48956ca\") " pod="openstack/dnsmasq-dns-69494d9f89-z9hd6" Nov 22 08:18:39 crc kubenswrapper[4789]: I1122 08:18:39.833360 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/75e7c40c-ba8c-45f6-b30c-f654f48956ca-dns-svc\") pod \"dnsmasq-dns-69494d9f89-z9hd6\" (UID: \"75e7c40c-ba8c-45f6-b30c-f654f48956ca\") " pod="openstack/dnsmasq-dns-69494d9f89-z9hd6" Nov 22 08:18:39 crc kubenswrapper[4789]: I1122 08:18:39.837812 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/75e7c40c-ba8c-45f6-b30c-f654f48956ca-ovsdbserver-nb\") pod \"dnsmasq-dns-69494d9f89-z9hd6\" (UID: \"75e7c40c-ba8c-45f6-b30c-f654f48956ca\") " pod="openstack/dnsmasq-dns-69494d9f89-z9hd6" Nov 22 08:18:39 crc kubenswrapper[4789]: I1122 08:18:39.841968 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cb3ad54-8a01-43df-b938-b6e68bcb704c-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"9cb3ad54-8a01-43df-b938-b6e68bcb704c\") " pod="openstack/nova-scheduler-0" Nov 22 08:18:39 crc kubenswrapper[4789]: I1122 08:18:39.843263 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e012dd92-8842-4e20-beca-54115a7c933e-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"e012dd92-8842-4e20-beca-54115a7c933e\") " pod="openstack/nova-cell1-novncproxy-0" Nov 22 08:18:39 crc kubenswrapper[4789]: I1122 08:18:39.844506 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e012dd92-8842-4e20-beca-54115a7c933e-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"e012dd92-8842-4e20-beca-54115a7c933e\") " pod="openstack/nova-cell1-novncproxy-0" Nov 22 08:18:39 crc kubenswrapper[4789]: I1122 08:18:39.849640 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-57kgk\" (UniqueName: \"kubernetes.io/projected/75e7c40c-ba8c-45f6-b30c-f654f48956ca-kube-api-access-57kgk\") pod \"dnsmasq-dns-69494d9f89-z9hd6\" (UID: \"75e7c40c-ba8c-45f6-b30c-f654f48956ca\") " pod="openstack/dnsmasq-dns-69494d9f89-z9hd6" Nov 22 08:18:39 crc kubenswrapper[4789]: I1122 08:18:39.851993 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9cb3ad54-8a01-43df-b938-b6e68bcb704c-config-data\") pod \"nova-scheduler-0\" (UID: \"9cb3ad54-8a01-43df-b938-b6e68bcb704c\") " pod="openstack/nova-scheduler-0" Nov 22 08:18:39 crc kubenswrapper[4789]: I1122 08:18:39.855115 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lcnnm\" (UniqueName: \"kubernetes.io/projected/9cb3ad54-8a01-43df-b938-b6e68bcb704c-kube-api-access-lcnnm\") pod \"nova-scheduler-0\" (UID: \"9cb3ad54-8a01-43df-b938-b6e68bcb704c\") " pod="openstack/nova-scheduler-0" Nov 22 08:18:39 crc kubenswrapper[4789]: I1122 08:18:39.855661 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pd4g6\" (UniqueName: \"kubernetes.io/projected/e012dd92-8842-4e20-beca-54115a7c933e-kube-api-access-pd4g6\") pod \"nova-cell1-novncproxy-0\" (UID: \"e012dd92-8842-4e20-beca-54115a7c933e\") " pod="openstack/nova-cell1-novncproxy-0" Nov 22 08:18:40 crc kubenswrapper[4789]: I1122 08:18:40.025472 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 22 08:18:40 crc kubenswrapper[4789]: I1122 08:18:40.053295 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-69494d9f89-z9hd6" Nov 22 08:18:40 crc kubenswrapper[4789]: I1122 08:18:40.061265 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 22 08:18:40 crc kubenswrapper[4789]: I1122 08:18:40.126413 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-fzs5c"] Nov 22 08:18:40 crc kubenswrapper[4789]: I1122 08:18:40.155595 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-skqfc"] Nov 22 08:18:40 crc kubenswrapper[4789]: I1122 08:18:40.159563 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-skqfc" Nov 22 08:18:40 crc kubenswrapper[4789]: I1122 08:18:40.164872 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Nov 22 08:18:40 crc kubenswrapper[4789]: I1122 08:18:40.165125 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Nov 22 08:18:40 crc kubenswrapper[4789]: I1122 08:18:40.170880 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-skqfc"] Nov 22 08:18:40 crc kubenswrapper[4789]: I1122 08:18:40.247650 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 22 08:18:40 crc kubenswrapper[4789]: I1122 08:18:40.250142 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4270eeb-3da5-464a-806a-358dbc4bd6ba-config-data\") pod \"nova-cell1-conductor-db-sync-skqfc\" (UID: \"d4270eeb-3da5-464a-806a-358dbc4bd6ba\") " pod="openstack/nova-cell1-conductor-db-sync-skqfc" Nov 22 08:18:40 crc kubenswrapper[4789]: I1122 08:18:40.250566 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4270eeb-3da5-464a-806a-358dbc4bd6ba-scripts\") pod \"nova-cell1-conductor-db-sync-skqfc\" (UID: \"d4270eeb-3da5-464a-806a-358dbc4bd6ba\") " pod="openstack/nova-cell1-conductor-db-sync-skqfc" Nov 22 08:18:40 crc kubenswrapper[4789]: I1122 08:18:40.250722 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4270eeb-3da5-464a-806a-358dbc4bd6ba-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-skqfc\" (UID: \"d4270eeb-3da5-464a-806a-358dbc4bd6ba\") " pod="openstack/nova-cell1-conductor-db-sync-skqfc" Nov 22 08:18:40 crc kubenswrapper[4789]: I1122 08:18:40.251330 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5ww4b\" (UniqueName: \"kubernetes.io/projected/d4270eeb-3da5-464a-806a-358dbc4bd6ba-kube-api-access-5ww4b\") pod \"nova-cell1-conductor-db-sync-skqfc\" (UID: \"d4270eeb-3da5-464a-806a-358dbc4bd6ba\") " pod="openstack/nova-cell1-conductor-db-sync-skqfc" Nov 22 08:18:40 crc kubenswrapper[4789]: I1122 08:18:40.252106 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-fzs5c" event={"ID":"79d1ad09-f02f-421e-b1e5-10860722cc77","Type":"ContainerStarted","Data":"e258e3652b4b1c90544937b7e2c9af190d01701361b8142f0630f586d2139c1a"} Nov 22 08:18:40 crc kubenswrapper[4789]: W1122 08:18:40.273324 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podebf4ac2c_3bd1_474e_94ee_e707e215a5f3.slice/crio-3fc32e52e36dd429f27a156b2223397ea30be918754d9e601ef9a702a7ab77ab WatchSource:0}: Error finding container 3fc32e52e36dd429f27a156b2223397ea30be918754d9e601ef9a702a7ab77ab: Status 404 returned error can't find the container with id 3fc32e52e36dd429f27a156b2223397ea30be918754d9e601ef9a702a7ab77ab Nov 22 08:18:40 crc kubenswrapper[4789]: I1122 08:18:40.316040 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 22 08:18:40 crc kubenswrapper[4789]: I1122 08:18:40.353116 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4270eeb-3da5-464a-806a-358dbc4bd6ba-config-data\") pod \"nova-cell1-conductor-db-sync-skqfc\" (UID: \"d4270eeb-3da5-464a-806a-358dbc4bd6ba\") " pod="openstack/nova-cell1-conductor-db-sync-skqfc" Nov 22 08:18:40 crc kubenswrapper[4789]: I1122 08:18:40.353239 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4270eeb-3da5-464a-806a-358dbc4bd6ba-scripts\") pod \"nova-cell1-conductor-db-sync-skqfc\" (UID: \"d4270eeb-3da5-464a-806a-358dbc4bd6ba\") " pod="openstack/nova-cell1-conductor-db-sync-skqfc" Nov 22 08:18:40 crc kubenswrapper[4789]: I1122 08:18:40.353268 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4270eeb-3da5-464a-806a-358dbc4bd6ba-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-skqfc\" (UID: \"d4270eeb-3da5-464a-806a-358dbc4bd6ba\") " pod="openstack/nova-cell1-conductor-db-sync-skqfc" Nov 22 08:18:40 crc kubenswrapper[4789]: I1122 08:18:40.353369 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5ww4b\" (UniqueName: \"kubernetes.io/projected/d4270eeb-3da5-464a-806a-358dbc4bd6ba-kube-api-access-5ww4b\") pod \"nova-cell1-conductor-db-sync-skqfc\" (UID: \"d4270eeb-3da5-464a-806a-358dbc4bd6ba\") " pod="openstack/nova-cell1-conductor-db-sync-skqfc" Nov 22 08:18:40 crc kubenswrapper[4789]: I1122 08:18:40.357925 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4270eeb-3da5-464a-806a-358dbc4bd6ba-config-data\") pod \"nova-cell1-conductor-db-sync-skqfc\" (UID: \"d4270eeb-3da5-464a-806a-358dbc4bd6ba\") " pod="openstack/nova-cell1-conductor-db-sync-skqfc" Nov 22 08:18:40 crc kubenswrapper[4789]: I1122 08:18:40.358239 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4270eeb-3da5-464a-806a-358dbc4bd6ba-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-skqfc\" (UID: \"d4270eeb-3da5-464a-806a-358dbc4bd6ba\") " pod="openstack/nova-cell1-conductor-db-sync-skqfc" Nov 22 08:18:40 crc kubenswrapper[4789]: I1122 08:18:40.358709 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4270eeb-3da5-464a-806a-358dbc4bd6ba-scripts\") pod \"nova-cell1-conductor-db-sync-skqfc\" (UID: \"d4270eeb-3da5-464a-806a-358dbc4bd6ba\") " pod="openstack/nova-cell1-conductor-db-sync-skqfc" Nov 22 08:18:40 crc kubenswrapper[4789]: I1122 08:18:40.371707 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5ww4b\" (UniqueName: \"kubernetes.io/projected/d4270eeb-3da5-464a-806a-358dbc4bd6ba-kube-api-access-5ww4b\") pod \"nova-cell1-conductor-db-sync-skqfc\" (UID: \"d4270eeb-3da5-464a-806a-358dbc4bd6ba\") " pod="openstack/nova-cell1-conductor-db-sync-skqfc" Nov 22 08:18:40 crc kubenswrapper[4789]: I1122 08:18:40.488222 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-skqfc" Nov 22 08:18:40 crc kubenswrapper[4789]: I1122 08:18:40.572634 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 22 08:18:40 crc kubenswrapper[4789]: W1122 08:18:40.576280 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9cb3ad54_8a01_43df_b938_b6e68bcb704c.slice/crio-d04c84c773790abb1711752e6d6fbb9acf91aff0aaf6d4dd561fc9e8bbae0f3a WatchSource:0}: Error finding container d04c84c773790abb1711752e6d6fbb9acf91aff0aaf6d4dd561fc9e8bbae0f3a: Status 404 returned error can't find the container with id d04c84c773790abb1711752e6d6fbb9acf91aff0aaf6d4dd561fc9e8bbae0f3a Nov 22 08:18:40 crc kubenswrapper[4789]: I1122 08:18:40.667434 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-69494d9f89-z9hd6"] Nov 22 08:18:40 crc kubenswrapper[4789]: I1122 08:18:40.678118 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 22 08:18:40 crc kubenswrapper[4789]: I1122 08:18:40.955783 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-skqfc"] Nov 22 08:18:40 crc kubenswrapper[4789]: W1122 08:18:40.972293 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd4270eeb_3da5_464a_806a_358dbc4bd6ba.slice/crio-b28cb0f340702a695e4339755c1b4af7f18c0a5874c60254b801ea4d53670d0b WatchSource:0}: Error finding container b28cb0f340702a695e4339755c1b4af7f18c0a5874c60254b801ea4d53670d0b: Status 404 returned error can't find the container with id b28cb0f340702a695e4339755c1b4af7f18c0a5874c60254b801ea4d53670d0b Nov 22 08:18:41 crc kubenswrapper[4789]: I1122 08:18:41.264760 4789 generic.go:334] "Generic (PLEG): container finished" podID="75e7c40c-ba8c-45f6-b30c-f654f48956ca" containerID="6783b7130455f20a22183eb8df83c513dd7d6d2621b36957429a75060739d0fb" exitCode=0 Nov 22 08:18:41 crc kubenswrapper[4789]: I1122 08:18:41.264832 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-69494d9f89-z9hd6" event={"ID":"75e7c40c-ba8c-45f6-b30c-f654f48956ca","Type":"ContainerDied","Data":"6783b7130455f20a22183eb8df83c513dd7d6d2621b36957429a75060739d0fb"} Nov 22 08:18:41 crc kubenswrapper[4789]: I1122 08:18:41.264901 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-69494d9f89-z9hd6" event={"ID":"75e7c40c-ba8c-45f6-b30c-f654f48956ca","Type":"ContainerStarted","Data":"4c282f547bcbdad803bad222cabb5fcdb560aa8c5982c873cdb4469fca1bc3a9"} Nov 22 08:18:41 crc kubenswrapper[4789]: I1122 08:18:41.275510 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-fzs5c" event={"ID":"79d1ad09-f02f-421e-b1e5-10860722cc77","Type":"ContainerStarted","Data":"1d99062b301747b7ae1bd0307943d6d1d286baa1210d5067e5b0ada008972cf3"} Nov 22 08:18:41 crc kubenswrapper[4789]: I1122 08:18:41.288692 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"e012dd92-8842-4e20-beca-54115a7c933e","Type":"ContainerStarted","Data":"0d170139d4b7ce975ed2ac1b8cbe87dfef8512dfae7c681824d624cca08b758b"} Nov 22 08:18:41 crc kubenswrapper[4789]: I1122 08:18:41.306878 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1e4eaef7-30a8-4a62-bb67-9aafc8f52b76","Type":"ContainerStarted","Data":"5ffc2d1e0ec6eeaff9400eddc9910feb0d5ef3b342b482a3356811c3dbbd55c6"} Nov 22 08:18:41 crc kubenswrapper[4789]: I1122 08:18:41.313505 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-skqfc" event={"ID":"d4270eeb-3da5-464a-806a-358dbc4bd6ba","Type":"ContainerStarted","Data":"7128ef0b5659bd07e32c2171483c5a7e4125643355ad974d47dde4aa001b1fec"} Nov 22 08:18:41 crc kubenswrapper[4789]: I1122 08:18:41.313546 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-skqfc" event={"ID":"d4270eeb-3da5-464a-806a-358dbc4bd6ba","Type":"ContainerStarted","Data":"b28cb0f340702a695e4339755c1b4af7f18c0a5874c60254b801ea4d53670d0b"} Nov 22 08:18:41 crc kubenswrapper[4789]: I1122 08:18:41.316549 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"9cb3ad54-8a01-43df-b938-b6e68bcb704c","Type":"ContainerStarted","Data":"d04c84c773790abb1711752e6d6fbb9acf91aff0aaf6d4dd561fc9e8bbae0f3a"} Nov 22 08:18:41 crc kubenswrapper[4789]: I1122 08:18:41.318667 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ebf4ac2c-3bd1-474e-94ee-e707e215a5f3","Type":"ContainerStarted","Data":"3fc32e52e36dd429f27a156b2223397ea30be918754d9e601ef9a702a7ab77ab"} Nov 22 08:18:41 crc kubenswrapper[4789]: I1122 08:18:41.322825 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-fzs5c" podStartSLOduration=3.322428728 podStartE2EDuration="3.322428728s" podCreationTimestamp="2025-11-22 08:18:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:18:41.307160411 +0000 UTC m=+1335.541560684" watchObservedRunningTime="2025-11-22 08:18:41.322428728 +0000 UTC m=+1335.556829001" Nov 22 08:18:41 crc kubenswrapper[4789]: I1122 08:18:41.334160 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-skqfc" podStartSLOduration=1.334144089 podStartE2EDuration="1.334144089s" podCreationTimestamp="2025-11-22 08:18:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:18:41.330583741 +0000 UTC m=+1335.564984014" watchObservedRunningTime="2025-11-22 08:18:41.334144089 +0000 UTC m=+1335.568544362" Nov 22 08:18:42 crc kubenswrapper[4789]: I1122 08:18:42.328643 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-69494d9f89-z9hd6" event={"ID":"75e7c40c-ba8c-45f6-b30c-f654f48956ca","Type":"ContainerStarted","Data":"209550f4b6c9910b78a3c4910da280b88da86bda2367421bdb30ba3154833e07"} Nov 22 08:18:42 crc kubenswrapper[4789]: I1122 08:18:42.347610 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-69494d9f89-z9hd6" podStartSLOduration=3.347591412 podStartE2EDuration="3.347591412s" podCreationTimestamp="2025-11-22 08:18:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:18:42.34678105 +0000 UTC m=+1336.581181343" watchObservedRunningTime="2025-11-22 08:18:42.347591412 +0000 UTC m=+1336.581991685" Nov 22 08:18:43 crc kubenswrapper[4789]: I1122 08:18:43.042344 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 22 08:18:43 crc kubenswrapper[4789]: I1122 08:18:43.051423 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 22 08:18:43 crc kubenswrapper[4789]: I1122 08:18:43.337637 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-69494d9f89-z9hd6" Nov 22 08:18:45 crc kubenswrapper[4789]: I1122 08:18:45.373360 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"9cb3ad54-8a01-43df-b938-b6e68bcb704c","Type":"ContainerStarted","Data":"4bc244b6de6b03bed55cd48a85231ca6f96de9810592b5ee76f208984a87adc0"} Nov 22 08:18:45 crc kubenswrapper[4789]: I1122 08:18:45.375479 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ebf4ac2c-3bd1-474e-94ee-e707e215a5f3","Type":"ContainerStarted","Data":"744dca789fc995b4de5c6d2192de27669bffceb0b684c0ebd2d95ff48881db10"} Nov 22 08:18:45 crc kubenswrapper[4789]: I1122 08:18:45.375630 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ebf4ac2c-3bd1-474e-94ee-e707e215a5f3","Type":"ContainerStarted","Data":"352c28c526ea7c4fb9b0d9a8f43aa191072de013667fe06518534c9c7609cace"} Nov 22 08:18:45 crc kubenswrapper[4789]: I1122 08:18:45.375525 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="ebf4ac2c-3bd1-474e-94ee-e707e215a5f3" containerName="nova-metadata-log" containerID="cri-o://352c28c526ea7c4fb9b0d9a8f43aa191072de013667fe06518534c9c7609cace" gracePeriod=30 Nov 22 08:18:45 crc kubenswrapper[4789]: I1122 08:18:45.375957 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="ebf4ac2c-3bd1-474e-94ee-e707e215a5f3" containerName="nova-metadata-metadata" containerID="cri-o://744dca789fc995b4de5c6d2192de27669bffceb0b684c0ebd2d95ff48881db10" gracePeriod=30 Nov 22 08:18:45 crc kubenswrapper[4789]: I1122 08:18:45.379377 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="e012dd92-8842-4e20-beca-54115a7c933e" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://b84bea1d6e8d82a040a02e9f4861206802c1408e3e270e86181e6fa19997b2a4" gracePeriod=30 Nov 22 08:18:45 crc kubenswrapper[4789]: I1122 08:18:45.379504 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"e012dd92-8842-4e20-beca-54115a7c933e","Type":"ContainerStarted","Data":"b84bea1d6e8d82a040a02e9f4861206802c1408e3e270e86181e6fa19997b2a4"} Nov 22 08:18:45 crc kubenswrapper[4789]: I1122 08:18:45.390335 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1e4eaef7-30a8-4a62-bb67-9aafc8f52b76","Type":"ContainerStarted","Data":"3cb2ac3c4243a5a7bc7f0d9e9e4c7f1a499314f692dbb42de7c8edaf4c703b4f"} Nov 22 08:18:45 crc kubenswrapper[4789]: I1122 08:18:45.390388 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1e4eaef7-30a8-4a62-bb67-9aafc8f52b76","Type":"ContainerStarted","Data":"6f051cc0e759f5f1a60b82855e4f3e14affc615ae4b07071c76bea4e60c971c2"} Nov 22 08:18:45 crc kubenswrapper[4789]: I1122 08:18:45.398596 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.85752368 podStartE2EDuration="6.398579493s" podCreationTimestamp="2025-11-22 08:18:39 +0000 UTC" firstStartedPulling="2025-11-22 08:18:40.578786912 +0000 UTC m=+1334.813187185" lastFinishedPulling="2025-11-22 08:18:44.119842725 +0000 UTC m=+1338.354242998" observedRunningTime="2025-11-22 08:18:45.389380022 +0000 UTC m=+1339.623780295" watchObservedRunningTime="2025-11-22 08:18:45.398579493 +0000 UTC m=+1339.632979766" Nov 22 08:18:45 crc kubenswrapper[4789]: I1122 08:18:45.412056 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.571102049 podStartE2EDuration="6.41203846s" podCreationTimestamp="2025-11-22 08:18:39 +0000 UTC" firstStartedPulling="2025-11-22 08:18:40.277132516 +0000 UTC m=+1334.511532789" lastFinishedPulling="2025-11-22 08:18:44.118068927 +0000 UTC m=+1338.352469200" observedRunningTime="2025-11-22 08:18:45.410309553 +0000 UTC m=+1339.644709836" watchObservedRunningTime="2025-11-22 08:18:45.41203846 +0000 UTC m=+1339.646438733" Nov 22 08:18:45 crc kubenswrapper[4789]: I1122 08:18:45.426694 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.979386298 podStartE2EDuration="6.42667776s" podCreationTimestamp="2025-11-22 08:18:39 +0000 UTC" firstStartedPulling="2025-11-22 08:18:40.672563123 +0000 UTC m=+1334.906963396" lastFinishedPulling="2025-11-22 08:18:44.119854585 +0000 UTC m=+1338.354254858" observedRunningTime="2025-11-22 08:18:45.426247838 +0000 UTC m=+1339.660648121" watchObservedRunningTime="2025-11-22 08:18:45.42667776 +0000 UTC m=+1339.661078033" Nov 22 08:18:45 crc kubenswrapper[4789]: I1122 08:18:45.446802 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.660847399 podStartE2EDuration="6.446785859s" podCreationTimestamp="2025-11-22 08:18:39 +0000 UTC" firstStartedPulling="2025-11-22 08:18:40.321391933 +0000 UTC m=+1334.555792206" lastFinishedPulling="2025-11-22 08:18:44.107330393 +0000 UTC m=+1338.341730666" observedRunningTime="2025-11-22 08:18:45.441987848 +0000 UTC m=+1339.676388121" watchObservedRunningTime="2025-11-22 08:18:45.446785859 +0000 UTC m=+1339.681186122" Nov 22 08:18:45 crc kubenswrapper[4789]: I1122 08:18:45.957258 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 22 08:18:46 crc kubenswrapper[4789]: I1122 08:18:46.061787 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ebf4ac2c-3bd1-474e-94ee-e707e215a5f3-logs\") pod \"ebf4ac2c-3bd1-474e-94ee-e707e215a5f3\" (UID: \"ebf4ac2c-3bd1-474e-94ee-e707e215a5f3\") " Nov 22 08:18:46 crc kubenswrapper[4789]: I1122 08:18:46.061886 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ebf4ac2c-3bd1-474e-94ee-e707e215a5f3-config-data\") pod \"ebf4ac2c-3bd1-474e-94ee-e707e215a5f3\" (UID: \"ebf4ac2c-3bd1-474e-94ee-e707e215a5f3\") " Nov 22 08:18:46 crc kubenswrapper[4789]: I1122 08:18:46.061959 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6rqgs\" (UniqueName: \"kubernetes.io/projected/ebf4ac2c-3bd1-474e-94ee-e707e215a5f3-kube-api-access-6rqgs\") pod \"ebf4ac2c-3bd1-474e-94ee-e707e215a5f3\" (UID: \"ebf4ac2c-3bd1-474e-94ee-e707e215a5f3\") " Nov 22 08:18:46 crc kubenswrapper[4789]: I1122 08:18:46.062025 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ebf4ac2c-3bd1-474e-94ee-e707e215a5f3-combined-ca-bundle\") pod \"ebf4ac2c-3bd1-474e-94ee-e707e215a5f3\" (UID: \"ebf4ac2c-3bd1-474e-94ee-e707e215a5f3\") " Nov 22 08:18:46 crc kubenswrapper[4789]: I1122 08:18:46.062361 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ebf4ac2c-3bd1-474e-94ee-e707e215a5f3-logs" (OuterVolumeSpecName: "logs") pod "ebf4ac2c-3bd1-474e-94ee-e707e215a5f3" (UID: "ebf4ac2c-3bd1-474e-94ee-e707e215a5f3"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:18:46 crc kubenswrapper[4789]: I1122 08:18:46.063475 4789 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ebf4ac2c-3bd1-474e-94ee-e707e215a5f3-logs\") on node \"crc\" DevicePath \"\"" Nov 22 08:18:46 crc kubenswrapper[4789]: I1122 08:18:46.066696 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ebf4ac2c-3bd1-474e-94ee-e707e215a5f3-kube-api-access-6rqgs" (OuterVolumeSpecName: "kube-api-access-6rqgs") pod "ebf4ac2c-3bd1-474e-94ee-e707e215a5f3" (UID: "ebf4ac2c-3bd1-474e-94ee-e707e215a5f3"). InnerVolumeSpecName "kube-api-access-6rqgs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:18:46 crc kubenswrapper[4789]: I1122 08:18:46.091090 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ebf4ac2c-3bd1-474e-94ee-e707e215a5f3-config-data" (OuterVolumeSpecName: "config-data") pod "ebf4ac2c-3bd1-474e-94ee-e707e215a5f3" (UID: "ebf4ac2c-3bd1-474e-94ee-e707e215a5f3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:18:46 crc kubenswrapper[4789]: I1122 08:18:46.091167 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ebf4ac2c-3bd1-474e-94ee-e707e215a5f3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ebf4ac2c-3bd1-474e-94ee-e707e215a5f3" (UID: "ebf4ac2c-3bd1-474e-94ee-e707e215a5f3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:18:46 crc kubenswrapper[4789]: I1122 08:18:46.164819 4789 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ebf4ac2c-3bd1-474e-94ee-e707e215a5f3-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:18:46 crc kubenswrapper[4789]: I1122 08:18:46.165084 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6rqgs\" (UniqueName: \"kubernetes.io/projected/ebf4ac2c-3bd1-474e-94ee-e707e215a5f3-kube-api-access-6rqgs\") on node \"crc\" DevicePath \"\"" Nov 22 08:18:46 crc kubenswrapper[4789]: I1122 08:18:46.165097 4789 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ebf4ac2c-3bd1-474e-94ee-e707e215a5f3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:18:46 crc kubenswrapper[4789]: I1122 08:18:46.400610 4789 generic.go:334] "Generic (PLEG): container finished" podID="ebf4ac2c-3bd1-474e-94ee-e707e215a5f3" containerID="744dca789fc995b4de5c6d2192de27669bffceb0b684c0ebd2d95ff48881db10" exitCode=0 Nov 22 08:18:46 crc kubenswrapper[4789]: I1122 08:18:46.400668 4789 generic.go:334] "Generic (PLEG): container finished" podID="ebf4ac2c-3bd1-474e-94ee-e707e215a5f3" containerID="352c28c526ea7c4fb9b0d9a8f43aa191072de013667fe06518534c9c7609cace" exitCode=143 Nov 22 08:18:46 crc kubenswrapper[4789]: I1122 08:18:46.400695 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 22 08:18:46 crc kubenswrapper[4789]: I1122 08:18:46.400730 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ebf4ac2c-3bd1-474e-94ee-e707e215a5f3","Type":"ContainerDied","Data":"744dca789fc995b4de5c6d2192de27669bffceb0b684c0ebd2d95ff48881db10"} Nov 22 08:18:46 crc kubenswrapper[4789]: I1122 08:18:46.400794 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ebf4ac2c-3bd1-474e-94ee-e707e215a5f3","Type":"ContainerDied","Data":"352c28c526ea7c4fb9b0d9a8f43aa191072de013667fe06518534c9c7609cace"} Nov 22 08:18:46 crc kubenswrapper[4789]: I1122 08:18:46.400807 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ebf4ac2c-3bd1-474e-94ee-e707e215a5f3","Type":"ContainerDied","Data":"3fc32e52e36dd429f27a156b2223397ea30be918754d9e601ef9a702a7ab77ab"} Nov 22 08:18:46 crc kubenswrapper[4789]: I1122 08:18:46.400837 4789 scope.go:117] "RemoveContainer" containerID="744dca789fc995b4de5c6d2192de27669bffceb0b684c0ebd2d95ff48881db10" Nov 22 08:18:46 crc kubenswrapper[4789]: I1122 08:18:46.438553 4789 scope.go:117] "RemoveContainer" containerID="352c28c526ea7c4fb9b0d9a8f43aa191072de013667fe06518534c9c7609cace" Nov 22 08:18:46 crc kubenswrapper[4789]: I1122 08:18:46.445252 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 22 08:18:46 crc kubenswrapper[4789]: I1122 08:18:46.454297 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 22 08:18:46 crc kubenswrapper[4789]: I1122 08:18:46.467964 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 22 08:18:46 crc kubenswrapper[4789]: E1122 08:18:46.468738 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebf4ac2c-3bd1-474e-94ee-e707e215a5f3" containerName="nova-metadata-log" Nov 22 08:18:46 crc kubenswrapper[4789]: I1122 08:18:46.468838 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebf4ac2c-3bd1-474e-94ee-e707e215a5f3" containerName="nova-metadata-log" Nov 22 08:18:46 crc kubenswrapper[4789]: E1122 08:18:46.468915 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebf4ac2c-3bd1-474e-94ee-e707e215a5f3" containerName="nova-metadata-metadata" Nov 22 08:18:46 crc kubenswrapper[4789]: I1122 08:18:46.468977 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebf4ac2c-3bd1-474e-94ee-e707e215a5f3" containerName="nova-metadata-metadata" Nov 22 08:18:46 crc kubenswrapper[4789]: I1122 08:18:46.469195 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="ebf4ac2c-3bd1-474e-94ee-e707e215a5f3" containerName="nova-metadata-log" Nov 22 08:18:46 crc kubenswrapper[4789]: I1122 08:18:46.469269 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="ebf4ac2c-3bd1-474e-94ee-e707e215a5f3" containerName="nova-metadata-metadata" Nov 22 08:18:46 crc kubenswrapper[4789]: I1122 08:18:46.470196 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 22 08:18:46 crc kubenswrapper[4789]: I1122 08:18:46.473467 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 22 08:18:46 crc kubenswrapper[4789]: I1122 08:18:46.473676 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 22 08:18:46 crc kubenswrapper[4789]: I1122 08:18:46.479432 4789 scope.go:117] "RemoveContainer" containerID="744dca789fc995b4de5c6d2192de27669bffceb0b684c0ebd2d95ff48881db10" Nov 22 08:18:46 crc kubenswrapper[4789]: I1122 08:18:46.480934 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 22 08:18:46 crc kubenswrapper[4789]: E1122 08:18:46.481780 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"744dca789fc995b4de5c6d2192de27669bffceb0b684c0ebd2d95ff48881db10\": container with ID starting with 744dca789fc995b4de5c6d2192de27669bffceb0b684c0ebd2d95ff48881db10 not found: ID does not exist" containerID="744dca789fc995b4de5c6d2192de27669bffceb0b684c0ebd2d95ff48881db10" Nov 22 08:18:46 crc kubenswrapper[4789]: I1122 08:18:46.481884 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"744dca789fc995b4de5c6d2192de27669bffceb0b684c0ebd2d95ff48881db10"} err="failed to get container status \"744dca789fc995b4de5c6d2192de27669bffceb0b684c0ebd2d95ff48881db10\": rpc error: code = NotFound desc = could not find container \"744dca789fc995b4de5c6d2192de27669bffceb0b684c0ebd2d95ff48881db10\": container with ID starting with 744dca789fc995b4de5c6d2192de27669bffceb0b684c0ebd2d95ff48881db10 not found: ID does not exist" Nov 22 08:18:46 crc kubenswrapper[4789]: I1122 08:18:46.481987 4789 scope.go:117] "RemoveContainer" containerID="352c28c526ea7c4fb9b0d9a8f43aa191072de013667fe06518534c9c7609cace" Nov 22 08:18:46 crc kubenswrapper[4789]: E1122 08:18:46.483108 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"352c28c526ea7c4fb9b0d9a8f43aa191072de013667fe06518534c9c7609cace\": container with ID starting with 352c28c526ea7c4fb9b0d9a8f43aa191072de013667fe06518534c9c7609cace not found: ID does not exist" containerID="352c28c526ea7c4fb9b0d9a8f43aa191072de013667fe06518534c9c7609cace" Nov 22 08:18:46 crc kubenswrapper[4789]: I1122 08:18:46.483151 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"352c28c526ea7c4fb9b0d9a8f43aa191072de013667fe06518534c9c7609cace"} err="failed to get container status \"352c28c526ea7c4fb9b0d9a8f43aa191072de013667fe06518534c9c7609cace\": rpc error: code = NotFound desc = could not find container \"352c28c526ea7c4fb9b0d9a8f43aa191072de013667fe06518534c9c7609cace\": container with ID starting with 352c28c526ea7c4fb9b0d9a8f43aa191072de013667fe06518534c9c7609cace not found: ID does not exist" Nov 22 08:18:46 crc kubenswrapper[4789]: I1122 08:18:46.483183 4789 scope.go:117] "RemoveContainer" containerID="744dca789fc995b4de5c6d2192de27669bffceb0b684c0ebd2d95ff48881db10" Nov 22 08:18:46 crc kubenswrapper[4789]: I1122 08:18:46.483577 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"744dca789fc995b4de5c6d2192de27669bffceb0b684c0ebd2d95ff48881db10"} err="failed to get container status \"744dca789fc995b4de5c6d2192de27669bffceb0b684c0ebd2d95ff48881db10\": rpc error: code = NotFound desc = could not find container \"744dca789fc995b4de5c6d2192de27669bffceb0b684c0ebd2d95ff48881db10\": container with ID starting with 744dca789fc995b4de5c6d2192de27669bffceb0b684c0ebd2d95ff48881db10 not found: ID does not exist" Nov 22 08:18:46 crc kubenswrapper[4789]: I1122 08:18:46.483686 4789 scope.go:117] "RemoveContainer" containerID="352c28c526ea7c4fb9b0d9a8f43aa191072de013667fe06518534c9c7609cace" Nov 22 08:18:46 crc kubenswrapper[4789]: I1122 08:18:46.484242 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"352c28c526ea7c4fb9b0d9a8f43aa191072de013667fe06518534c9c7609cace"} err="failed to get container status \"352c28c526ea7c4fb9b0d9a8f43aa191072de013667fe06518534c9c7609cace\": rpc error: code = NotFound desc = could not find container \"352c28c526ea7c4fb9b0d9a8f43aa191072de013667fe06518534c9c7609cace\": container with ID starting with 352c28c526ea7c4fb9b0d9a8f43aa191072de013667fe06518534c9c7609cace not found: ID does not exist" Nov 22 08:18:46 crc kubenswrapper[4789]: I1122 08:18:46.572128 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1f547dc8-eab8-420e-af46-30615fd4da38-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"1f547dc8-eab8-420e-af46-30615fd4da38\") " pod="openstack/nova-metadata-0" Nov 22 08:18:46 crc kubenswrapper[4789]: I1122 08:18:46.572212 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vwsmc\" (UniqueName: \"kubernetes.io/projected/1f547dc8-eab8-420e-af46-30615fd4da38-kube-api-access-vwsmc\") pod \"nova-metadata-0\" (UID: \"1f547dc8-eab8-420e-af46-30615fd4da38\") " pod="openstack/nova-metadata-0" Nov 22 08:18:46 crc kubenswrapper[4789]: I1122 08:18:46.572284 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f547dc8-eab8-420e-af46-30615fd4da38-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"1f547dc8-eab8-420e-af46-30615fd4da38\") " pod="openstack/nova-metadata-0" Nov 22 08:18:46 crc kubenswrapper[4789]: I1122 08:18:46.572341 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f547dc8-eab8-420e-af46-30615fd4da38-config-data\") pod \"nova-metadata-0\" (UID: \"1f547dc8-eab8-420e-af46-30615fd4da38\") " pod="openstack/nova-metadata-0" Nov 22 08:18:46 crc kubenswrapper[4789]: I1122 08:18:46.572399 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1f547dc8-eab8-420e-af46-30615fd4da38-logs\") pod \"nova-metadata-0\" (UID: \"1f547dc8-eab8-420e-af46-30615fd4da38\") " pod="openstack/nova-metadata-0" Nov 22 08:18:46 crc kubenswrapper[4789]: I1122 08:18:46.674116 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vwsmc\" (UniqueName: \"kubernetes.io/projected/1f547dc8-eab8-420e-af46-30615fd4da38-kube-api-access-vwsmc\") pod \"nova-metadata-0\" (UID: \"1f547dc8-eab8-420e-af46-30615fd4da38\") " pod="openstack/nova-metadata-0" Nov 22 08:18:46 crc kubenswrapper[4789]: I1122 08:18:46.674180 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f547dc8-eab8-420e-af46-30615fd4da38-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"1f547dc8-eab8-420e-af46-30615fd4da38\") " pod="openstack/nova-metadata-0" Nov 22 08:18:46 crc kubenswrapper[4789]: I1122 08:18:46.674244 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f547dc8-eab8-420e-af46-30615fd4da38-config-data\") pod \"nova-metadata-0\" (UID: \"1f547dc8-eab8-420e-af46-30615fd4da38\") " pod="openstack/nova-metadata-0" Nov 22 08:18:46 crc kubenswrapper[4789]: I1122 08:18:46.674300 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1f547dc8-eab8-420e-af46-30615fd4da38-logs\") pod \"nova-metadata-0\" (UID: \"1f547dc8-eab8-420e-af46-30615fd4da38\") " pod="openstack/nova-metadata-0" Nov 22 08:18:46 crc kubenswrapper[4789]: I1122 08:18:46.674346 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1f547dc8-eab8-420e-af46-30615fd4da38-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"1f547dc8-eab8-420e-af46-30615fd4da38\") " pod="openstack/nova-metadata-0" Nov 22 08:18:46 crc kubenswrapper[4789]: I1122 08:18:46.674772 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1f547dc8-eab8-420e-af46-30615fd4da38-logs\") pod \"nova-metadata-0\" (UID: \"1f547dc8-eab8-420e-af46-30615fd4da38\") " pod="openstack/nova-metadata-0" Nov 22 08:18:46 crc kubenswrapper[4789]: I1122 08:18:46.678234 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f547dc8-eab8-420e-af46-30615fd4da38-config-data\") pod \"nova-metadata-0\" (UID: \"1f547dc8-eab8-420e-af46-30615fd4da38\") " pod="openstack/nova-metadata-0" Nov 22 08:18:46 crc kubenswrapper[4789]: I1122 08:18:46.678495 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1f547dc8-eab8-420e-af46-30615fd4da38-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"1f547dc8-eab8-420e-af46-30615fd4da38\") " pod="openstack/nova-metadata-0" Nov 22 08:18:46 crc kubenswrapper[4789]: I1122 08:18:46.681657 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f547dc8-eab8-420e-af46-30615fd4da38-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"1f547dc8-eab8-420e-af46-30615fd4da38\") " pod="openstack/nova-metadata-0" Nov 22 08:18:46 crc kubenswrapper[4789]: I1122 08:18:46.700156 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vwsmc\" (UniqueName: \"kubernetes.io/projected/1f547dc8-eab8-420e-af46-30615fd4da38-kube-api-access-vwsmc\") pod \"nova-metadata-0\" (UID: \"1f547dc8-eab8-420e-af46-30615fd4da38\") " pod="openstack/nova-metadata-0" Nov 22 08:18:46 crc kubenswrapper[4789]: I1122 08:18:46.790583 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 22 08:18:47 crc kubenswrapper[4789]: W1122 08:18:47.233201 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1f547dc8_eab8_420e_af46_30615fd4da38.slice/crio-e9559a8790ccb63d080a958a6a2c257277081521598026de833486eb8cd46e98 WatchSource:0}: Error finding container e9559a8790ccb63d080a958a6a2c257277081521598026de833486eb8cd46e98: Status 404 returned error can't find the container with id e9559a8790ccb63d080a958a6a2c257277081521598026de833486eb8cd46e98 Nov 22 08:18:47 crc kubenswrapper[4789]: I1122 08:18:47.245501 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 22 08:18:47 crc kubenswrapper[4789]: I1122 08:18:47.412477 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1f547dc8-eab8-420e-af46-30615fd4da38","Type":"ContainerStarted","Data":"e9559a8790ccb63d080a958a6a2c257277081521598026de833486eb8cd46e98"} Nov 22 08:18:47 crc kubenswrapper[4789]: I1122 08:18:47.992775 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ebf4ac2c-3bd1-474e-94ee-e707e215a5f3" path="/var/lib/kubelet/pods/ebf4ac2c-3bd1-474e-94ee-e707e215a5f3/volumes" Nov 22 08:18:48 crc kubenswrapper[4789]: E1122 08:18:48.165061 4789 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod79d1ad09_f02f_421e_b1e5_10860722cc77.slice/crio-1d99062b301747b7ae1bd0307943d6d1d286baa1210d5067e5b0ada008972cf3.scope\": RecentStats: unable to find data in memory cache]" Nov 22 08:18:48 crc kubenswrapper[4789]: I1122 08:18:48.431233 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1f547dc8-eab8-420e-af46-30615fd4da38","Type":"ContainerStarted","Data":"bb703be4b250f73809414a5541515acfe36ef7e7ea0de717dc765cd6395809c4"} Nov 22 08:18:48 crc kubenswrapper[4789]: I1122 08:18:48.431281 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1f547dc8-eab8-420e-af46-30615fd4da38","Type":"ContainerStarted","Data":"c629553dadc27644fcbb1ea47c619fa4669d7b280aa764c6c907d6900cbe794e"} Nov 22 08:18:48 crc kubenswrapper[4789]: I1122 08:18:48.433039 4789 generic.go:334] "Generic (PLEG): container finished" podID="79d1ad09-f02f-421e-b1e5-10860722cc77" containerID="1d99062b301747b7ae1bd0307943d6d1d286baa1210d5067e5b0ada008972cf3" exitCode=0 Nov 22 08:18:48 crc kubenswrapper[4789]: I1122 08:18:48.433074 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-fzs5c" event={"ID":"79d1ad09-f02f-421e-b1e5-10860722cc77","Type":"ContainerDied","Data":"1d99062b301747b7ae1bd0307943d6d1d286baa1210d5067e5b0ada008972cf3"} Nov 22 08:18:48 crc kubenswrapper[4789]: I1122 08:18:48.463486 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.463463724 podStartE2EDuration="2.463463724s" podCreationTimestamp="2025-11-22 08:18:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:18:48.456416941 +0000 UTC m=+1342.690817214" watchObservedRunningTime="2025-11-22 08:18:48.463463724 +0000 UTC m=+1342.697863997" Nov 22 08:18:49 crc kubenswrapper[4789]: I1122 08:18:49.443147 4789 generic.go:334] "Generic (PLEG): container finished" podID="d4270eeb-3da5-464a-806a-358dbc4bd6ba" containerID="7128ef0b5659bd07e32c2171483c5a7e4125643355ad974d47dde4aa001b1fec" exitCode=0 Nov 22 08:18:49 crc kubenswrapper[4789]: I1122 08:18:49.443213 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-skqfc" event={"ID":"d4270eeb-3da5-464a-806a-358dbc4bd6ba","Type":"ContainerDied","Data":"7128ef0b5659bd07e32c2171483c5a7e4125643355ad974d47dde4aa001b1fec"} Nov 22 08:18:49 crc kubenswrapper[4789]: I1122 08:18:49.781658 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 22 08:18:49 crc kubenswrapper[4789]: I1122 08:18:49.781722 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 22 08:18:49 crc kubenswrapper[4789]: I1122 08:18:49.788254 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-fzs5c" Nov 22 08:18:49 crc kubenswrapper[4789]: I1122 08:18:49.826699 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/79d1ad09-f02f-421e-b1e5-10860722cc77-scripts\") pod \"79d1ad09-f02f-421e-b1e5-10860722cc77\" (UID: \"79d1ad09-f02f-421e-b1e5-10860722cc77\") " Nov 22 08:18:49 crc kubenswrapper[4789]: I1122 08:18:49.826942 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mb5qv\" (UniqueName: \"kubernetes.io/projected/79d1ad09-f02f-421e-b1e5-10860722cc77-kube-api-access-mb5qv\") pod \"79d1ad09-f02f-421e-b1e5-10860722cc77\" (UID: \"79d1ad09-f02f-421e-b1e5-10860722cc77\") " Nov 22 08:18:49 crc kubenswrapper[4789]: I1122 08:18:49.827069 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79d1ad09-f02f-421e-b1e5-10860722cc77-combined-ca-bundle\") pod \"79d1ad09-f02f-421e-b1e5-10860722cc77\" (UID: \"79d1ad09-f02f-421e-b1e5-10860722cc77\") " Nov 22 08:18:49 crc kubenswrapper[4789]: I1122 08:18:49.827155 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79d1ad09-f02f-421e-b1e5-10860722cc77-config-data\") pod \"79d1ad09-f02f-421e-b1e5-10860722cc77\" (UID: \"79d1ad09-f02f-421e-b1e5-10860722cc77\") " Nov 22 08:18:49 crc kubenswrapper[4789]: I1122 08:18:49.832554 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79d1ad09-f02f-421e-b1e5-10860722cc77-kube-api-access-mb5qv" (OuterVolumeSpecName: "kube-api-access-mb5qv") pod "79d1ad09-f02f-421e-b1e5-10860722cc77" (UID: "79d1ad09-f02f-421e-b1e5-10860722cc77"). InnerVolumeSpecName "kube-api-access-mb5qv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:18:49 crc kubenswrapper[4789]: I1122 08:18:49.834166 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79d1ad09-f02f-421e-b1e5-10860722cc77-scripts" (OuterVolumeSpecName: "scripts") pod "79d1ad09-f02f-421e-b1e5-10860722cc77" (UID: "79d1ad09-f02f-421e-b1e5-10860722cc77"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:18:49 crc kubenswrapper[4789]: I1122 08:18:49.856173 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79d1ad09-f02f-421e-b1e5-10860722cc77-config-data" (OuterVolumeSpecName: "config-data") pod "79d1ad09-f02f-421e-b1e5-10860722cc77" (UID: "79d1ad09-f02f-421e-b1e5-10860722cc77"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:18:49 crc kubenswrapper[4789]: I1122 08:18:49.856570 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79d1ad09-f02f-421e-b1e5-10860722cc77-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "79d1ad09-f02f-421e-b1e5-10860722cc77" (UID: "79d1ad09-f02f-421e-b1e5-10860722cc77"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:18:49 crc kubenswrapper[4789]: I1122 08:18:49.928520 4789 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79d1ad09-f02f-421e-b1e5-10860722cc77-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:18:49 crc kubenswrapper[4789]: I1122 08:18:49.928560 4789 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79d1ad09-f02f-421e-b1e5-10860722cc77-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:18:49 crc kubenswrapper[4789]: I1122 08:18:49.928568 4789 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/79d1ad09-f02f-421e-b1e5-10860722cc77-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:18:49 crc kubenswrapper[4789]: I1122 08:18:49.928577 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mb5qv\" (UniqueName: \"kubernetes.io/projected/79d1ad09-f02f-421e-b1e5-10860722cc77-kube-api-access-mb5qv\") on node \"crc\" DevicePath \"\"" Nov 22 08:18:50 crc kubenswrapper[4789]: I1122 08:18:50.026510 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 22 08:18:50 crc kubenswrapper[4789]: I1122 08:18:50.026609 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 22 08:18:50 crc kubenswrapper[4789]: I1122 08:18:50.053683 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 22 08:18:50 crc kubenswrapper[4789]: I1122 08:18:50.054936 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-69494d9f89-z9hd6" Nov 22 08:18:50 crc kubenswrapper[4789]: I1122 08:18:50.061854 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Nov 22 08:18:50 crc kubenswrapper[4789]: I1122 08:18:50.128716 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-775457b975-x8hvd"] Nov 22 08:18:50 crc kubenswrapper[4789]: I1122 08:18:50.128982 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-775457b975-x8hvd" podUID="63d8628c-ab9f-4c35-8550-08d647beffa4" containerName="dnsmasq-dns" containerID="cri-o://edb4df9f694719c6d72cb115eb8dcf575170c30a9b7954588cb0ade5ba02390f" gracePeriod=10 Nov 22 08:18:50 crc kubenswrapper[4789]: I1122 08:18:50.472697 4789 generic.go:334] "Generic (PLEG): container finished" podID="63d8628c-ab9f-4c35-8550-08d647beffa4" containerID="edb4df9f694719c6d72cb115eb8dcf575170c30a9b7954588cb0ade5ba02390f" exitCode=0 Nov 22 08:18:50 crc kubenswrapper[4789]: I1122 08:18:50.472861 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-775457b975-x8hvd" event={"ID":"63d8628c-ab9f-4c35-8550-08d647beffa4","Type":"ContainerDied","Data":"edb4df9f694719c6d72cb115eb8dcf575170c30a9b7954588cb0ade5ba02390f"} Nov 22 08:18:50 crc kubenswrapper[4789]: I1122 08:18:50.475815 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-fzs5c" Nov 22 08:18:50 crc kubenswrapper[4789]: I1122 08:18:50.475881 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-fzs5c" event={"ID":"79d1ad09-f02f-421e-b1e5-10860722cc77","Type":"ContainerDied","Data":"e258e3652b4b1c90544937b7e2c9af190d01701361b8142f0630f586d2139c1a"} Nov 22 08:18:50 crc kubenswrapper[4789]: I1122 08:18:50.475912 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e258e3652b4b1c90544937b7e2c9af190d01701361b8142f0630f586d2139c1a" Nov 22 08:18:50 crc kubenswrapper[4789]: I1122 08:18:50.530125 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 22 08:18:50 crc kubenswrapper[4789]: I1122 08:18:50.559065 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-775457b975-x8hvd" Nov 22 08:18:50 crc kubenswrapper[4789]: I1122 08:18:50.648009 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/63d8628c-ab9f-4c35-8550-08d647beffa4-ovsdbserver-nb\") pod \"63d8628c-ab9f-4c35-8550-08d647beffa4\" (UID: \"63d8628c-ab9f-4c35-8550-08d647beffa4\") " Nov 22 08:18:50 crc kubenswrapper[4789]: I1122 08:18:50.648183 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/63d8628c-ab9f-4c35-8550-08d647beffa4-config\") pod \"63d8628c-ab9f-4c35-8550-08d647beffa4\" (UID: \"63d8628c-ab9f-4c35-8550-08d647beffa4\") " Nov 22 08:18:50 crc kubenswrapper[4789]: I1122 08:18:50.648247 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/63d8628c-ab9f-4c35-8550-08d647beffa4-dns-svc\") pod \"63d8628c-ab9f-4c35-8550-08d647beffa4\" (UID: \"63d8628c-ab9f-4c35-8550-08d647beffa4\") " Nov 22 08:18:50 crc kubenswrapper[4789]: I1122 08:18:50.648287 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/63d8628c-ab9f-4c35-8550-08d647beffa4-ovsdbserver-sb\") pod \"63d8628c-ab9f-4c35-8550-08d647beffa4\" (UID: \"63d8628c-ab9f-4c35-8550-08d647beffa4\") " Nov 22 08:18:50 crc kubenswrapper[4789]: I1122 08:18:50.648312 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p49cm\" (UniqueName: \"kubernetes.io/projected/63d8628c-ab9f-4c35-8550-08d647beffa4-kube-api-access-p49cm\") pod \"63d8628c-ab9f-4c35-8550-08d647beffa4\" (UID: \"63d8628c-ab9f-4c35-8550-08d647beffa4\") " Nov 22 08:18:50 crc kubenswrapper[4789]: I1122 08:18:50.658732 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63d8628c-ab9f-4c35-8550-08d647beffa4-kube-api-access-p49cm" (OuterVolumeSpecName: "kube-api-access-p49cm") pod "63d8628c-ab9f-4c35-8550-08d647beffa4" (UID: "63d8628c-ab9f-4c35-8550-08d647beffa4"). InnerVolumeSpecName "kube-api-access-p49cm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:18:50 crc kubenswrapper[4789]: I1122 08:18:50.733575 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/63d8628c-ab9f-4c35-8550-08d647beffa4-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "63d8628c-ab9f-4c35-8550-08d647beffa4" (UID: "63d8628c-ab9f-4c35-8550-08d647beffa4"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:18:50 crc kubenswrapper[4789]: I1122 08:18:50.738048 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/63d8628c-ab9f-4c35-8550-08d647beffa4-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "63d8628c-ab9f-4c35-8550-08d647beffa4" (UID: "63d8628c-ab9f-4c35-8550-08d647beffa4"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:18:50 crc kubenswrapper[4789]: I1122 08:18:50.750329 4789 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/63d8628c-ab9f-4c35-8550-08d647beffa4-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 22 08:18:50 crc kubenswrapper[4789]: I1122 08:18:50.750366 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p49cm\" (UniqueName: \"kubernetes.io/projected/63d8628c-ab9f-4c35-8550-08d647beffa4-kube-api-access-p49cm\") on node \"crc\" DevicePath \"\"" Nov 22 08:18:50 crc kubenswrapper[4789]: I1122 08:18:50.750377 4789 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/63d8628c-ab9f-4c35-8550-08d647beffa4-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 22 08:18:50 crc kubenswrapper[4789]: I1122 08:18:50.755047 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/63d8628c-ab9f-4c35-8550-08d647beffa4-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "63d8628c-ab9f-4c35-8550-08d647beffa4" (UID: "63d8628c-ab9f-4c35-8550-08d647beffa4"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:18:50 crc kubenswrapper[4789]: I1122 08:18:50.764478 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 22 08:18:50 crc kubenswrapper[4789]: I1122 08:18:50.764731 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="1e4eaef7-30a8-4a62-bb67-9aafc8f52b76" containerName="nova-api-log" containerID="cri-o://6f051cc0e759f5f1a60b82855e4f3e14affc615ae4b07071c76bea4e60c971c2" gracePeriod=30 Nov 22 08:18:50 crc kubenswrapper[4789]: I1122 08:18:50.764901 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="1e4eaef7-30a8-4a62-bb67-9aafc8f52b76" containerName="nova-api-api" containerID="cri-o://3cb2ac3c4243a5a7bc7f0d9e9e4c7f1a499314f692dbb42de7c8edaf4c703b4f" gracePeriod=30 Nov 22 08:18:50 crc kubenswrapper[4789]: I1122 08:18:50.775073 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="1e4eaef7-30a8-4a62-bb67-9aafc8f52b76" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.174:8774/\": EOF" Nov 22 08:18:50 crc kubenswrapper[4789]: I1122 08:18:50.775685 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="1e4eaef7-30a8-4a62-bb67-9aafc8f52b76" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.174:8774/\": EOF" Nov 22 08:18:50 crc kubenswrapper[4789]: I1122 08:18:50.777419 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/63d8628c-ab9f-4c35-8550-08d647beffa4-config" (OuterVolumeSpecName: "config") pod "63d8628c-ab9f-4c35-8550-08d647beffa4" (UID: "63d8628c-ab9f-4c35-8550-08d647beffa4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:18:50 crc kubenswrapper[4789]: I1122 08:18:50.839455 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 22 08:18:50 crc kubenswrapper[4789]: I1122 08:18:50.839872 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="1f547dc8-eab8-420e-af46-30615fd4da38" containerName="nova-metadata-metadata" containerID="cri-o://c629553dadc27644fcbb1ea47c619fa4669d7b280aa764c6c907d6900cbe794e" gracePeriod=30 Nov 22 08:18:50 crc kubenswrapper[4789]: I1122 08:18:50.839859 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="1f547dc8-eab8-420e-af46-30615fd4da38" containerName="nova-metadata-log" containerID="cri-o://bb703be4b250f73809414a5541515acfe36ef7e7ea0de717dc765cd6395809c4" gracePeriod=30 Nov 22 08:18:50 crc kubenswrapper[4789]: I1122 08:18:50.853074 4789 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/63d8628c-ab9f-4c35-8550-08d647beffa4-config\") on node \"crc\" DevicePath \"\"" Nov 22 08:18:50 crc kubenswrapper[4789]: I1122 08:18:50.853139 4789 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/63d8628c-ab9f-4c35-8550-08d647beffa4-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 22 08:18:50 crc kubenswrapper[4789]: I1122 08:18:50.868533 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-skqfc" Nov 22 08:18:50 crc kubenswrapper[4789]: I1122 08:18:50.954116 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4270eeb-3da5-464a-806a-358dbc4bd6ba-combined-ca-bundle\") pod \"d4270eeb-3da5-464a-806a-358dbc4bd6ba\" (UID: \"d4270eeb-3da5-464a-806a-358dbc4bd6ba\") " Nov 22 08:18:50 crc kubenswrapper[4789]: I1122 08:18:50.954303 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4270eeb-3da5-464a-806a-358dbc4bd6ba-config-data\") pod \"d4270eeb-3da5-464a-806a-358dbc4bd6ba\" (UID: \"d4270eeb-3da5-464a-806a-358dbc4bd6ba\") " Nov 22 08:18:50 crc kubenswrapper[4789]: I1122 08:18:50.954401 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5ww4b\" (UniqueName: \"kubernetes.io/projected/d4270eeb-3da5-464a-806a-358dbc4bd6ba-kube-api-access-5ww4b\") pod \"d4270eeb-3da5-464a-806a-358dbc4bd6ba\" (UID: \"d4270eeb-3da5-464a-806a-358dbc4bd6ba\") " Nov 22 08:18:50 crc kubenswrapper[4789]: I1122 08:18:50.954429 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4270eeb-3da5-464a-806a-358dbc4bd6ba-scripts\") pod \"d4270eeb-3da5-464a-806a-358dbc4bd6ba\" (UID: \"d4270eeb-3da5-464a-806a-358dbc4bd6ba\") " Nov 22 08:18:50 crc kubenswrapper[4789]: I1122 08:18:50.958895 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4270eeb-3da5-464a-806a-358dbc4bd6ba-scripts" (OuterVolumeSpecName: "scripts") pod "d4270eeb-3da5-464a-806a-358dbc4bd6ba" (UID: "d4270eeb-3da5-464a-806a-358dbc4bd6ba"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:18:50 crc kubenswrapper[4789]: I1122 08:18:50.959414 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4270eeb-3da5-464a-806a-358dbc4bd6ba-kube-api-access-5ww4b" (OuterVolumeSpecName: "kube-api-access-5ww4b") pod "d4270eeb-3da5-464a-806a-358dbc4bd6ba" (UID: "d4270eeb-3da5-464a-806a-358dbc4bd6ba"). InnerVolumeSpecName "kube-api-access-5ww4b". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:18:50 crc kubenswrapper[4789]: I1122 08:18:50.987627 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4270eeb-3da5-464a-806a-358dbc4bd6ba-config-data" (OuterVolumeSpecName: "config-data") pod "d4270eeb-3da5-464a-806a-358dbc4bd6ba" (UID: "d4270eeb-3da5-464a-806a-358dbc4bd6ba"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:18:50 crc kubenswrapper[4789]: I1122 08:18:50.989246 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4270eeb-3da5-464a-806a-358dbc4bd6ba-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d4270eeb-3da5-464a-806a-358dbc4bd6ba" (UID: "d4270eeb-3da5-464a-806a-358dbc4bd6ba"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:18:51 crc kubenswrapper[4789]: I1122 08:18:51.028536 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 22 08:18:51 crc kubenswrapper[4789]: I1122 08:18:51.057945 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5ww4b\" (UniqueName: \"kubernetes.io/projected/d4270eeb-3da5-464a-806a-358dbc4bd6ba-kube-api-access-5ww4b\") on node \"crc\" DevicePath \"\"" Nov 22 08:18:51 crc kubenswrapper[4789]: I1122 08:18:51.057985 4789 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4270eeb-3da5-464a-806a-358dbc4bd6ba-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:18:51 crc kubenswrapper[4789]: I1122 08:18:51.057996 4789 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4270eeb-3da5-464a-806a-358dbc4bd6ba-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:18:51 crc kubenswrapper[4789]: I1122 08:18:51.058018 4789 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4270eeb-3da5-464a-806a-358dbc4bd6ba-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:18:51 crc kubenswrapper[4789]: I1122 08:18:51.487747 4789 generic.go:334] "Generic (PLEG): container finished" podID="1e4eaef7-30a8-4a62-bb67-9aafc8f52b76" containerID="6f051cc0e759f5f1a60b82855e4f3e14affc615ae4b07071c76bea4e60c971c2" exitCode=143 Nov 22 08:18:51 crc kubenswrapper[4789]: I1122 08:18:51.487795 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1e4eaef7-30a8-4a62-bb67-9aafc8f52b76","Type":"ContainerDied","Data":"6f051cc0e759f5f1a60b82855e4f3e14affc615ae4b07071c76bea4e60c971c2"} Nov 22 08:18:51 crc kubenswrapper[4789]: I1122 08:18:51.490864 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-skqfc" event={"ID":"d4270eeb-3da5-464a-806a-358dbc4bd6ba","Type":"ContainerDied","Data":"b28cb0f340702a695e4339755c1b4af7f18c0a5874c60254b801ea4d53670d0b"} Nov 22 08:18:51 crc kubenswrapper[4789]: I1122 08:18:51.490930 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b28cb0f340702a695e4339755c1b4af7f18c0a5874c60254b801ea4d53670d0b" Nov 22 08:18:51 crc kubenswrapper[4789]: I1122 08:18:51.491024 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-skqfc" Nov 22 08:18:51 crc kubenswrapper[4789]: I1122 08:18:51.495483 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-775457b975-x8hvd" event={"ID":"63d8628c-ab9f-4c35-8550-08d647beffa4","Type":"ContainerDied","Data":"4e97eb89335c67cd4648cf09d9cf3029ec59539eee8a712ea386948b010051d1"} Nov 22 08:18:51 crc kubenswrapper[4789]: I1122 08:18:51.495522 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-775457b975-x8hvd" Nov 22 08:18:51 crc kubenswrapper[4789]: I1122 08:18:51.495584 4789 scope.go:117] "RemoveContainer" containerID="edb4df9f694719c6d72cb115eb8dcf575170c30a9b7954588cb0ade5ba02390f" Nov 22 08:18:51 crc kubenswrapper[4789]: I1122 08:18:51.505662 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 22 08:18:51 crc kubenswrapper[4789]: I1122 08:18:51.506072 4789 generic.go:334] "Generic (PLEG): container finished" podID="1f547dc8-eab8-420e-af46-30615fd4da38" containerID="c629553dadc27644fcbb1ea47c619fa4669d7b280aa764c6c907d6900cbe794e" exitCode=0 Nov 22 08:18:51 crc kubenswrapper[4789]: I1122 08:18:51.506094 4789 generic.go:334] "Generic (PLEG): container finished" podID="1f547dc8-eab8-420e-af46-30615fd4da38" containerID="bb703be4b250f73809414a5541515acfe36ef7e7ea0de717dc765cd6395809c4" exitCode=143 Nov 22 08:18:51 crc kubenswrapper[4789]: I1122 08:18:51.506120 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1f547dc8-eab8-420e-af46-30615fd4da38","Type":"ContainerDied","Data":"c629553dadc27644fcbb1ea47c619fa4669d7b280aa764c6c907d6900cbe794e"} Nov 22 08:18:51 crc kubenswrapper[4789]: I1122 08:18:51.506146 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1f547dc8-eab8-420e-af46-30615fd4da38","Type":"ContainerDied","Data":"bb703be4b250f73809414a5541515acfe36ef7e7ea0de717dc765cd6395809c4"} Nov 22 08:18:51 crc kubenswrapper[4789]: I1122 08:18:51.537049 4789 scope.go:117] "RemoveContainer" containerID="8d5b43af47dc8c714c823ec10bfe8850e47ceb36a28d5618349498082600cfa2" Nov 22 08:18:51 crc kubenswrapper[4789]: I1122 08:18:51.569628 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f547dc8-eab8-420e-af46-30615fd4da38-combined-ca-bundle\") pod \"1f547dc8-eab8-420e-af46-30615fd4da38\" (UID: \"1f547dc8-eab8-420e-af46-30615fd4da38\") " Nov 22 08:18:51 crc kubenswrapper[4789]: I1122 08:18:51.569721 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vwsmc\" (UniqueName: \"kubernetes.io/projected/1f547dc8-eab8-420e-af46-30615fd4da38-kube-api-access-vwsmc\") pod \"1f547dc8-eab8-420e-af46-30615fd4da38\" (UID: \"1f547dc8-eab8-420e-af46-30615fd4da38\") " Nov 22 08:18:51 crc kubenswrapper[4789]: I1122 08:18:51.569825 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1f547dc8-eab8-420e-af46-30615fd4da38-logs\") pod \"1f547dc8-eab8-420e-af46-30615fd4da38\" (UID: \"1f547dc8-eab8-420e-af46-30615fd4da38\") " Nov 22 08:18:51 crc kubenswrapper[4789]: I1122 08:18:51.569883 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f547dc8-eab8-420e-af46-30615fd4da38-config-data\") pod \"1f547dc8-eab8-420e-af46-30615fd4da38\" (UID: \"1f547dc8-eab8-420e-af46-30615fd4da38\") " Nov 22 08:18:51 crc kubenswrapper[4789]: I1122 08:18:51.569915 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1f547dc8-eab8-420e-af46-30615fd4da38-nova-metadata-tls-certs\") pod \"1f547dc8-eab8-420e-af46-30615fd4da38\" (UID: \"1f547dc8-eab8-420e-af46-30615fd4da38\") " Nov 22 08:18:51 crc kubenswrapper[4789]: I1122 08:18:51.570836 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 22 08:18:51 crc kubenswrapper[4789]: E1122 08:18:51.571279 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63d8628c-ab9f-4c35-8550-08d647beffa4" containerName="init" Nov 22 08:18:51 crc kubenswrapper[4789]: I1122 08:18:51.571296 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="63d8628c-ab9f-4c35-8550-08d647beffa4" containerName="init" Nov 22 08:18:51 crc kubenswrapper[4789]: E1122 08:18:51.571318 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f547dc8-eab8-420e-af46-30615fd4da38" containerName="nova-metadata-log" Nov 22 08:18:51 crc kubenswrapper[4789]: I1122 08:18:51.571326 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f547dc8-eab8-420e-af46-30615fd4da38" containerName="nova-metadata-log" Nov 22 08:18:51 crc kubenswrapper[4789]: E1122 08:18:51.571348 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4270eeb-3da5-464a-806a-358dbc4bd6ba" containerName="nova-cell1-conductor-db-sync" Nov 22 08:18:51 crc kubenswrapper[4789]: I1122 08:18:51.571357 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4270eeb-3da5-464a-806a-358dbc4bd6ba" containerName="nova-cell1-conductor-db-sync" Nov 22 08:18:51 crc kubenswrapper[4789]: E1122 08:18:51.571369 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79d1ad09-f02f-421e-b1e5-10860722cc77" containerName="nova-manage" Nov 22 08:18:51 crc kubenswrapper[4789]: I1122 08:18:51.571376 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="79d1ad09-f02f-421e-b1e5-10860722cc77" containerName="nova-manage" Nov 22 08:18:51 crc kubenswrapper[4789]: E1122 08:18:51.571393 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f547dc8-eab8-420e-af46-30615fd4da38" containerName="nova-metadata-metadata" Nov 22 08:18:51 crc kubenswrapper[4789]: I1122 08:18:51.571399 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f547dc8-eab8-420e-af46-30615fd4da38" containerName="nova-metadata-metadata" Nov 22 08:18:51 crc kubenswrapper[4789]: E1122 08:18:51.571416 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63d8628c-ab9f-4c35-8550-08d647beffa4" containerName="dnsmasq-dns" Nov 22 08:18:51 crc kubenswrapper[4789]: I1122 08:18:51.571423 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="63d8628c-ab9f-4c35-8550-08d647beffa4" containerName="dnsmasq-dns" Nov 22 08:18:51 crc kubenswrapper[4789]: I1122 08:18:51.571638 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f547dc8-eab8-420e-af46-30615fd4da38" containerName="nova-metadata-log" Nov 22 08:18:51 crc kubenswrapper[4789]: I1122 08:18:51.571662 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4270eeb-3da5-464a-806a-358dbc4bd6ba" containerName="nova-cell1-conductor-db-sync" Nov 22 08:18:51 crc kubenswrapper[4789]: I1122 08:18:51.571679 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="79d1ad09-f02f-421e-b1e5-10860722cc77" containerName="nova-manage" Nov 22 08:18:51 crc kubenswrapper[4789]: I1122 08:18:51.571693 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="63d8628c-ab9f-4c35-8550-08d647beffa4" containerName="dnsmasq-dns" Nov 22 08:18:51 crc kubenswrapper[4789]: I1122 08:18:51.571702 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f547dc8-eab8-420e-af46-30615fd4da38" containerName="nova-metadata-metadata" Nov 22 08:18:51 crc kubenswrapper[4789]: I1122 08:18:51.572379 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 22 08:18:51 crc kubenswrapper[4789]: I1122 08:18:51.574117 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1f547dc8-eab8-420e-af46-30615fd4da38-logs" (OuterVolumeSpecName: "logs") pod "1f547dc8-eab8-420e-af46-30615fd4da38" (UID: "1f547dc8-eab8-420e-af46-30615fd4da38"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:18:51 crc kubenswrapper[4789]: I1122 08:18:51.576352 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Nov 22 08:18:51 crc kubenswrapper[4789]: I1122 08:18:51.578770 4789 scope.go:117] "RemoveContainer" containerID="c629553dadc27644fcbb1ea47c619fa4669d7b280aa764c6c907d6900cbe794e" Nov 22 08:18:51 crc kubenswrapper[4789]: I1122 08:18:51.580503 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1f547dc8-eab8-420e-af46-30615fd4da38-kube-api-access-vwsmc" (OuterVolumeSpecName: "kube-api-access-vwsmc") pod "1f547dc8-eab8-420e-af46-30615fd4da38" (UID: "1f547dc8-eab8-420e-af46-30615fd4da38"). InnerVolumeSpecName "kube-api-access-vwsmc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:18:51 crc kubenswrapper[4789]: I1122 08:18:51.598169 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-775457b975-x8hvd"] Nov 22 08:18:51 crc kubenswrapper[4789]: I1122 08:18:51.606568 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f547dc8-eab8-420e-af46-30615fd4da38-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1f547dc8-eab8-420e-af46-30615fd4da38" (UID: "1f547dc8-eab8-420e-af46-30615fd4da38"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:18:51 crc kubenswrapper[4789]: I1122 08:18:51.610103 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-775457b975-x8hvd"] Nov 22 08:18:51 crc kubenswrapper[4789]: I1122 08:18:51.619403 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f547dc8-eab8-420e-af46-30615fd4da38-config-data" (OuterVolumeSpecName: "config-data") pod "1f547dc8-eab8-420e-af46-30615fd4da38" (UID: "1f547dc8-eab8-420e-af46-30615fd4da38"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:18:51 crc kubenswrapper[4789]: I1122 08:18:51.623060 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 22 08:18:51 crc kubenswrapper[4789]: I1122 08:18:51.623480 4789 scope.go:117] "RemoveContainer" containerID="bb703be4b250f73809414a5541515acfe36ef7e7ea0de717dc765cd6395809c4" Nov 22 08:18:51 crc kubenswrapper[4789]: I1122 08:18:51.637918 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f547dc8-eab8-420e-af46-30615fd4da38-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "1f547dc8-eab8-420e-af46-30615fd4da38" (UID: "1f547dc8-eab8-420e-af46-30615fd4da38"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:18:51 crc kubenswrapper[4789]: I1122 08:18:51.641272 4789 scope.go:117] "RemoveContainer" containerID="c629553dadc27644fcbb1ea47c619fa4669d7b280aa764c6c907d6900cbe794e" Nov 22 08:18:51 crc kubenswrapper[4789]: E1122 08:18:51.641838 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c629553dadc27644fcbb1ea47c619fa4669d7b280aa764c6c907d6900cbe794e\": container with ID starting with c629553dadc27644fcbb1ea47c619fa4669d7b280aa764c6c907d6900cbe794e not found: ID does not exist" containerID="c629553dadc27644fcbb1ea47c619fa4669d7b280aa764c6c907d6900cbe794e" Nov 22 08:18:51 crc kubenswrapper[4789]: I1122 08:18:51.641879 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c629553dadc27644fcbb1ea47c619fa4669d7b280aa764c6c907d6900cbe794e"} err="failed to get container status \"c629553dadc27644fcbb1ea47c619fa4669d7b280aa764c6c907d6900cbe794e\": rpc error: code = NotFound desc = could not find container \"c629553dadc27644fcbb1ea47c619fa4669d7b280aa764c6c907d6900cbe794e\": container with ID starting with c629553dadc27644fcbb1ea47c619fa4669d7b280aa764c6c907d6900cbe794e not found: ID does not exist" Nov 22 08:18:51 crc kubenswrapper[4789]: I1122 08:18:51.641901 4789 scope.go:117] "RemoveContainer" containerID="bb703be4b250f73809414a5541515acfe36ef7e7ea0de717dc765cd6395809c4" Nov 22 08:18:51 crc kubenswrapper[4789]: E1122 08:18:51.642245 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bb703be4b250f73809414a5541515acfe36ef7e7ea0de717dc765cd6395809c4\": container with ID starting with bb703be4b250f73809414a5541515acfe36ef7e7ea0de717dc765cd6395809c4 not found: ID does not exist" containerID="bb703be4b250f73809414a5541515acfe36ef7e7ea0de717dc765cd6395809c4" Nov 22 08:18:51 crc kubenswrapper[4789]: I1122 08:18:51.642266 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bb703be4b250f73809414a5541515acfe36ef7e7ea0de717dc765cd6395809c4"} err="failed to get container status \"bb703be4b250f73809414a5541515acfe36ef7e7ea0de717dc765cd6395809c4\": rpc error: code = NotFound desc = could not find container \"bb703be4b250f73809414a5541515acfe36ef7e7ea0de717dc765cd6395809c4\": container with ID starting with bb703be4b250f73809414a5541515acfe36ef7e7ea0de717dc765cd6395809c4 not found: ID does not exist" Nov 22 08:18:51 crc kubenswrapper[4789]: I1122 08:18:51.671723 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/298b1252-2a3e-4f1a-a5af-5e573964ddd4-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"298b1252-2a3e-4f1a-a5af-5e573964ddd4\") " pod="openstack/nova-cell1-conductor-0" Nov 22 08:18:51 crc kubenswrapper[4789]: I1122 08:18:51.671805 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2777m\" (UniqueName: \"kubernetes.io/projected/298b1252-2a3e-4f1a-a5af-5e573964ddd4-kube-api-access-2777m\") pod \"nova-cell1-conductor-0\" (UID: \"298b1252-2a3e-4f1a-a5af-5e573964ddd4\") " pod="openstack/nova-cell1-conductor-0" Nov 22 08:18:51 crc kubenswrapper[4789]: I1122 08:18:51.671860 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/298b1252-2a3e-4f1a-a5af-5e573964ddd4-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"298b1252-2a3e-4f1a-a5af-5e573964ddd4\") " pod="openstack/nova-cell1-conductor-0" Nov 22 08:18:51 crc kubenswrapper[4789]: I1122 08:18:51.671951 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vwsmc\" (UniqueName: \"kubernetes.io/projected/1f547dc8-eab8-420e-af46-30615fd4da38-kube-api-access-vwsmc\") on node \"crc\" DevicePath \"\"" Nov 22 08:18:51 crc kubenswrapper[4789]: I1122 08:18:51.671962 4789 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1f547dc8-eab8-420e-af46-30615fd4da38-logs\") on node \"crc\" DevicePath \"\"" Nov 22 08:18:51 crc kubenswrapper[4789]: I1122 08:18:51.671971 4789 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f547dc8-eab8-420e-af46-30615fd4da38-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:18:51 crc kubenswrapper[4789]: I1122 08:18:51.671981 4789 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1f547dc8-eab8-420e-af46-30615fd4da38-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 22 08:18:51 crc kubenswrapper[4789]: I1122 08:18:51.671989 4789 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f547dc8-eab8-420e-af46-30615fd4da38-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:18:51 crc kubenswrapper[4789]: I1122 08:18:51.773125 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/298b1252-2a3e-4f1a-a5af-5e573964ddd4-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"298b1252-2a3e-4f1a-a5af-5e573964ddd4\") " pod="openstack/nova-cell1-conductor-0" Nov 22 08:18:51 crc kubenswrapper[4789]: I1122 08:18:51.773201 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2777m\" (UniqueName: \"kubernetes.io/projected/298b1252-2a3e-4f1a-a5af-5e573964ddd4-kube-api-access-2777m\") pod \"nova-cell1-conductor-0\" (UID: \"298b1252-2a3e-4f1a-a5af-5e573964ddd4\") " pod="openstack/nova-cell1-conductor-0" Nov 22 08:18:51 crc kubenswrapper[4789]: I1122 08:18:51.773298 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/298b1252-2a3e-4f1a-a5af-5e573964ddd4-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"298b1252-2a3e-4f1a-a5af-5e573964ddd4\") " pod="openstack/nova-cell1-conductor-0" Nov 22 08:18:51 crc kubenswrapper[4789]: I1122 08:18:51.777704 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/298b1252-2a3e-4f1a-a5af-5e573964ddd4-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"298b1252-2a3e-4f1a-a5af-5e573964ddd4\") " pod="openstack/nova-cell1-conductor-0" Nov 22 08:18:51 crc kubenswrapper[4789]: I1122 08:18:51.777855 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/298b1252-2a3e-4f1a-a5af-5e573964ddd4-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"298b1252-2a3e-4f1a-a5af-5e573964ddd4\") " pod="openstack/nova-cell1-conductor-0" Nov 22 08:18:51 crc kubenswrapper[4789]: I1122 08:18:51.789942 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2777m\" (UniqueName: \"kubernetes.io/projected/298b1252-2a3e-4f1a-a5af-5e573964ddd4-kube-api-access-2777m\") pod \"nova-cell1-conductor-0\" (UID: \"298b1252-2a3e-4f1a-a5af-5e573964ddd4\") " pod="openstack/nova-cell1-conductor-0" Nov 22 08:18:51 crc kubenswrapper[4789]: I1122 08:18:51.897058 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 22 08:18:51 crc kubenswrapper[4789]: I1122 08:18:51.981718 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="63d8628c-ab9f-4c35-8550-08d647beffa4" path="/var/lib/kubelet/pods/63d8628c-ab9f-4c35-8550-08d647beffa4/volumes" Nov 22 08:18:52 crc kubenswrapper[4789]: I1122 08:18:52.352655 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 22 08:18:52 crc kubenswrapper[4789]: W1122 08:18:52.358886 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod298b1252_2a3e_4f1a_a5af_5e573964ddd4.slice/crio-bbe2d0f3dad1b9ecd3c5df2135c03a7f1d72cce2f15037c75391330de7698a27 WatchSource:0}: Error finding container bbe2d0f3dad1b9ecd3c5df2135c03a7f1d72cce2f15037c75391330de7698a27: Status 404 returned error can't find the container with id bbe2d0f3dad1b9ecd3c5df2135c03a7f1d72cce2f15037c75391330de7698a27 Nov 22 08:18:52 crc kubenswrapper[4789]: I1122 08:18:52.517823 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"298b1252-2a3e-4f1a-a5af-5e573964ddd4","Type":"ContainerStarted","Data":"bbe2d0f3dad1b9ecd3c5df2135c03a7f1d72cce2f15037c75391330de7698a27"} Nov 22 08:18:52 crc kubenswrapper[4789]: I1122 08:18:52.520605 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1f547dc8-eab8-420e-af46-30615fd4da38","Type":"ContainerDied","Data":"e9559a8790ccb63d080a958a6a2c257277081521598026de833486eb8cd46e98"} Nov 22 08:18:52 crc kubenswrapper[4789]: I1122 08:18:52.520639 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="9cb3ad54-8a01-43df-b938-b6e68bcb704c" containerName="nova-scheduler-scheduler" containerID="cri-o://4bc244b6de6b03bed55cd48a85231ca6f96de9810592b5ee76f208984a87adc0" gracePeriod=30 Nov 22 08:18:52 crc kubenswrapper[4789]: I1122 08:18:52.520816 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 22 08:18:52 crc kubenswrapper[4789]: I1122 08:18:52.550337 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 22 08:18:52 crc kubenswrapper[4789]: I1122 08:18:52.598250 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 22 08:18:52 crc kubenswrapper[4789]: I1122 08:18:52.610095 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 22 08:18:52 crc kubenswrapper[4789]: I1122 08:18:52.611631 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 22 08:18:52 crc kubenswrapper[4789]: I1122 08:18:52.613765 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 22 08:18:52 crc kubenswrapper[4789]: I1122 08:18:52.616531 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 22 08:18:52 crc kubenswrapper[4789]: I1122 08:18:52.629582 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 22 08:18:52 crc kubenswrapper[4789]: I1122 08:18:52.694192 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/08796aa5-e098-4917-8eb3-dca574b4670a-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"08796aa5-e098-4917-8eb3-dca574b4670a\") " pod="openstack/nova-metadata-0" Nov 22 08:18:52 crc kubenswrapper[4789]: I1122 08:18:52.700831 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4wzs4\" (UniqueName: \"kubernetes.io/projected/08796aa5-e098-4917-8eb3-dca574b4670a-kube-api-access-4wzs4\") pod \"nova-metadata-0\" (UID: \"08796aa5-e098-4917-8eb3-dca574b4670a\") " pod="openstack/nova-metadata-0" Nov 22 08:18:52 crc kubenswrapper[4789]: I1122 08:18:52.700875 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08796aa5-e098-4917-8eb3-dca574b4670a-config-data\") pod \"nova-metadata-0\" (UID: \"08796aa5-e098-4917-8eb3-dca574b4670a\") " pod="openstack/nova-metadata-0" Nov 22 08:18:52 crc kubenswrapper[4789]: I1122 08:18:52.700983 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/08796aa5-e098-4917-8eb3-dca574b4670a-logs\") pod \"nova-metadata-0\" (UID: \"08796aa5-e098-4917-8eb3-dca574b4670a\") " pod="openstack/nova-metadata-0" Nov 22 08:18:52 crc kubenswrapper[4789]: I1122 08:18:52.701069 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08796aa5-e098-4917-8eb3-dca574b4670a-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"08796aa5-e098-4917-8eb3-dca574b4670a\") " pod="openstack/nova-metadata-0" Nov 22 08:18:52 crc kubenswrapper[4789]: I1122 08:18:52.803246 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4wzs4\" (UniqueName: \"kubernetes.io/projected/08796aa5-e098-4917-8eb3-dca574b4670a-kube-api-access-4wzs4\") pod \"nova-metadata-0\" (UID: \"08796aa5-e098-4917-8eb3-dca574b4670a\") " pod="openstack/nova-metadata-0" Nov 22 08:18:52 crc kubenswrapper[4789]: I1122 08:18:52.803294 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08796aa5-e098-4917-8eb3-dca574b4670a-config-data\") pod \"nova-metadata-0\" (UID: \"08796aa5-e098-4917-8eb3-dca574b4670a\") " pod="openstack/nova-metadata-0" Nov 22 08:18:52 crc kubenswrapper[4789]: I1122 08:18:52.803338 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/08796aa5-e098-4917-8eb3-dca574b4670a-logs\") pod \"nova-metadata-0\" (UID: \"08796aa5-e098-4917-8eb3-dca574b4670a\") " pod="openstack/nova-metadata-0" Nov 22 08:18:52 crc kubenswrapper[4789]: I1122 08:18:52.803381 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08796aa5-e098-4917-8eb3-dca574b4670a-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"08796aa5-e098-4917-8eb3-dca574b4670a\") " pod="openstack/nova-metadata-0" Nov 22 08:18:52 crc kubenswrapper[4789]: I1122 08:18:52.803447 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/08796aa5-e098-4917-8eb3-dca574b4670a-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"08796aa5-e098-4917-8eb3-dca574b4670a\") " pod="openstack/nova-metadata-0" Nov 22 08:18:52 crc kubenswrapper[4789]: I1122 08:18:52.804057 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/08796aa5-e098-4917-8eb3-dca574b4670a-logs\") pod \"nova-metadata-0\" (UID: \"08796aa5-e098-4917-8eb3-dca574b4670a\") " pod="openstack/nova-metadata-0" Nov 22 08:18:52 crc kubenswrapper[4789]: I1122 08:18:52.808764 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08796aa5-e098-4917-8eb3-dca574b4670a-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"08796aa5-e098-4917-8eb3-dca574b4670a\") " pod="openstack/nova-metadata-0" Nov 22 08:18:52 crc kubenswrapper[4789]: I1122 08:18:52.809503 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08796aa5-e098-4917-8eb3-dca574b4670a-config-data\") pod \"nova-metadata-0\" (UID: \"08796aa5-e098-4917-8eb3-dca574b4670a\") " pod="openstack/nova-metadata-0" Nov 22 08:18:52 crc kubenswrapper[4789]: I1122 08:18:52.818829 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/08796aa5-e098-4917-8eb3-dca574b4670a-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"08796aa5-e098-4917-8eb3-dca574b4670a\") " pod="openstack/nova-metadata-0" Nov 22 08:18:52 crc kubenswrapper[4789]: I1122 08:18:52.819207 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4wzs4\" (UniqueName: \"kubernetes.io/projected/08796aa5-e098-4917-8eb3-dca574b4670a-kube-api-access-4wzs4\") pod \"nova-metadata-0\" (UID: \"08796aa5-e098-4917-8eb3-dca574b4670a\") " pod="openstack/nova-metadata-0" Nov 22 08:18:52 crc kubenswrapper[4789]: I1122 08:18:52.934703 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 22 08:18:53 crc kubenswrapper[4789]: I1122 08:18:53.369951 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 22 08:18:53 crc kubenswrapper[4789]: W1122 08:18:53.381085 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod08796aa5_e098_4917_8eb3_dca574b4670a.slice/crio-b5d645d57033db6a068d4d4f1f001a1baf3ffa53f8062e3a9af095ad32653cee WatchSource:0}: Error finding container b5d645d57033db6a068d4d4f1f001a1baf3ffa53f8062e3a9af095ad32653cee: Status 404 returned error can't find the container with id b5d645d57033db6a068d4d4f1f001a1baf3ffa53f8062e3a9af095ad32653cee Nov 22 08:18:53 crc kubenswrapper[4789]: I1122 08:18:53.531695 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"298b1252-2a3e-4f1a-a5af-5e573964ddd4","Type":"ContainerStarted","Data":"6ca08c33934691fa7e915b6b1dcb852a62d1e74696ce81adbd5aaae1ba57e8d6"} Nov 22 08:18:53 crc kubenswrapper[4789]: I1122 08:18:53.531785 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Nov 22 08:18:53 crc kubenswrapper[4789]: I1122 08:18:53.543723 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"08796aa5-e098-4917-8eb3-dca574b4670a","Type":"ContainerStarted","Data":"b5d645d57033db6a068d4d4f1f001a1baf3ffa53f8062e3a9af095ad32653cee"} Nov 22 08:18:53 crc kubenswrapper[4789]: I1122 08:18:53.552159 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.552139936 podStartE2EDuration="2.552139936s" podCreationTimestamp="2025-11-22 08:18:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:18:53.550617114 +0000 UTC m=+1347.785017397" watchObservedRunningTime="2025-11-22 08:18:53.552139936 +0000 UTC m=+1347.786540209" Nov 22 08:18:53 crc kubenswrapper[4789]: I1122 08:18:53.981726 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1f547dc8-eab8-420e-af46-30615fd4da38" path="/var/lib/kubelet/pods/1f547dc8-eab8-420e-af46-30615fd4da38/volumes" Nov 22 08:18:54 crc kubenswrapper[4789]: I1122 08:18:54.559663 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"08796aa5-e098-4917-8eb3-dca574b4670a","Type":"ContainerStarted","Data":"e31cb8624b2f37521918d4a58cde08dbb6df0e2b3f29c3df0479fbaf3b71a3e0"} Nov 22 08:18:54 crc kubenswrapper[4789]: I1122 08:18:54.559729 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"08796aa5-e098-4917-8eb3-dca574b4670a","Type":"ContainerStarted","Data":"725dfda767bf2e943a2e5632fba3109331a9370abb6b7cd8ff809b9175abb2d5"} Nov 22 08:18:54 crc kubenswrapper[4789]: I1122 08:18:54.591614 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.591588169 podStartE2EDuration="2.591588169s" podCreationTimestamp="2025-11-22 08:18:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:18:54.589875012 +0000 UTC m=+1348.824275285" watchObservedRunningTime="2025-11-22 08:18:54.591588169 +0000 UTC m=+1348.825988462" Nov 22 08:18:55 crc kubenswrapper[4789]: E1122 08:18:55.029337 4789 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="4bc244b6de6b03bed55cd48a85231ca6f96de9810592b5ee76f208984a87adc0" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 22 08:18:55 crc kubenswrapper[4789]: E1122 08:18:55.030728 4789 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="4bc244b6de6b03bed55cd48a85231ca6f96de9810592b5ee76f208984a87adc0" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 22 08:18:55 crc kubenswrapper[4789]: E1122 08:18:55.031795 4789 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="4bc244b6de6b03bed55cd48a85231ca6f96de9810592b5ee76f208984a87adc0" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 22 08:18:55 crc kubenswrapper[4789]: E1122 08:18:55.031878 4789 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="9cb3ad54-8a01-43df-b938-b6e68bcb704c" containerName="nova-scheduler-scheduler" Nov 22 08:18:56 crc kubenswrapper[4789]: I1122 08:18:56.023142 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 22 08:18:56 crc kubenswrapper[4789]: I1122 08:18:56.060668 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lcnnm\" (UniqueName: \"kubernetes.io/projected/9cb3ad54-8a01-43df-b938-b6e68bcb704c-kube-api-access-lcnnm\") pod \"9cb3ad54-8a01-43df-b938-b6e68bcb704c\" (UID: \"9cb3ad54-8a01-43df-b938-b6e68bcb704c\") " Nov 22 08:18:56 crc kubenswrapper[4789]: I1122 08:18:56.060850 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9cb3ad54-8a01-43df-b938-b6e68bcb704c-config-data\") pod \"9cb3ad54-8a01-43df-b938-b6e68bcb704c\" (UID: \"9cb3ad54-8a01-43df-b938-b6e68bcb704c\") " Nov 22 08:18:56 crc kubenswrapper[4789]: I1122 08:18:56.060988 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cb3ad54-8a01-43df-b938-b6e68bcb704c-combined-ca-bundle\") pod \"9cb3ad54-8a01-43df-b938-b6e68bcb704c\" (UID: \"9cb3ad54-8a01-43df-b938-b6e68bcb704c\") " Nov 22 08:18:56 crc kubenswrapper[4789]: I1122 08:18:56.067578 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9cb3ad54-8a01-43df-b938-b6e68bcb704c-kube-api-access-lcnnm" (OuterVolumeSpecName: "kube-api-access-lcnnm") pod "9cb3ad54-8a01-43df-b938-b6e68bcb704c" (UID: "9cb3ad54-8a01-43df-b938-b6e68bcb704c"). InnerVolumeSpecName "kube-api-access-lcnnm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:18:56 crc kubenswrapper[4789]: I1122 08:18:56.092890 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9cb3ad54-8a01-43df-b938-b6e68bcb704c-config-data" (OuterVolumeSpecName: "config-data") pod "9cb3ad54-8a01-43df-b938-b6e68bcb704c" (UID: "9cb3ad54-8a01-43df-b938-b6e68bcb704c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:18:56 crc kubenswrapper[4789]: I1122 08:18:56.095703 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9cb3ad54-8a01-43df-b938-b6e68bcb704c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9cb3ad54-8a01-43df-b938-b6e68bcb704c" (UID: "9cb3ad54-8a01-43df-b938-b6e68bcb704c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:18:56 crc kubenswrapper[4789]: I1122 08:18:56.163516 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lcnnm\" (UniqueName: \"kubernetes.io/projected/9cb3ad54-8a01-43df-b938-b6e68bcb704c-kube-api-access-lcnnm\") on node \"crc\" DevicePath \"\"" Nov 22 08:18:56 crc kubenswrapper[4789]: I1122 08:18:56.163547 4789 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9cb3ad54-8a01-43df-b938-b6e68bcb704c-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:18:56 crc kubenswrapper[4789]: I1122 08:18:56.163556 4789 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cb3ad54-8a01-43df-b938-b6e68bcb704c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:18:56 crc kubenswrapper[4789]: I1122 08:18:56.494420 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 22 08:18:56 crc kubenswrapper[4789]: I1122 08:18:56.575633 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e4eaef7-30a8-4a62-bb67-9aafc8f52b76-combined-ca-bundle\") pod \"1e4eaef7-30a8-4a62-bb67-9aafc8f52b76\" (UID: \"1e4eaef7-30a8-4a62-bb67-9aafc8f52b76\") " Nov 22 08:18:56 crc kubenswrapper[4789]: I1122 08:18:56.575871 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1e4eaef7-30a8-4a62-bb67-9aafc8f52b76-logs\") pod \"1e4eaef7-30a8-4a62-bb67-9aafc8f52b76\" (UID: \"1e4eaef7-30a8-4a62-bb67-9aafc8f52b76\") " Nov 22 08:18:56 crc kubenswrapper[4789]: I1122 08:18:56.575993 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lgcp8\" (UniqueName: \"kubernetes.io/projected/1e4eaef7-30a8-4a62-bb67-9aafc8f52b76-kube-api-access-lgcp8\") pod \"1e4eaef7-30a8-4a62-bb67-9aafc8f52b76\" (UID: \"1e4eaef7-30a8-4a62-bb67-9aafc8f52b76\") " Nov 22 08:18:56 crc kubenswrapper[4789]: I1122 08:18:56.576024 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e4eaef7-30a8-4a62-bb67-9aafc8f52b76-config-data\") pod \"1e4eaef7-30a8-4a62-bb67-9aafc8f52b76\" (UID: \"1e4eaef7-30a8-4a62-bb67-9aafc8f52b76\") " Nov 22 08:18:56 crc kubenswrapper[4789]: I1122 08:18:56.576839 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1e4eaef7-30a8-4a62-bb67-9aafc8f52b76-logs" (OuterVolumeSpecName: "logs") pod "1e4eaef7-30a8-4a62-bb67-9aafc8f52b76" (UID: "1e4eaef7-30a8-4a62-bb67-9aafc8f52b76"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:18:56 crc kubenswrapper[4789]: I1122 08:18:56.577951 4789 generic.go:334] "Generic (PLEG): container finished" podID="1e4eaef7-30a8-4a62-bb67-9aafc8f52b76" containerID="3cb2ac3c4243a5a7bc7f0d9e9e4c7f1a499314f692dbb42de7c8edaf4c703b4f" exitCode=0 Nov 22 08:18:56 crc kubenswrapper[4789]: I1122 08:18:56.578019 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1e4eaef7-30a8-4a62-bb67-9aafc8f52b76","Type":"ContainerDied","Data":"3cb2ac3c4243a5a7bc7f0d9e9e4c7f1a499314f692dbb42de7c8edaf4c703b4f"} Nov 22 08:18:56 crc kubenswrapper[4789]: I1122 08:18:56.578044 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1e4eaef7-30a8-4a62-bb67-9aafc8f52b76","Type":"ContainerDied","Data":"5ffc2d1e0ec6eeaff9400eddc9910feb0d5ef3b342b482a3356811c3dbbd55c6"} Nov 22 08:18:56 crc kubenswrapper[4789]: I1122 08:18:56.578060 4789 scope.go:117] "RemoveContainer" containerID="3cb2ac3c4243a5a7bc7f0d9e9e4c7f1a499314f692dbb42de7c8edaf4c703b4f" Nov 22 08:18:56 crc kubenswrapper[4789]: I1122 08:18:56.578230 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 22 08:18:56 crc kubenswrapper[4789]: I1122 08:18:56.580052 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e4eaef7-30a8-4a62-bb67-9aafc8f52b76-kube-api-access-lgcp8" (OuterVolumeSpecName: "kube-api-access-lgcp8") pod "1e4eaef7-30a8-4a62-bb67-9aafc8f52b76" (UID: "1e4eaef7-30a8-4a62-bb67-9aafc8f52b76"). InnerVolumeSpecName "kube-api-access-lgcp8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:18:56 crc kubenswrapper[4789]: I1122 08:18:56.581095 4789 generic.go:334] "Generic (PLEG): container finished" podID="9cb3ad54-8a01-43df-b938-b6e68bcb704c" containerID="4bc244b6de6b03bed55cd48a85231ca6f96de9810592b5ee76f208984a87adc0" exitCode=0 Nov 22 08:18:56 crc kubenswrapper[4789]: I1122 08:18:56.581154 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"9cb3ad54-8a01-43df-b938-b6e68bcb704c","Type":"ContainerDied","Data":"4bc244b6de6b03bed55cd48a85231ca6f96de9810592b5ee76f208984a87adc0"} Nov 22 08:18:56 crc kubenswrapper[4789]: I1122 08:18:56.581175 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"9cb3ad54-8a01-43df-b938-b6e68bcb704c","Type":"ContainerDied","Data":"d04c84c773790abb1711752e6d6fbb9acf91aff0aaf6d4dd561fc9e8bbae0f3a"} Nov 22 08:18:56 crc kubenswrapper[4789]: I1122 08:18:56.581256 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 22 08:18:56 crc kubenswrapper[4789]: I1122 08:18:56.611106 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e4eaef7-30a8-4a62-bb67-9aafc8f52b76-config-data" (OuterVolumeSpecName: "config-data") pod "1e4eaef7-30a8-4a62-bb67-9aafc8f52b76" (UID: "1e4eaef7-30a8-4a62-bb67-9aafc8f52b76"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:18:56 crc kubenswrapper[4789]: I1122 08:18:56.623026 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e4eaef7-30a8-4a62-bb67-9aafc8f52b76-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1e4eaef7-30a8-4a62-bb67-9aafc8f52b76" (UID: "1e4eaef7-30a8-4a62-bb67-9aafc8f52b76"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:18:56 crc kubenswrapper[4789]: I1122 08:18:56.633712 4789 scope.go:117] "RemoveContainer" containerID="6f051cc0e759f5f1a60b82855e4f3e14affc615ae4b07071c76bea4e60c971c2" Nov 22 08:18:56 crc kubenswrapper[4789]: I1122 08:18:56.648062 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 22 08:18:56 crc kubenswrapper[4789]: I1122 08:18:56.671199 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Nov 22 08:18:56 crc kubenswrapper[4789]: I1122 08:18:56.677897 4789 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1e4eaef7-30a8-4a62-bb67-9aafc8f52b76-logs\") on node \"crc\" DevicePath \"\"" Nov 22 08:18:56 crc kubenswrapper[4789]: I1122 08:18:56.677928 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lgcp8\" (UniqueName: \"kubernetes.io/projected/1e4eaef7-30a8-4a62-bb67-9aafc8f52b76-kube-api-access-lgcp8\") on node \"crc\" DevicePath \"\"" Nov 22 08:18:56 crc kubenswrapper[4789]: I1122 08:18:56.677938 4789 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e4eaef7-30a8-4a62-bb67-9aafc8f52b76-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:18:56 crc kubenswrapper[4789]: I1122 08:18:56.677951 4789 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e4eaef7-30a8-4a62-bb67-9aafc8f52b76-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:18:56 crc kubenswrapper[4789]: I1122 08:18:56.695307 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 22 08:18:56 crc kubenswrapper[4789]: E1122 08:18:56.695702 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e4eaef7-30a8-4a62-bb67-9aafc8f52b76" containerName="nova-api-log" Nov 22 08:18:56 crc kubenswrapper[4789]: I1122 08:18:56.695719 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e4eaef7-30a8-4a62-bb67-9aafc8f52b76" containerName="nova-api-log" Nov 22 08:18:56 crc kubenswrapper[4789]: E1122 08:18:56.695728 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e4eaef7-30a8-4a62-bb67-9aafc8f52b76" containerName="nova-api-api" Nov 22 08:18:56 crc kubenswrapper[4789]: I1122 08:18:56.695736 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e4eaef7-30a8-4a62-bb67-9aafc8f52b76" containerName="nova-api-api" Nov 22 08:18:56 crc kubenswrapper[4789]: E1122 08:18:56.695807 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cb3ad54-8a01-43df-b938-b6e68bcb704c" containerName="nova-scheduler-scheduler" Nov 22 08:18:56 crc kubenswrapper[4789]: I1122 08:18:56.695814 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cb3ad54-8a01-43df-b938-b6e68bcb704c" containerName="nova-scheduler-scheduler" Nov 22 08:18:56 crc kubenswrapper[4789]: I1122 08:18:56.695981 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e4eaef7-30a8-4a62-bb67-9aafc8f52b76" containerName="nova-api-api" Nov 22 08:18:56 crc kubenswrapper[4789]: I1122 08:18:56.696000 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="9cb3ad54-8a01-43df-b938-b6e68bcb704c" containerName="nova-scheduler-scheduler" Nov 22 08:18:56 crc kubenswrapper[4789]: I1122 08:18:56.696024 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e4eaef7-30a8-4a62-bb67-9aafc8f52b76" containerName="nova-api-log" Nov 22 08:18:56 crc kubenswrapper[4789]: I1122 08:18:56.696880 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 22 08:18:56 crc kubenswrapper[4789]: I1122 08:18:56.701614 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 22 08:18:56 crc kubenswrapper[4789]: I1122 08:18:56.708999 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 22 08:18:56 crc kubenswrapper[4789]: I1122 08:18:56.716955 4789 scope.go:117] "RemoveContainer" containerID="3cb2ac3c4243a5a7bc7f0d9e9e4c7f1a499314f692dbb42de7c8edaf4c703b4f" Nov 22 08:18:56 crc kubenswrapper[4789]: E1122 08:18:56.717612 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3cb2ac3c4243a5a7bc7f0d9e9e4c7f1a499314f692dbb42de7c8edaf4c703b4f\": container with ID starting with 3cb2ac3c4243a5a7bc7f0d9e9e4c7f1a499314f692dbb42de7c8edaf4c703b4f not found: ID does not exist" containerID="3cb2ac3c4243a5a7bc7f0d9e9e4c7f1a499314f692dbb42de7c8edaf4c703b4f" Nov 22 08:18:56 crc kubenswrapper[4789]: I1122 08:18:56.717660 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3cb2ac3c4243a5a7bc7f0d9e9e4c7f1a499314f692dbb42de7c8edaf4c703b4f"} err="failed to get container status \"3cb2ac3c4243a5a7bc7f0d9e9e4c7f1a499314f692dbb42de7c8edaf4c703b4f\": rpc error: code = NotFound desc = could not find container \"3cb2ac3c4243a5a7bc7f0d9e9e4c7f1a499314f692dbb42de7c8edaf4c703b4f\": container with ID starting with 3cb2ac3c4243a5a7bc7f0d9e9e4c7f1a499314f692dbb42de7c8edaf4c703b4f not found: ID does not exist" Nov 22 08:18:56 crc kubenswrapper[4789]: I1122 08:18:56.717688 4789 scope.go:117] "RemoveContainer" containerID="6f051cc0e759f5f1a60b82855e4f3e14affc615ae4b07071c76bea4e60c971c2" Nov 22 08:18:56 crc kubenswrapper[4789]: E1122 08:18:56.718663 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6f051cc0e759f5f1a60b82855e4f3e14affc615ae4b07071c76bea4e60c971c2\": container with ID starting with 6f051cc0e759f5f1a60b82855e4f3e14affc615ae4b07071c76bea4e60c971c2 not found: ID does not exist" containerID="6f051cc0e759f5f1a60b82855e4f3e14affc615ae4b07071c76bea4e60c971c2" Nov 22 08:18:56 crc kubenswrapper[4789]: I1122 08:18:56.718739 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f051cc0e759f5f1a60b82855e4f3e14affc615ae4b07071c76bea4e60c971c2"} err="failed to get container status \"6f051cc0e759f5f1a60b82855e4f3e14affc615ae4b07071c76bea4e60c971c2\": rpc error: code = NotFound desc = could not find container \"6f051cc0e759f5f1a60b82855e4f3e14affc615ae4b07071c76bea4e60c971c2\": container with ID starting with 6f051cc0e759f5f1a60b82855e4f3e14affc615ae4b07071c76bea4e60c971c2 not found: ID does not exist" Nov 22 08:18:56 crc kubenswrapper[4789]: I1122 08:18:56.718811 4789 scope.go:117] "RemoveContainer" containerID="4bc244b6de6b03bed55cd48a85231ca6f96de9810592b5ee76f208984a87adc0" Nov 22 08:18:56 crc kubenswrapper[4789]: I1122 08:18:56.742369 4789 scope.go:117] "RemoveContainer" containerID="4bc244b6de6b03bed55cd48a85231ca6f96de9810592b5ee76f208984a87adc0" Nov 22 08:18:56 crc kubenswrapper[4789]: E1122 08:18:56.742844 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4bc244b6de6b03bed55cd48a85231ca6f96de9810592b5ee76f208984a87adc0\": container with ID starting with 4bc244b6de6b03bed55cd48a85231ca6f96de9810592b5ee76f208984a87adc0 not found: ID does not exist" containerID="4bc244b6de6b03bed55cd48a85231ca6f96de9810592b5ee76f208984a87adc0" Nov 22 08:18:56 crc kubenswrapper[4789]: I1122 08:18:56.742891 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4bc244b6de6b03bed55cd48a85231ca6f96de9810592b5ee76f208984a87adc0"} err="failed to get container status \"4bc244b6de6b03bed55cd48a85231ca6f96de9810592b5ee76f208984a87adc0\": rpc error: code = NotFound desc = could not find container \"4bc244b6de6b03bed55cd48a85231ca6f96de9810592b5ee76f208984a87adc0\": container with ID starting with 4bc244b6de6b03bed55cd48a85231ca6f96de9810592b5ee76f208984a87adc0 not found: ID does not exist" Nov 22 08:18:56 crc kubenswrapper[4789]: I1122 08:18:56.779259 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tff8g\" (UniqueName: \"kubernetes.io/projected/88a4d40a-bffd-416e-a813-a2dd3ddcf47c-kube-api-access-tff8g\") pod \"nova-scheduler-0\" (UID: \"88a4d40a-bffd-416e-a813-a2dd3ddcf47c\") " pod="openstack/nova-scheduler-0" Nov 22 08:18:56 crc kubenswrapper[4789]: I1122 08:18:56.779330 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88a4d40a-bffd-416e-a813-a2dd3ddcf47c-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"88a4d40a-bffd-416e-a813-a2dd3ddcf47c\") " pod="openstack/nova-scheduler-0" Nov 22 08:18:56 crc kubenswrapper[4789]: I1122 08:18:56.779494 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88a4d40a-bffd-416e-a813-a2dd3ddcf47c-config-data\") pod \"nova-scheduler-0\" (UID: \"88a4d40a-bffd-416e-a813-a2dd3ddcf47c\") " pod="openstack/nova-scheduler-0" Nov 22 08:18:56 crc kubenswrapper[4789]: I1122 08:18:56.882231 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88a4d40a-bffd-416e-a813-a2dd3ddcf47c-config-data\") pod \"nova-scheduler-0\" (UID: \"88a4d40a-bffd-416e-a813-a2dd3ddcf47c\") " pod="openstack/nova-scheduler-0" Nov 22 08:18:56 crc kubenswrapper[4789]: I1122 08:18:56.882406 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tff8g\" (UniqueName: \"kubernetes.io/projected/88a4d40a-bffd-416e-a813-a2dd3ddcf47c-kube-api-access-tff8g\") pod \"nova-scheduler-0\" (UID: \"88a4d40a-bffd-416e-a813-a2dd3ddcf47c\") " pod="openstack/nova-scheduler-0" Nov 22 08:18:56 crc kubenswrapper[4789]: I1122 08:18:56.882443 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88a4d40a-bffd-416e-a813-a2dd3ddcf47c-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"88a4d40a-bffd-416e-a813-a2dd3ddcf47c\") " pod="openstack/nova-scheduler-0" Nov 22 08:18:56 crc kubenswrapper[4789]: I1122 08:18:56.886007 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88a4d40a-bffd-416e-a813-a2dd3ddcf47c-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"88a4d40a-bffd-416e-a813-a2dd3ddcf47c\") " pod="openstack/nova-scheduler-0" Nov 22 08:18:56 crc kubenswrapper[4789]: I1122 08:18:56.885999 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88a4d40a-bffd-416e-a813-a2dd3ddcf47c-config-data\") pod \"nova-scheduler-0\" (UID: \"88a4d40a-bffd-416e-a813-a2dd3ddcf47c\") " pod="openstack/nova-scheduler-0" Nov 22 08:18:56 crc kubenswrapper[4789]: I1122 08:18:56.901043 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tff8g\" (UniqueName: \"kubernetes.io/projected/88a4d40a-bffd-416e-a813-a2dd3ddcf47c-kube-api-access-tff8g\") pod \"nova-scheduler-0\" (UID: \"88a4d40a-bffd-416e-a813-a2dd3ddcf47c\") " pod="openstack/nova-scheduler-0" Nov 22 08:18:56 crc kubenswrapper[4789]: I1122 08:18:56.978564 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 22 08:18:56 crc kubenswrapper[4789]: I1122 08:18:56.986256 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 22 08:18:56 crc kubenswrapper[4789]: I1122 08:18:56.994937 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 22 08:18:56 crc kubenswrapper[4789]: I1122 08:18:56.996343 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 22 08:18:56 crc kubenswrapper[4789]: I1122 08:18:56.999320 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 22 08:18:57 crc kubenswrapper[4789]: I1122 08:18:57.012379 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 22 08:18:57 crc kubenswrapper[4789]: I1122 08:18:57.018157 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 22 08:18:57 crc kubenswrapper[4789]: I1122 08:18:57.085800 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1b3c74bc-088d-4ffa-ae56-933e93be3414-logs\") pod \"nova-api-0\" (UID: \"1b3c74bc-088d-4ffa-ae56-933e93be3414\") " pod="openstack/nova-api-0" Nov 22 08:18:57 crc kubenswrapper[4789]: I1122 08:18:57.086111 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b3c74bc-088d-4ffa-ae56-933e93be3414-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"1b3c74bc-088d-4ffa-ae56-933e93be3414\") " pod="openstack/nova-api-0" Nov 22 08:18:57 crc kubenswrapper[4789]: I1122 08:18:57.086150 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b3c74bc-088d-4ffa-ae56-933e93be3414-config-data\") pod \"nova-api-0\" (UID: \"1b3c74bc-088d-4ffa-ae56-933e93be3414\") " pod="openstack/nova-api-0" Nov 22 08:18:57 crc kubenswrapper[4789]: I1122 08:18:57.086238 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q46wb\" (UniqueName: \"kubernetes.io/projected/1b3c74bc-088d-4ffa-ae56-933e93be3414-kube-api-access-q46wb\") pod \"nova-api-0\" (UID: \"1b3c74bc-088d-4ffa-ae56-933e93be3414\") " pod="openstack/nova-api-0" Nov 22 08:18:57 crc kubenswrapper[4789]: I1122 08:18:57.188823 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1b3c74bc-088d-4ffa-ae56-933e93be3414-logs\") pod \"nova-api-0\" (UID: \"1b3c74bc-088d-4ffa-ae56-933e93be3414\") " pod="openstack/nova-api-0" Nov 22 08:18:57 crc kubenswrapper[4789]: I1122 08:18:57.188973 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b3c74bc-088d-4ffa-ae56-933e93be3414-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"1b3c74bc-088d-4ffa-ae56-933e93be3414\") " pod="openstack/nova-api-0" Nov 22 08:18:57 crc kubenswrapper[4789]: I1122 08:18:57.189019 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b3c74bc-088d-4ffa-ae56-933e93be3414-config-data\") pod \"nova-api-0\" (UID: \"1b3c74bc-088d-4ffa-ae56-933e93be3414\") " pod="openstack/nova-api-0" Nov 22 08:18:57 crc kubenswrapper[4789]: I1122 08:18:57.189103 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q46wb\" (UniqueName: \"kubernetes.io/projected/1b3c74bc-088d-4ffa-ae56-933e93be3414-kube-api-access-q46wb\") pod \"nova-api-0\" (UID: \"1b3c74bc-088d-4ffa-ae56-933e93be3414\") " pod="openstack/nova-api-0" Nov 22 08:18:57 crc kubenswrapper[4789]: I1122 08:18:57.189484 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1b3c74bc-088d-4ffa-ae56-933e93be3414-logs\") pod \"nova-api-0\" (UID: \"1b3c74bc-088d-4ffa-ae56-933e93be3414\") " pod="openstack/nova-api-0" Nov 22 08:18:57 crc kubenswrapper[4789]: I1122 08:18:57.194241 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b3c74bc-088d-4ffa-ae56-933e93be3414-config-data\") pod \"nova-api-0\" (UID: \"1b3c74bc-088d-4ffa-ae56-933e93be3414\") " pod="openstack/nova-api-0" Nov 22 08:18:57 crc kubenswrapper[4789]: I1122 08:18:57.194820 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b3c74bc-088d-4ffa-ae56-933e93be3414-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"1b3c74bc-088d-4ffa-ae56-933e93be3414\") " pod="openstack/nova-api-0" Nov 22 08:18:57 crc kubenswrapper[4789]: I1122 08:18:57.211687 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q46wb\" (UniqueName: \"kubernetes.io/projected/1b3c74bc-088d-4ffa-ae56-933e93be3414-kube-api-access-q46wb\") pod \"nova-api-0\" (UID: \"1b3c74bc-088d-4ffa-ae56-933e93be3414\") " pod="openstack/nova-api-0" Nov 22 08:18:57 crc kubenswrapper[4789]: I1122 08:18:57.318103 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 22 08:18:57 crc kubenswrapper[4789]: I1122 08:18:57.450370 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 22 08:18:57 crc kubenswrapper[4789]: I1122 08:18:57.592295 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"88a4d40a-bffd-416e-a813-a2dd3ddcf47c","Type":"ContainerStarted","Data":"aa984d4d1a30af98cecf4aa9492ab6e6e910fc2f1b87761a0cd618d94808551a"} Nov 22 08:18:57 crc kubenswrapper[4789]: I1122 08:18:57.769619 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 22 08:18:57 crc kubenswrapper[4789]: W1122 08:18:57.771346 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1b3c74bc_088d_4ffa_ae56_933e93be3414.slice/crio-21eba1bbd95d8b47073662a8abea1541b3925bde0c74c6f565d1a23735b140da WatchSource:0}: Error finding container 21eba1bbd95d8b47073662a8abea1541b3925bde0c74c6f565d1a23735b140da: Status 404 returned error can't find the container with id 21eba1bbd95d8b47073662a8abea1541b3925bde0c74c6f565d1a23735b140da Nov 22 08:18:57 crc kubenswrapper[4789]: I1122 08:18:57.936081 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 22 08:18:57 crc kubenswrapper[4789]: I1122 08:18:57.936142 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 22 08:18:57 crc kubenswrapper[4789]: I1122 08:18:57.976681 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1e4eaef7-30a8-4a62-bb67-9aafc8f52b76" path="/var/lib/kubelet/pods/1e4eaef7-30a8-4a62-bb67-9aafc8f52b76/volumes" Nov 22 08:18:57 crc kubenswrapper[4789]: I1122 08:18:57.977505 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9cb3ad54-8a01-43df-b938-b6e68bcb704c" path="/var/lib/kubelet/pods/9cb3ad54-8a01-43df-b938-b6e68bcb704c/volumes" Nov 22 08:18:58 crc kubenswrapper[4789]: I1122 08:18:58.606512 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"88a4d40a-bffd-416e-a813-a2dd3ddcf47c","Type":"ContainerStarted","Data":"f984fc3abdac330f1fc81a8a4c0e461ee37ef7101c674d47b577b6b4b64761f9"} Nov 22 08:18:58 crc kubenswrapper[4789]: I1122 08:18:58.613500 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1b3c74bc-088d-4ffa-ae56-933e93be3414","Type":"ContainerStarted","Data":"02def008793adfc97f8a0f02c99797d3bc4826e9dbecdf67ddbc48b3937a54cc"} Nov 22 08:18:58 crc kubenswrapper[4789]: I1122 08:18:58.613546 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1b3c74bc-088d-4ffa-ae56-933e93be3414","Type":"ContainerStarted","Data":"d1ffc76583baf25f8c868b7ccd9e18ffeb0bf395b6f644e070f422339b0eb730"} Nov 22 08:18:58 crc kubenswrapper[4789]: I1122 08:18:58.613561 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1b3c74bc-088d-4ffa-ae56-933e93be3414","Type":"ContainerStarted","Data":"21eba1bbd95d8b47073662a8abea1541b3925bde0c74c6f565d1a23735b140da"} Nov 22 08:18:58 crc kubenswrapper[4789]: I1122 08:18:58.642426 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.642408331 podStartE2EDuration="2.642408331s" podCreationTimestamp="2025-11-22 08:18:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:18:58.623027353 +0000 UTC m=+1352.857427626" watchObservedRunningTime="2025-11-22 08:18:58.642408331 +0000 UTC m=+1352.876808604" Nov 22 08:18:58 crc kubenswrapper[4789]: I1122 08:18:58.647054 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.647035638 podStartE2EDuration="2.647035638s" podCreationTimestamp="2025-11-22 08:18:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:18:58.641315532 +0000 UTC m=+1352.875715815" watchObservedRunningTime="2025-11-22 08:18:58.647035638 +0000 UTC m=+1352.881435931" Nov 22 08:19:01 crc kubenswrapper[4789]: I1122 08:19:01.922619 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Nov 22 08:19:02 crc kubenswrapper[4789]: I1122 08:19:02.019098 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 22 08:19:02 crc kubenswrapper[4789]: I1122 08:19:02.936954 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 22 08:19:02 crc kubenswrapper[4789]: I1122 08:19:02.937022 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 22 08:19:03 crc kubenswrapper[4789]: I1122 08:19:03.946992 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="08796aa5-e098-4917-8eb3-dca574b4670a" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.181:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 22 08:19:03 crc kubenswrapper[4789]: I1122 08:19:03.947081 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="08796aa5-e098-4917-8eb3-dca574b4670a" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.181:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 22 08:19:05 crc kubenswrapper[4789]: I1122 08:19:05.372718 4789 patch_prober.go:28] interesting pod/machine-config-daemon-pxvrm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 08:19:05 crc kubenswrapper[4789]: I1122 08:19:05.373104 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 08:19:07 crc kubenswrapper[4789]: I1122 08:19:07.019830 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 22 08:19:07 crc kubenswrapper[4789]: I1122 08:19:07.044146 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 22 08:19:07 crc kubenswrapper[4789]: I1122 08:19:07.318969 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 22 08:19:07 crc kubenswrapper[4789]: I1122 08:19:07.319256 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 22 08:19:07 crc kubenswrapper[4789]: I1122 08:19:07.725382 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 22 08:19:08 crc kubenswrapper[4789]: I1122 08:19:08.402018 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="1b3c74bc-088d-4ffa-ae56-933e93be3414" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.183:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 22 08:19:08 crc kubenswrapper[4789]: I1122 08:19:08.402033 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="1b3c74bc-088d-4ffa-ae56-933e93be3414" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.183:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 22 08:19:12 crc kubenswrapper[4789]: I1122 08:19:12.943309 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 22 08:19:12 crc kubenswrapper[4789]: I1122 08:19:12.944006 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 22 08:19:12 crc kubenswrapper[4789]: I1122 08:19:12.949475 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 22 08:19:12 crc kubenswrapper[4789]: I1122 08:19:12.950023 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 22 08:19:15 crc kubenswrapper[4789]: I1122 08:19:15.765281 4789 generic.go:334] "Generic (PLEG): container finished" podID="e012dd92-8842-4e20-beca-54115a7c933e" containerID="b84bea1d6e8d82a040a02e9f4861206802c1408e3e270e86181e6fa19997b2a4" exitCode=137 Nov 22 08:19:15 crc kubenswrapper[4789]: I1122 08:19:15.765834 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"e012dd92-8842-4e20-beca-54115a7c933e","Type":"ContainerDied","Data":"b84bea1d6e8d82a040a02e9f4861206802c1408e3e270e86181e6fa19997b2a4"} Nov 22 08:19:15 crc kubenswrapper[4789]: I1122 08:19:15.765862 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"e012dd92-8842-4e20-beca-54115a7c933e","Type":"ContainerDied","Data":"0d170139d4b7ce975ed2ac1b8cbe87dfef8512dfae7c681824d624cca08b758b"} Nov 22 08:19:15 crc kubenswrapper[4789]: I1122 08:19:15.765873 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0d170139d4b7ce975ed2ac1b8cbe87dfef8512dfae7c681824d624cca08b758b" Nov 22 08:19:15 crc kubenswrapper[4789]: I1122 08:19:15.771129 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 22 08:19:15 crc kubenswrapper[4789]: I1122 08:19:15.915559 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e012dd92-8842-4e20-beca-54115a7c933e-config-data\") pod \"e012dd92-8842-4e20-beca-54115a7c933e\" (UID: \"e012dd92-8842-4e20-beca-54115a7c933e\") " Nov 22 08:19:15 crc kubenswrapper[4789]: I1122 08:19:15.915851 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pd4g6\" (UniqueName: \"kubernetes.io/projected/e012dd92-8842-4e20-beca-54115a7c933e-kube-api-access-pd4g6\") pod \"e012dd92-8842-4e20-beca-54115a7c933e\" (UID: \"e012dd92-8842-4e20-beca-54115a7c933e\") " Nov 22 08:19:15 crc kubenswrapper[4789]: I1122 08:19:15.915917 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e012dd92-8842-4e20-beca-54115a7c933e-combined-ca-bundle\") pod \"e012dd92-8842-4e20-beca-54115a7c933e\" (UID: \"e012dd92-8842-4e20-beca-54115a7c933e\") " Nov 22 08:19:15 crc kubenswrapper[4789]: I1122 08:19:15.922597 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e012dd92-8842-4e20-beca-54115a7c933e-kube-api-access-pd4g6" (OuterVolumeSpecName: "kube-api-access-pd4g6") pod "e012dd92-8842-4e20-beca-54115a7c933e" (UID: "e012dd92-8842-4e20-beca-54115a7c933e"). InnerVolumeSpecName "kube-api-access-pd4g6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:19:15 crc kubenswrapper[4789]: I1122 08:19:15.945294 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e012dd92-8842-4e20-beca-54115a7c933e-config-data" (OuterVolumeSpecName: "config-data") pod "e012dd92-8842-4e20-beca-54115a7c933e" (UID: "e012dd92-8842-4e20-beca-54115a7c933e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:19:15 crc kubenswrapper[4789]: I1122 08:19:15.946447 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e012dd92-8842-4e20-beca-54115a7c933e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e012dd92-8842-4e20-beca-54115a7c933e" (UID: "e012dd92-8842-4e20-beca-54115a7c933e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:19:16 crc kubenswrapper[4789]: I1122 08:19:16.018233 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pd4g6\" (UniqueName: \"kubernetes.io/projected/e012dd92-8842-4e20-beca-54115a7c933e-kube-api-access-pd4g6\") on node \"crc\" DevicePath \"\"" Nov 22 08:19:16 crc kubenswrapper[4789]: I1122 08:19:16.018262 4789 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e012dd92-8842-4e20-beca-54115a7c933e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:19:16 crc kubenswrapper[4789]: I1122 08:19:16.018274 4789 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e012dd92-8842-4e20-beca-54115a7c933e-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:19:16 crc kubenswrapper[4789]: I1122 08:19:16.773308 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 22 08:19:16 crc kubenswrapper[4789]: I1122 08:19:16.794514 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 22 08:19:16 crc kubenswrapper[4789]: I1122 08:19:16.801388 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 22 08:19:16 crc kubenswrapper[4789]: I1122 08:19:16.817003 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 22 08:19:16 crc kubenswrapper[4789]: E1122 08:19:16.817372 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e012dd92-8842-4e20-beca-54115a7c933e" containerName="nova-cell1-novncproxy-novncproxy" Nov 22 08:19:16 crc kubenswrapper[4789]: I1122 08:19:16.817391 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="e012dd92-8842-4e20-beca-54115a7c933e" containerName="nova-cell1-novncproxy-novncproxy" Nov 22 08:19:16 crc kubenswrapper[4789]: I1122 08:19:16.817555 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="e012dd92-8842-4e20-beca-54115a7c933e" containerName="nova-cell1-novncproxy-novncproxy" Nov 22 08:19:16 crc kubenswrapper[4789]: I1122 08:19:16.818144 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 22 08:19:16 crc kubenswrapper[4789]: I1122 08:19:16.820048 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Nov 22 08:19:16 crc kubenswrapper[4789]: I1122 08:19:16.820518 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Nov 22 08:19:16 crc kubenswrapper[4789]: I1122 08:19:16.820725 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Nov 22 08:19:16 crc kubenswrapper[4789]: I1122 08:19:16.831057 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 22 08:19:16 crc kubenswrapper[4789]: I1122 08:19:16.932816 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/99fff521-c886-48d2-8858-1ad7fb4a33f2-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"99fff521-c886-48d2-8858-1ad7fb4a33f2\") " pod="openstack/nova-cell1-novncproxy-0" Nov 22 08:19:16 crc kubenswrapper[4789]: I1122 08:19:16.933108 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xdhr7\" (UniqueName: \"kubernetes.io/projected/99fff521-c886-48d2-8858-1ad7fb4a33f2-kube-api-access-xdhr7\") pod \"nova-cell1-novncproxy-0\" (UID: \"99fff521-c886-48d2-8858-1ad7fb4a33f2\") " pod="openstack/nova-cell1-novncproxy-0" Nov 22 08:19:16 crc kubenswrapper[4789]: I1122 08:19:16.933224 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99fff521-c886-48d2-8858-1ad7fb4a33f2-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"99fff521-c886-48d2-8858-1ad7fb4a33f2\") " pod="openstack/nova-cell1-novncproxy-0" Nov 22 08:19:16 crc kubenswrapper[4789]: I1122 08:19:16.933368 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/99fff521-c886-48d2-8858-1ad7fb4a33f2-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"99fff521-c886-48d2-8858-1ad7fb4a33f2\") " pod="openstack/nova-cell1-novncproxy-0" Nov 22 08:19:16 crc kubenswrapper[4789]: I1122 08:19:16.933501 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99fff521-c886-48d2-8858-1ad7fb4a33f2-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"99fff521-c886-48d2-8858-1ad7fb4a33f2\") " pod="openstack/nova-cell1-novncproxy-0" Nov 22 08:19:17 crc kubenswrapper[4789]: I1122 08:19:17.034815 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99fff521-c886-48d2-8858-1ad7fb4a33f2-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"99fff521-c886-48d2-8858-1ad7fb4a33f2\") " pod="openstack/nova-cell1-novncproxy-0" Nov 22 08:19:17 crc kubenswrapper[4789]: I1122 08:19:17.034951 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/99fff521-c886-48d2-8858-1ad7fb4a33f2-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"99fff521-c886-48d2-8858-1ad7fb4a33f2\") " pod="openstack/nova-cell1-novncproxy-0" Nov 22 08:19:17 crc kubenswrapper[4789]: I1122 08:19:17.035003 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xdhr7\" (UniqueName: \"kubernetes.io/projected/99fff521-c886-48d2-8858-1ad7fb4a33f2-kube-api-access-xdhr7\") pod \"nova-cell1-novncproxy-0\" (UID: \"99fff521-c886-48d2-8858-1ad7fb4a33f2\") " pod="openstack/nova-cell1-novncproxy-0" Nov 22 08:19:17 crc kubenswrapper[4789]: I1122 08:19:17.035041 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99fff521-c886-48d2-8858-1ad7fb4a33f2-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"99fff521-c886-48d2-8858-1ad7fb4a33f2\") " pod="openstack/nova-cell1-novncproxy-0" Nov 22 08:19:17 crc kubenswrapper[4789]: I1122 08:19:17.035077 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/99fff521-c886-48d2-8858-1ad7fb4a33f2-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"99fff521-c886-48d2-8858-1ad7fb4a33f2\") " pod="openstack/nova-cell1-novncproxy-0" Nov 22 08:19:17 crc kubenswrapper[4789]: I1122 08:19:17.040437 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99fff521-c886-48d2-8858-1ad7fb4a33f2-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"99fff521-c886-48d2-8858-1ad7fb4a33f2\") " pod="openstack/nova-cell1-novncproxy-0" Nov 22 08:19:17 crc kubenswrapper[4789]: I1122 08:19:17.040437 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99fff521-c886-48d2-8858-1ad7fb4a33f2-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"99fff521-c886-48d2-8858-1ad7fb4a33f2\") " pod="openstack/nova-cell1-novncproxy-0" Nov 22 08:19:17 crc kubenswrapper[4789]: I1122 08:19:17.040447 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/99fff521-c886-48d2-8858-1ad7fb4a33f2-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"99fff521-c886-48d2-8858-1ad7fb4a33f2\") " pod="openstack/nova-cell1-novncproxy-0" Nov 22 08:19:17 crc kubenswrapper[4789]: I1122 08:19:17.041523 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/99fff521-c886-48d2-8858-1ad7fb4a33f2-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"99fff521-c886-48d2-8858-1ad7fb4a33f2\") " pod="openstack/nova-cell1-novncproxy-0" Nov 22 08:19:17 crc kubenswrapper[4789]: I1122 08:19:17.054017 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xdhr7\" (UniqueName: \"kubernetes.io/projected/99fff521-c886-48d2-8858-1ad7fb4a33f2-kube-api-access-xdhr7\") pod \"nova-cell1-novncproxy-0\" (UID: \"99fff521-c886-48d2-8858-1ad7fb4a33f2\") " pod="openstack/nova-cell1-novncproxy-0" Nov 22 08:19:17 crc kubenswrapper[4789]: I1122 08:19:17.145704 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 22 08:19:17 crc kubenswrapper[4789]: I1122 08:19:17.323174 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 22 08:19:17 crc kubenswrapper[4789]: I1122 08:19:17.323497 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 22 08:19:17 crc kubenswrapper[4789]: I1122 08:19:17.323954 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 22 08:19:17 crc kubenswrapper[4789]: I1122 08:19:17.324018 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 22 08:19:17 crc kubenswrapper[4789]: I1122 08:19:17.327858 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 22 08:19:17 crc kubenswrapper[4789]: I1122 08:19:17.328074 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 22 08:19:17 crc kubenswrapper[4789]: I1122 08:19:17.496860 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-c9b558957-c7k8k"] Nov 22 08:19:17 crc kubenswrapper[4789]: I1122 08:19:17.498211 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-c9b558957-c7k8k" Nov 22 08:19:17 crc kubenswrapper[4789]: I1122 08:19:17.524923 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-c9b558957-c7k8k"] Nov 22 08:19:17 crc kubenswrapper[4789]: I1122 08:19:17.567614 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 22 08:19:17 crc kubenswrapper[4789]: W1122 08:19:17.582834 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod99fff521_c886_48d2_8858_1ad7fb4a33f2.slice/crio-7b3718edd0cec2fb7f991628375fc27fc81b21a9f56b4189d719b2d23d7823ca WatchSource:0}: Error finding container 7b3718edd0cec2fb7f991628375fc27fc81b21a9f56b4189d719b2d23d7823ca: Status 404 returned error can't find the container with id 7b3718edd0cec2fb7f991628375fc27fc81b21a9f56b4189d719b2d23d7823ca Nov 22 08:19:17 crc kubenswrapper[4789]: I1122 08:19:17.648828 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1899d9cc-6fb5-499f-8a8a-c7fe3b923e96-dns-svc\") pod \"dnsmasq-dns-c9b558957-c7k8k\" (UID: \"1899d9cc-6fb5-499f-8a8a-c7fe3b923e96\") " pod="openstack/dnsmasq-dns-c9b558957-c7k8k" Nov 22 08:19:17 crc kubenswrapper[4789]: I1122 08:19:17.648921 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1899d9cc-6fb5-499f-8a8a-c7fe3b923e96-ovsdbserver-nb\") pod \"dnsmasq-dns-c9b558957-c7k8k\" (UID: \"1899d9cc-6fb5-499f-8a8a-c7fe3b923e96\") " pod="openstack/dnsmasq-dns-c9b558957-c7k8k" Nov 22 08:19:17 crc kubenswrapper[4789]: I1122 08:19:17.648972 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1899d9cc-6fb5-499f-8a8a-c7fe3b923e96-ovsdbserver-sb\") pod \"dnsmasq-dns-c9b558957-c7k8k\" (UID: \"1899d9cc-6fb5-499f-8a8a-c7fe3b923e96\") " pod="openstack/dnsmasq-dns-c9b558957-c7k8k" Nov 22 08:19:17 crc kubenswrapper[4789]: I1122 08:19:17.649052 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1899d9cc-6fb5-499f-8a8a-c7fe3b923e96-config\") pod \"dnsmasq-dns-c9b558957-c7k8k\" (UID: \"1899d9cc-6fb5-499f-8a8a-c7fe3b923e96\") " pod="openstack/dnsmasq-dns-c9b558957-c7k8k" Nov 22 08:19:17 crc kubenswrapper[4789]: I1122 08:19:17.649387 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fg72r\" (UniqueName: \"kubernetes.io/projected/1899d9cc-6fb5-499f-8a8a-c7fe3b923e96-kube-api-access-fg72r\") pod \"dnsmasq-dns-c9b558957-c7k8k\" (UID: \"1899d9cc-6fb5-499f-8a8a-c7fe3b923e96\") " pod="openstack/dnsmasq-dns-c9b558957-c7k8k" Nov 22 08:19:17 crc kubenswrapper[4789]: I1122 08:19:17.750935 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1899d9cc-6fb5-499f-8a8a-c7fe3b923e96-ovsdbserver-sb\") pod \"dnsmasq-dns-c9b558957-c7k8k\" (UID: \"1899d9cc-6fb5-499f-8a8a-c7fe3b923e96\") " pod="openstack/dnsmasq-dns-c9b558957-c7k8k" Nov 22 08:19:17 crc kubenswrapper[4789]: I1122 08:19:17.751052 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1899d9cc-6fb5-499f-8a8a-c7fe3b923e96-config\") pod \"dnsmasq-dns-c9b558957-c7k8k\" (UID: \"1899d9cc-6fb5-499f-8a8a-c7fe3b923e96\") " pod="openstack/dnsmasq-dns-c9b558957-c7k8k" Nov 22 08:19:17 crc kubenswrapper[4789]: I1122 08:19:17.751117 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fg72r\" (UniqueName: \"kubernetes.io/projected/1899d9cc-6fb5-499f-8a8a-c7fe3b923e96-kube-api-access-fg72r\") pod \"dnsmasq-dns-c9b558957-c7k8k\" (UID: \"1899d9cc-6fb5-499f-8a8a-c7fe3b923e96\") " pod="openstack/dnsmasq-dns-c9b558957-c7k8k" Nov 22 08:19:17 crc kubenswrapper[4789]: I1122 08:19:17.751173 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1899d9cc-6fb5-499f-8a8a-c7fe3b923e96-dns-svc\") pod \"dnsmasq-dns-c9b558957-c7k8k\" (UID: \"1899d9cc-6fb5-499f-8a8a-c7fe3b923e96\") " pod="openstack/dnsmasq-dns-c9b558957-c7k8k" Nov 22 08:19:17 crc kubenswrapper[4789]: I1122 08:19:17.751211 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1899d9cc-6fb5-499f-8a8a-c7fe3b923e96-ovsdbserver-nb\") pod \"dnsmasq-dns-c9b558957-c7k8k\" (UID: \"1899d9cc-6fb5-499f-8a8a-c7fe3b923e96\") " pod="openstack/dnsmasq-dns-c9b558957-c7k8k" Nov 22 08:19:17 crc kubenswrapper[4789]: I1122 08:19:17.752850 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1899d9cc-6fb5-499f-8a8a-c7fe3b923e96-config\") pod \"dnsmasq-dns-c9b558957-c7k8k\" (UID: \"1899d9cc-6fb5-499f-8a8a-c7fe3b923e96\") " pod="openstack/dnsmasq-dns-c9b558957-c7k8k" Nov 22 08:19:17 crc kubenswrapper[4789]: I1122 08:19:17.753008 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1899d9cc-6fb5-499f-8a8a-c7fe3b923e96-ovsdbserver-nb\") pod \"dnsmasq-dns-c9b558957-c7k8k\" (UID: \"1899d9cc-6fb5-499f-8a8a-c7fe3b923e96\") " pod="openstack/dnsmasq-dns-c9b558957-c7k8k" Nov 22 08:19:17 crc kubenswrapper[4789]: I1122 08:19:17.753249 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1899d9cc-6fb5-499f-8a8a-c7fe3b923e96-dns-svc\") pod \"dnsmasq-dns-c9b558957-c7k8k\" (UID: \"1899d9cc-6fb5-499f-8a8a-c7fe3b923e96\") " pod="openstack/dnsmasq-dns-c9b558957-c7k8k" Nov 22 08:19:17 crc kubenswrapper[4789]: I1122 08:19:17.756888 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1899d9cc-6fb5-499f-8a8a-c7fe3b923e96-ovsdbserver-sb\") pod \"dnsmasq-dns-c9b558957-c7k8k\" (UID: \"1899d9cc-6fb5-499f-8a8a-c7fe3b923e96\") " pod="openstack/dnsmasq-dns-c9b558957-c7k8k" Nov 22 08:19:17 crc kubenswrapper[4789]: I1122 08:19:17.770489 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fg72r\" (UniqueName: \"kubernetes.io/projected/1899d9cc-6fb5-499f-8a8a-c7fe3b923e96-kube-api-access-fg72r\") pod \"dnsmasq-dns-c9b558957-c7k8k\" (UID: \"1899d9cc-6fb5-499f-8a8a-c7fe3b923e96\") " pod="openstack/dnsmasq-dns-c9b558957-c7k8k" Nov 22 08:19:17 crc kubenswrapper[4789]: I1122 08:19:17.793848 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"99fff521-c886-48d2-8858-1ad7fb4a33f2","Type":"ContainerStarted","Data":"7b3718edd0cec2fb7f991628375fc27fc81b21a9f56b4189d719b2d23d7823ca"} Nov 22 08:19:17 crc kubenswrapper[4789]: I1122 08:19:17.823112 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-c9b558957-c7k8k" Nov 22 08:19:17 crc kubenswrapper[4789]: I1122 08:19:17.981539 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e012dd92-8842-4e20-beca-54115a7c933e" path="/var/lib/kubelet/pods/e012dd92-8842-4e20-beca-54115a7c933e/volumes" Nov 22 08:19:18 crc kubenswrapper[4789]: I1122 08:19:18.299393 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-c9b558957-c7k8k"] Nov 22 08:19:18 crc kubenswrapper[4789]: I1122 08:19:18.799399 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"99fff521-c886-48d2-8858-1ad7fb4a33f2","Type":"ContainerStarted","Data":"58f992e490657b552a90a8b6ce8351f30ff8dd89112f21fd952eb3baa6bb095e"} Nov 22 08:19:18 crc kubenswrapper[4789]: I1122 08:19:18.801912 4789 generic.go:334] "Generic (PLEG): container finished" podID="1899d9cc-6fb5-499f-8a8a-c7fe3b923e96" containerID="298646ec1930685846c3079c5df5819831f8d41c2dcf045279ed940d562894d6" exitCode=0 Nov 22 08:19:18 crc kubenswrapper[4789]: I1122 08:19:18.801983 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-c9b558957-c7k8k" event={"ID":"1899d9cc-6fb5-499f-8a8a-c7fe3b923e96","Type":"ContainerDied","Data":"298646ec1930685846c3079c5df5819831f8d41c2dcf045279ed940d562894d6"} Nov 22 08:19:18 crc kubenswrapper[4789]: I1122 08:19:18.802091 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-c9b558957-c7k8k" event={"ID":"1899d9cc-6fb5-499f-8a8a-c7fe3b923e96","Type":"ContainerStarted","Data":"9ee8362ad000dde85bad9761a8221d23863d0602d9bdf5f0ca85557aedb2fedc"} Nov 22 08:19:18 crc kubenswrapper[4789]: I1122 08:19:18.837003 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.836976969 podStartE2EDuration="2.836976969s" podCreationTimestamp="2025-11-22 08:19:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:19:18.820382757 +0000 UTC m=+1373.054783040" watchObservedRunningTime="2025-11-22 08:19:18.836976969 +0000 UTC m=+1373.071377242" Nov 22 08:19:19 crc kubenswrapper[4789]: I1122 08:19:19.688319 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:19:19 crc kubenswrapper[4789]: I1122 08:19:19.689400 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6ff6994a-950e-4aae-8126-1f55702f7eba" containerName="ceilometer-central-agent" containerID="cri-o://7211383e403d85898341931feec5b06f1e65e2c29b5288f6582db9b01610d7c9" gracePeriod=30 Nov 22 08:19:19 crc kubenswrapper[4789]: I1122 08:19:19.689866 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6ff6994a-950e-4aae-8126-1f55702f7eba" containerName="proxy-httpd" containerID="cri-o://7f2db1070da806a8e0a6409cda690ba4b4cfadd5f9454f9ef89548607b6c2de9" gracePeriod=30 Nov 22 08:19:19 crc kubenswrapper[4789]: I1122 08:19:19.689935 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6ff6994a-950e-4aae-8126-1f55702f7eba" containerName="sg-core" containerID="cri-o://2bfdd608fd028e7cbf93dfab2e22298c8e6476fc7c4897dc406c59747d223d3a" gracePeriod=30 Nov 22 08:19:19 crc kubenswrapper[4789]: I1122 08:19:19.689977 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6ff6994a-950e-4aae-8126-1f55702f7eba" containerName="ceilometer-notification-agent" containerID="cri-o://264d4c5e4620a285e31eeaf8220e3156dbbfc1893c0d8b9f1a01d6947b8a9e7b" gracePeriod=30 Nov 22 08:19:19 crc kubenswrapper[4789]: I1122 08:19:19.734537 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 22 08:19:19 crc kubenswrapper[4789]: I1122 08:19:19.811696 4789 generic.go:334] "Generic (PLEG): container finished" podID="6ff6994a-950e-4aae-8126-1f55702f7eba" containerID="2bfdd608fd028e7cbf93dfab2e22298c8e6476fc7c4897dc406c59747d223d3a" exitCode=2 Nov 22 08:19:19 crc kubenswrapper[4789]: I1122 08:19:19.811776 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6ff6994a-950e-4aae-8126-1f55702f7eba","Type":"ContainerDied","Data":"2bfdd608fd028e7cbf93dfab2e22298c8e6476fc7c4897dc406c59747d223d3a"} Nov 22 08:19:19 crc kubenswrapper[4789]: I1122 08:19:19.813344 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-c9b558957-c7k8k" event={"ID":"1899d9cc-6fb5-499f-8a8a-c7fe3b923e96","Type":"ContainerStarted","Data":"56819ef5f9657982c799cca0b86e2b25a4b3e1c3c10a456999068c3f20196c16"} Nov 22 08:19:19 crc kubenswrapper[4789]: I1122 08:19:19.813514 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="1b3c74bc-088d-4ffa-ae56-933e93be3414" containerName="nova-api-log" containerID="cri-o://d1ffc76583baf25f8c868b7ccd9e18ffeb0bf395b6f644e070f422339b0eb730" gracePeriod=30 Nov 22 08:19:19 crc kubenswrapper[4789]: I1122 08:19:19.813602 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="1b3c74bc-088d-4ffa-ae56-933e93be3414" containerName="nova-api-api" containerID="cri-o://02def008793adfc97f8a0f02c99797d3bc4826e9dbecdf67ddbc48b3937a54cc" gracePeriod=30 Nov 22 08:19:19 crc kubenswrapper[4789]: I1122 08:19:19.839135 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-c9b558957-c7k8k" podStartSLOduration=2.839114706 podStartE2EDuration="2.839114706s" podCreationTimestamp="2025-11-22 08:19:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:19:19.836649549 +0000 UTC m=+1374.071049822" watchObservedRunningTime="2025-11-22 08:19:19.839114706 +0000 UTC m=+1374.073514979" Nov 22 08:19:20 crc kubenswrapper[4789]: I1122 08:19:20.826201 4789 generic.go:334] "Generic (PLEG): container finished" podID="6ff6994a-950e-4aae-8126-1f55702f7eba" containerID="7f2db1070da806a8e0a6409cda690ba4b4cfadd5f9454f9ef89548607b6c2de9" exitCode=0 Nov 22 08:19:20 crc kubenswrapper[4789]: I1122 08:19:20.826238 4789 generic.go:334] "Generic (PLEG): container finished" podID="6ff6994a-950e-4aae-8126-1f55702f7eba" containerID="7211383e403d85898341931feec5b06f1e65e2c29b5288f6582db9b01610d7c9" exitCode=0 Nov 22 08:19:20 crc kubenswrapper[4789]: I1122 08:19:20.826272 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6ff6994a-950e-4aae-8126-1f55702f7eba","Type":"ContainerDied","Data":"7f2db1070da806a8e0a6409cda690ba4b4cfadd5f9454f9ef89548607b6c2de9"} Nov 22 08:19:20 crc kubenswrapper[4789]: I1122 08:19:20.826311 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6ff6994a-950e-4aae-8126-1f55702f7eba","Type":"ContainerDied","Data":"7211383e403d85898341931feec5b06f1e65e2c29b5288f6582db9b01610d7c9"} Nov 22 08:19:20 crc kubenswrapper[4789]: I1122 08:19:20.829657 4789 generic.go:334] "Generic (PLEG): container finished" podID="1b3c74bc-088d-4ffa-ae56-933e93be3414" containerID="d1ffc76583baf25f8c868b7ccd9e18ffeb0bf395b6f644e070f422339b0eb730" exitCode=143 Nov 22 08:19:20 crc kubenswrapper[4789]: I1122 08:19:20.829812 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1b3c74bc-088d-4ffa-ae56-933e93be3414","Type":"ContainerDied","Data":"d1ffc76583baf25f8c868b7ccd9e18ffeb0bf395b6f644e070f422339b0eb730"} Nov 22 08:19:20 crc kubenswrapper[4789]: I1122 08:19:20.829976 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-c9b558957-c7k8k" Nov 22 08:19:21 crc kubenswrapper[4789]: I1122 08:19:21.704373 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 08:19:21 crc kubenswrapper[4789]: I1122 08:19:21.820482 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/6ff6994a-950e-4aae-8126-1f55702f7eba-ceilometer-tls-certs\") pod \"6ff6994a-950e-4aae-8126-1f55702f7eba\" (UID: \"6ff6994a-950e-4aae-8126-1f55702f7eba\") " Nov 22 08:19:21 crc kubenswrapper[4789]: I1122 08:19:21.820905 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h2gvj\" (UniqueName: \"kubernetes.io/projected/6ff6994a-950e-4aae-8126-1f55702f7eba-kube-api-access-h2gvj\") pod \"6ff6994a-950e-4aae-8126-1f55702f7eba\" (UID: \"6ff6994a-950e-4aae-8126-1f55702f7eba\") " Nov 22 08:19:21 crc kubenswrapper[4789]: I1122 08:19:21.821022 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6ff6994a-950e-4aae-8126-1f55702f7eba-run-httpd\") pod \"6ff6994a-950e-4aae-8126-1f55702f7eba\" (UID: \"6ff6994a-950e-4aae-8126-1f55702f7eba\") " Nov 22 08:19:21 crc kubenswrapper[4789]: I1122 08:19:21.821091 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ff6994a-950e-4aae-8126-1f55702f7eba-combined-ca-bundle\") pod \"6ff6994a-950e-4aae-8126-1f55702f7eba\" (UID: \"6ff6994a-950e-4aae-8126-1f55702f7eba\") " Nov 22 08:19:21 crc kubenswrapper[4789]: I1122 08:19:21.821122 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6ff6994a-950e-4aae-8126-1f55702f7eba-log-httpd\") pod \"6ff6994a-950e-4aae-8126-1f55702f7eba\" (UID: \"6ff6994a-950e-4aae-8126-1f55702f7eba\") " Nov 22 08:19:21 crc kubenswrapper[4789]: I1122 08:19:21.821152 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ff6994a-950e-4aae-8126-1f55702f7eba-config-data\") pod \"6ff6994a-950e-4aae-8126-1f55702f7eba\" (UID: \"6ff6994a-950e-4aae-8126-1f55702f7eba\") " Nov 22 08:19:21 crc kubenswrapper[4789]: I1122 08:19:21.821224 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6ff6994a-950e-4aae-8126-1f55702f7eba-scripts\") pod \"6ff6994a-950e-4aae-8126-1f55702f7eba\" (UID: \"6ff6994a-950e-4aae-8126-1f55702f7eba\") " Nov 22 08:19:21 crc kubenswrapper[4789]: I1122 08:19:21.821301 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6ff6994a-950e-4aae-8126-1f55702f7eba-sg-core-conf-yaml\") pod \"6ff6994a-950e-4aae-8126-1f55702f7eba\" (UID: \"6ff6994a-950e-4aae-8126-1f55702f7eba\") " Nov 22 08:19:21 crc kubenswrapper[4789]: I1122 08:19:21.821707 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6ff6994a-950e-4aae-8126-1f55702f7eba-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "6ff6994a-950e-4aae-8126-1f55702f7eba" (UID: "6ff6994a-950e-4aae-8126-1f55702f7eba"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:19:21 crc kubenswrapper[4789]: I1122 08:19:21.821773 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6ff6994a-950e-4aae-8126-1f55702f7eba-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "6ff6994a-950e-4aae-8126-1f55702f7eba" (UID: "6ff6994a-950e-4aae-8126-1f55702f7eba"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:19:21 crc kubenswrapper[4789]: I1122 08:19:21.822021 4789 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6ff6994a-950e-4aae-8126-1f55702f7eba-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 22 08:19:21 crc kubenswrapper[4789]: I1122 08:19:21.822041 4789 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6ff6994a-950e-4aae-8126-1f55702f7eba-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 22 08:19:21 crc kubenswrapper[4789]: I1122 08:19:21.827507 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ff6994a-950e-4aae-8126-1f55702f7eba-kube-api-access-h2gvj" (OuterVolumeSpecName: "kube-api-access-h2gvj") pod "6ff6994a-950e-4aae-8126-1f55702f7eba" (UID: "6ff6994a-950e-4aae-8126-1f55702f7eba"). InnerVolumeSpecName "kube-api-access-h2gvj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:19:21 crc kubenswrapper[4789]: I1122 08:19:21.828691 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ff6994a-950e-4aae-8126-1f55702f7eba-scripts" (OuterVolumeSpecName: "scripts") pod "6ff6994a-950e-4aae-8126-1f55702f7eba" (UID: "6ff6994a-950e-4aae-8126-1f55702f7eba"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:19:21 crc kubenswrapper[4789]: I1122 08:19:21.845922 4789 generic.go:334] "Generic (PLEG): container finished" podID="6ff6994a-950e-4aae-8126-1f55702f7eba" containerID="264d4c5e4620a285e31eeaf8220e3156dbbfc1893c0d8b9f1a01d6947b8a9e7b" exitCode=0 Nov 22 08:19:21 crc kubenswrapper[4789]: I1122 08:19:21.845977 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 08:19:21 crc kubenswrapper[4789]: I1122 08:19:21.845982 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6ff6994a-950e-4aae-8126-1f55702f7eba","Type":"ContainerDied","Data":"264d4c5e4620a285e31eeaf8220e3156dbbfc1893c0d8b9f1a01d6947b8a9e7b"} Nov 22 08:19:21 crc kubenswrapper[4789]: I1122 08:19:21.847262 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6ff6994a-950e-4aae-8126-1f55702f7eba","Type":"ContainerDied","Data":"8b9efd471d7b1a57c2270c235f707634e2d025f454cdebccd3eb354c61e8027b"} Nov 22 08:19:21 crc kubenswrapper[4789]: I1122 08:19:21.847293 4789 scope.go:117] "RemoveContainer" containerID="7f2db1070da806a8e0a6409cda690ba4b4cfadd5f9454f9ef89548607b6c2de9" Nov 22 08:19:21 crc kubenswrapper[4789]: I1122 08:19:21.859406 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ff6994a-950e-4aae-8126-1f55702f7eba-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "6ff6994a-950e-4aae-8126-1f55702f7eba" (UID: "6ff6994a-950e-4aae-8126-1f55702f7eba"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:19:21 crc kubenswrapper[4789]: I1122 08:19:21.875721 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ff6994a-950e-4aae-8126-1f55702f7eba-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "6ff6994a-950e-4aae-8126-1f55702f7eba" (UID: "6ff6994a-950e-4aae-8126-1f55702f7eba"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:19:21 crc kubenswrapper[4789]: I1122 08:19:21.926667 4789 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6ff6994a-950e-4aae-8126-1f55702f7eba-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:19:21 crc kubenswrapper[4789]: I1122 08:19:21.926711 4789 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6ff6994a-950e-4aae-8126-1f55702f7eba-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 22 08:19:21 crc kubenswrapper[4789]: I1122 08:19:21.926725 4789 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/6ff6994a-950e-4aae-8126-1f55702f7eba-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 22 08:19:21 crc kubenswrapper[4789]: I1122 08:19:21.926735 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h2gvj\" (UniqueName: \"kubernetes.io/projected/6ff6994a-950e-4aae-8126-1f55702f7eba-kube-api-access-h2gvj\") on node \"crc\" DevicePath \"\"" Nov 22 08:19:21 crc kubenswrapper[4789]: I1122 08:19:21.928519 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ff6994a-950e-4aae-8126-1f55702f7eba-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6ff6994a-950e-4aae-8126-1f55702f7eba" (UID: "6ff6994a-950e-4aae-8126-1f55702f7eba"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:19:21 crc kubenswrapper[4789]: I1122 08:19:21.937838 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ff6994a-950e-4aae-8126-1f55702f7eba-config-data" (OuterVolumeSpecName: "config-data") pod "6ff6994a-950e-4aae-8126-1f55702f7eba" (UID: "6ff6994a-950e-4aae-8126-1f55702f7eba"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:19:21 crc kubenswrapper[4789]: I1122 08:19:21.968363 4789 scope.go:117] "RemoveContainer" containerID="2bfdd608fd028e7cbf93dfab2e22298c8e6476fc7c4897dc406c59747d223d3a" Nov 22 08:19:22 crc kubenswrapper[4789]: I1122 08:19:22.006710 4789 scope.go:117] "RemoveContainer" containerID="264d4c5e4620a285e31eeaf8220e3156dbbfc1893c0d8b9f1a01d6947b8a9e7b" Nov 22 08:19:22 crc kubenswrapper[4789]: I1122 08:19:22.028502 4789 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ff6994a-950e-4aae-8126-1f55702f7eba-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:19:22 crc kubenswrapper[4789]: I1122 08:19:22.028546 4789 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ff6994a-950e-4aae-8126-1f55702f7eba-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:19:22 crc kubenswrapper[4789]: I1122 08:19:22.036235 4789 scope.go:117] "RemoveContainer" containerID="7211383e403d85898341931feec5b06f1e65e2c29b5288f6582db9b01610d7c9" Nov 22 08:19:22 crc kubenswrapper[4789]: I1122 08:19:22.055915 4789 scope.go:117] "RemoveContainer" containerID="7f2db1070da806a8e0a6409cda690ba4b4cfadd5f9454f9ef89548607b6c2de9" Nov 22 08:19:22 crc kubenswrapper[4789]: E1122 08:19:22.056434 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7f2db1070da806a8e0a6409cda690ba4b4cfadd5f9454f9ef89548607b6c2de9\": container with ID starting with 7f2db1070da806a8e0a6409cda690ba4b4cfadd5f9454f9ef89548607b6c2de9 not found: ID does not exist" containerID="7f2db1070da806a8e0a6409cda690ba4b4cfadd5f9454f9ef89548607b6c2de9" Nov 22 08:19:22 crc kubenswrapper[4789]: I1122 08:19:22.056462 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f2db1070da806a8e0a6409cda690ba4b4cfadd5f9454f9ef89548607b6c2de9"} err="failed to get container status \"7f2db1070da806a8e0a6409cda690ba4b4cfadd5f9454f9ef89548607b6c2de9\": rpc error: code = NotFound desc = could not find container \"7f2db1070da806a8e0a6409cda690ba4b4cfadd5f9454f9ef89548607b6c2de9\": container with ID starting with 7f2db1070da806a8e0a6409cda690ba4b4cfadd5f9454f9ef89548607b6c2de9 not found: ID does not exist" Nov 22 08:19:22 crc kubenswrapper[4789]: I1122 08:19:22.056483 4789 scope.go:117] "RemoveContainer" containerID="2bfdd608fd028e7cbf93dfab2e22298c8e6476fc7c4897dc406c59747d223d3a" Nov 22 08:19:22 crc kubenswrapper[4789]: E1122 08:19:22.056861 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2bfdd608fd028e7cbf93dfab2e22298c8e6476fc7c4897dc406c59747d223d3a\": container with ID starting with 2bfdd608fd028e7cbf93dfab2e22298c8e6476fc7c4897dc406c59747d223d3a not found: ID does not exist" containerID="2bfdd608fd028e7cbf93dfab2e22298c8e6476fc7c4897dc406c59747d223d3a" Nov 22 08:19:22 crc kubenswrapper[4789]: I1122 08:19:22.056977 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2bfdd608fd028e7cbf93dfab2e22298c8e6476fc7c4897dc406c59747d223d3a"} err="failed to get container status \"2bfdd608fd028e7cbf93dfab2e22298c8e6476fc7c4897dc406c59747d223d3a\": rpc error: code = NotFound desc = could not find container \"2bfdd608fd028e7cbf93dfab2e22298c8e6476fc7c4897dc406c59747d223d3a\": container with ID starting with 2bfdd608fd028e7cbf93dfab2e22298c8e6476fc7c4897dc406c59747d223d3a not found: ID does not exist" Nov 22 08:19:22 crc kubenswrapper[4789]: I1122 08:19:22.057067 4789 scope.go:117] "RemoveContainer" containerID="264d4c5e4620a285e31eeaf8220e3156dbbfc1893c0d8b9f1a01d6947b8a9e7b" Nov 22 08:19:22 crc kubenswrapper[4789]: E1122 08:19:22.057429 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"264d4c5e4620a285e31eeaf8220e3156dbbfc1893c0d8b9f1a01d6947b8a9e7b\": container with ID starting with 264d4c5e4620a285e31eeaf8220e3156dbbfc1893c0d8b9f1a01d6947b8a9e7b not found: ID does not exist" containerID="264d4c5e4620a285e31eeaf8220e3156dbbfc1893c0d8b9f1a01d6947b8a9e7b" Nov 22 08:19:22 crc kubenswrapper[4789]: I1122 08:19:22.057454 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"264d4c5e4620a285e31eeaf8220e3156dbbfc1893c0d8b9f1a01d6947b8a9e7b"} err="failed to get container status \"264d4c5e4620a285e31eeaf8220e3156dbbfc1893c0d8b9f1a01d6947b8a9e7b\": rpc error: code = NotFound desc = could not find container \"264d4c5e4620a285e31eeaf8220e3156dbbfc1893c0d8b9f1a01d6947b8a9e7b\": container with ID starting with 264d4c5e4620a285e31eeaf8220e3156dbbfc1893c0d8b9f1a01d6947b8a9e7b not found: ID does not exist" Nov 22 08:19:22 crc kubenswrapper[4789]: I1122 08:19:22.057473 4789 scope.go:117] "RemoveContainer" containerID="7211383e403d85898341931feec5b06f1e65e2c29b5288f6582db9b01610d7c9" Nov 22 08:19:22 crc kubenswrapper[4789]: E1122 08:19:22.057702 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7211383e403d85898341931feec5b06f1e65e2c29b5288f6582db9b01610d7c9\": container with ID starting with 7211383e403d85898341931feec5b06f1e65e2c29b5288f6582db9b01610d7c9 not found: ID does not exist" containerID="7211383e403d85898341931feec5b06f1e65e2c29b5288f6582db9b01610d7c9" Nov 22 08:19:22 crc kubenswrapper[4789]: I1122 08:19:22.057732 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7211383e403d85898341931feec5b06f1e65e2c29b5288f6582db9b01610d7c9"} err="failed to get container status \"7211383e403d85898341931feec5b06f1e65e2c29b5288f6582db9b01610d7c9\": rpc error: code = NotFound desc = could not find container \"7211383e403d85898341931feec5b06f1e65e2c29b5288f6582db9b01610d7c9\": container with ID starting with 7211383e403d85898341931feec5b06f1e65e2c29b5288f6582db9b01610d7c9 not found: ID does not exist" Nov 22 08:19:22 crc kubenswrapper[4789]: I1122 08:19:22.146555 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Nov 22 08:19:22 crc kubenswrapper[4789]: I1122 08:19:22.175298 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:19:22 crc kubenswrapper[4789]: I1122 08:19:22.182982 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:19:22 crc kubenswrapper[4789]: I1122 08:19:22.194363 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:19:22 crc kubenswrapper[4789]: E1122 08:19:22.194979 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ff6994a-950e-4aae-8126-1f55702f7eba" containerName="proxy-httpd" Nov 22 08:19:22 crc kubenswrapper[4789]: I1122 08:19:22.195012 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ff6994a-950e-4aae-8126-1f55702f7eba" containerName="proxy-httpd" Nov 22 08:19:22 crc kubenswrapper[4789]: E1122 08:19:22.195066 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ff6994a-950e-4aae-8126-1f55702f7eba" containerName="ceilometer-central-agent" Nov 22 08:19:22 crc kubenswrapper[4789]: I1122 08:19:22.195075 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ff6994a-950e-4aae-8126-1f55702f7eba" containerName="ceilometer-central-agent" Nov 22 08:19:22 crc kubenswrapper[4789]: E1122 08:19:22.195091 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ff6994a-950e-4aae-8126-1f55702f7eba" containerName="sg-core" Nov 22 08:19:22 crc kubenswrapper[4789]: I1122 08:19:22.195099 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ff6994a-950e-4aae-8126-1f55702f7eba" containerName="sg-core" Nov 22 08:19:22 crc kubenswrapper[4789]: E1122 08:19:22.195114 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ff6994a-950e-4aae-8126-1f55702f7eba" containerName="ceilometer-notification-agent" Nov 22 08:19:22 crc kubenswrapper[4789]: I1122 08:19:22.195123 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ff6994a-950e-4aae-8126-1f55702f7eba" containerName="ceilometer-notification-agent" Nov 22 08:19:22 crc kubenswrapper[4789]: I1122 08:19:22.195384 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ff6994a-950e-4aae-8126-1f55702f7eba" containerName="proxy-httpd" Nov 22 08:19:22 crc kubenswrapper[4789]: I1122 08:19:22.195402 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ff6994a-950e-4aae-8126-1f55702f7eba" containerName="sg-core" Nov 22 08:19:22 crc kubenswrapper[4789]: I1122 08:19:22.195420 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ff6994a-950e-4aae-8126-1f55702f7eba" containerName="ceilometer-central-agent" Nov 22 08:19:22 crc kubenswrapper[4789]: I1122 08:19:22.195437 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ff6994a-950e-4aae-8126-1f55702f7eba" containerName="ceilometer-notification-agent" Nov 22 08:19:22 crc kubenswrapper[4789]: I1122 08:19:22.197579 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 08:19:22 crc kubenswrapper[4789]: I1122 08:19:22.200812 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 22 08:19:22 crc kubenswrapper[4789]: I1122 08:19:22.200883 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 22 08:19:22 crc kubenswrapper[4789]: I1122 08:19:22.200898 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 22 08:19:22 crc kubenswrapper[4789]: I1122 08:19:22.206521 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:19:22 crc kubenswrapper[4789]: I1122 08:19:22.334140 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8ctgf\" (UniqueName: \"kubernetes.io/projected/b3e0f7a4-cde7-4da6-8339-615054807f46-kube-api-access-8ctgf\") pod \"ceilometer-0\" (UID: \"b3e0f7a4-cde7-4da6-8339-615054807f46\") " pod="openstack/ceilometer-0" Nov 22 08:19:22 crc kubenswrapper[4789]: I1122 08:19:22.334271 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3e0f7a4-cde7-4da6-8339-615054807f46-config-data\") pod \"ceilometer-0\" (UID: \"b3e0f7a4-cde7-4da6-8339-615054807f46\") " pod="openstack/ceilometer-0" Nov 22 08:19:22 crc kubenswrapper[4789]: I1122 08:19:22.334324 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3e0f7a4-cde7-4da6-8339-615054807f46-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b3e0f7a4-cde7-4da6-8339-615054807f46\") " pod="openstack/ceilometer-0" Nov 22 08:19:22 crc kubenswrapper[4789]: I1122 08:19:22.334350 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b3e0f7a4-cde7-4da6-8339-615054807f46-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b3e0f7a4-cde7-4da6-8339-615054807f46\") " pod="openstack/ceilometer-0" Nov 22 08:19:22 crc kubenswrapper[4789]: I1122 08:19:22.334395 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b3e0f7a4-cde7-4da6-8339-615054807f46-run-httpd\") pod \"ceilometer-0\" (UID: \"b3e0f7a4-cde7-4da6-8339-615054807f46\") " pod="openstack/ceilometer-0" Nov 22 08:19:22 crc kubenswrapper[4789]: I1122 08:19:22.334476 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b3e0f7a4-cde7-4da6-8339-615054807f46-scripts\") pod \"ceilometer-0\" (UID: \"b3e0f7a4-cde7-4da6-8339-615054807f46\") " pod="openstack/ceilometer-0" Nov 22 08:19:22 crc kubenswrapper[4789]: I1122 08:19:22.334497 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b3e0f7a4-cde7-4da6-8339-615054807f46-log-httpd\") pod \"ceilometer-0\" (UID: \"b3e0f7a4-cde7-4da6-8339-615054807f46\") " pod="openstack/ceilometer-0" Nov 22 08:19:22 crc kubenswrapper[4789]: I1122 08:19:22.334679 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/b3e0f7a4-cde7-4da6-8339-615054807f46-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"b3e0f7a4-cde7-4da6-8339-615054807f46\") " pod="openstack/ceilometer-0" Nov 22 08:19:22 crc kubenswrapper[4789]: I1122 08:19:22.436164 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/b3e0f7a4-cde7-4da6-8339-615054807f46-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"b3e0f7a4-cde7-4da6-8339-615054807f46\") " pod="openstack/ceilometer-0" Nov 22 08:19:22 crc kubenswrapper[4789]: I1122 08:19:22.436241 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8ctgf\" (UniqueName: \"kubernetes.io/projected/b3e0f7a4-cde7-4da6-8339-615054807f46-kube-api-access-8ctgf\") pod \"ceilometer-0\" (UID: \"b3e0f7a4-cde7-4da6-8339-615054807f46\") " pod="openstack/ceilometer-0" Nov 22 08:19:22 crc kubenswrapper[4789]: I1122 08:19:22.436309 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3e0f7a4-cde7-4da6-8339-615054807f46-config-data\") pod \"ceilometer-0\" (UID: \"b3e0f7a4-cde7-4da6-8339-615054807f46\") " pod="openstack/ceilometer-0" Nov 22 08:19:22 crc kubenswrapper[4789]: I1122 08:19:22.436406 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3e0f7a4-cde7-4da6-8339-615054807f46-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b3e0f7a4-cde7-4da6-8339-615054807f46\") " pod="openstack/ceilometer-0" Nov 22 08:19:22 crc kubenswrapper[4789]: I1122 08:19:22.437355 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b3e0f7a4-cde7-4da6-8339-615054807f46-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b3e0f7a4-cde7-4da6-8339-615054807f46\") " pod="openstack/ceilometer-0" Nov 22 08:19:22 crc kubenswrapper[4789]: I1122 08:19:22.437527 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b3e0f7a4-cde7-4da6-8339-615054807f46-run-httpd\") pod \"ceilometer-0\" (UID: \"b3e0f7a4-cde7-4da6-8339-615054807f46\") " pod="openstack/ceilometer-0" Nov 22 08:19:22 crc kubenswrapper[4789]: I1122 08:19:22.438125 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b3e0f7a4-cde7-4da6-8339-615054807f46-run-httpd\") pod \"ceilometer-0\" (UID: \"b3e0f7a4-cde7-4da6-8339-615054807f46\") " pod="openstack/ceilometer-0" Nov 22 08:19:22 crc kubenswrapper[4789]: I1122 08:19:22.438131 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b3e0f7a4-cde7-4da6-8339-615054807f46-scripts\") pod \"ceilometer-0\" (UID: \"b3e0f7a4-cde7-4da6-8339-615054807f46\") " pod="openstack/ceilometer-0" Nov 22 08:19:22 crc kubenswrapper[4789]: I1122 08:19:22.438178 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b3e0f7a4-cde7-4da6-8339-615054807f46-log-httpd\") pod \"ceilometer-0\" (UID: \"b3e0f7a4-cde7-4da6-8339-615054807f46\") " pod="openstack/ceilometer-0" Nov 22 08:19:22 crc kubenswrapper[4789]: I1122 08:19:22.438777 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b3e0f7a4-cde7-4da6-8339-615054807f46-log-httpd\") pod \"ceilometer-0\" (UID: \"b3e0f7a4-cde7-4da6-8339-615054807f46\") " pod="openstack/ceilometer-0" Nov 22 08:19:22 crc kubenswrapper[4789]: I1122 08:19:22.442518 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/b3e0f7a4-cde7-4da6-8339-615054807f46-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"b3e0f7a4-cde7-4da6-8339-615054807f46\") " pod="openstack/ceilometer-0" Nov 22 08:19:22 crc kubenswrapper[4789]: I1122 08:19:22.445534 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b3e0f7a4-cde7-4da6-8339-615054807f46-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b3e0f7a4-cde7-4da6-8339-615054807f46\") " pod="openstack/ceilometer-0" Nov 22 08:19:22 crc kubenswrapper[4789]: I1122 08:19:22.445843 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3e0f7a4-cde7-4da6-8339-615054807f46-config-data\") pod \"ceilometer-0\" (UID: \"b3e0f7a4-cde7-4da6-8339-615054807f46\") " pod="openstack/ceilometer-0" Nov 22 08:19:22 crc kubenswrapper[4789]: I1122 08:19:22.445858 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b3e0f7a4-cde7-4da6-8339-615054807f46-scripts\") pod \"ceilometer-0\" (UID: \"b3e0f7a4-cde7-4da6-8339-615054807f46\") " pod="openstack/ceilometer-0" Nov 22 08:19:22 crc kubenswrapper[4789]: I1122 08:19:22.446602 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3e0f7a4-cde7-4da6-8339-615054807f46-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b3e0f7a4-cde7-4da6-8339-615054807f46\") " pod="openstack/ceilometer-0" Nov 22 08:19:22 crc kubenswrapper[4789]: I1122 08:19:22.458090 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8ctgf\" (UniqueName: \"kubernetes.io/projected/b3e0f7a4-cde7-4da6-8339-615054807f46-kube-api-access-8ctgf\") pod \"ceilometer-0\" (UID: \"b3e0f7a4-cde7-4da6-8339-615054807f46\") " pod="openstack/ceilometer-0" Nov 22 08:19:22 crc kubenswrapper[4789]: I1122 08:19:22.517305 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 08:19:22 crc kubenswrapper[4789]: I1122 08:19:22.962274 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:19:22 crc kubenswrapper[4789]: W1122 08:19:22.965812 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb3e0f7a4_cde7_4da6_8339_615054807f46.slice/crio-4c51d2ca1d082219d1fe4a73d0ff639d303d7e5c7e73c3a9868682aade936de1 WatchSource:0}: Error finding container 4c51d2ca1d082219d1fe4a73d0ff639d303d7e5c7e73c3a9868682aade936de1: Status 404 returned error can't find the container with id 4c51d2ca1d082219d1fe4a73d0ff639d303d7e5c7e73c3a9868682aade936de1 Nov 22 08:19:22 crc kubenswrapper[4789]: I1122 08:19:22.968769 4789 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 22 08:19:23 crc kubenswrapper[4789]: I1122 08:19:23.276808 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 22 08:19:23 crc kubenswrapper[4789]: I1122 08:19:23.459595 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b3c74bc-088d-4ffa-ae56-933e93be3414-config-data\") pod \"1b3c74bc-088d-4ffa-ae56-933e93be3414\" (UID: \"1b3c74bc-088d-4ffa-ae56-933e93be3414\") " Nov 22 08:19:23 crc kubenswrapper[4789]: I1122 08:19:23.459989 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b3c74bc-088d-4ffa-ae56-933e93be3414-combined-ca-bundle\") pod \"1b3c74bc-088d-4ffa-ae56-933e93be3414\" (UID: \"1b3c74bc-088d-4ffa-ae56-933e93be3414\") " Nov 22 08:19:23 crc kubenswrapper[4789]: I1122 08:19:23.460033 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1b3c74bc-088d-4ffa-ae56-933e93be3414-logs\") pod \"1b3c74bc-088d-4ffa-ae56-933e93be3414\" (UID: \"1b3c74bc-088d-4ffa-ae56-933e93be3414\") " Nov 22 08:19:23 crc kubenswrapper[4789]: I1122 08:19:23.460189 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q46wb\" (UniqueName: \"kubernetes.io/projected/1b3c74bc-088d-4ffa-ae56-933e93be3414-kube-api-access-q46wb\") pod \"1b3c74bc-088d-4ffa-ae56-933e93be3414\" (UID: \"1b3c74bc-088d-4ffa-ae56-933e93be3414\") " Nov 22 08:19:23 crc kubenswrapper[4789]: I1122 08:19:23.461331 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1b3c74bc-088d-4ffa-ae56-933e93be3414-logs" (OuterVolumeSpecName: "logs") pod "1b3c74bc-088d-4ffa-ae56-933e93be3414" (UID: "1b3c74bc-088d-4ffa-ae56-933e93be3414"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:19:23 crc kubenswrapper[4789]: I1122 08:19:23.468200 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b3c74bc-088d-4ffa-ae56-933e93be3414-kube-api-access-q46wb" (OuterVolumeSpecName: "kube-api-access-q46wb") pod "1b3c74bc-088d-4ffa-ae56-933e93be3414" (UID: "1b3c74bc-088d-4ffa-ae56-933e93be3414"). InnerVolumeSpecName "kube-api-access-q46wb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:19:23 crc kubenswrapper[4789]: E1122 08:19:23.485424 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1b3c74bc-088d-4ffa-ae56-933e93be3414-config-data podName:1b3c74bc-088d-4ffa-ae56-933e93be3414 nodeName:}" failed. No retries permitted until 2025-11-22 08:19:23.985387032 +0000 UTC m=+1378.219787305 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "config-data" (UniqueName: "kubernetes.io/secret/1b3c74bc-088d-4ffa-ae56-933e93be3414-config-data") pod "1b3c74bc-088d-4ffa-ae56-933e93be3414" (UID: "1b3c74bc-088d-4ffa-ae56-933e93be3414") : error deleting /var/lib/kubelet/pods/1b3c74bc-088d-4ffa-ae56-933e93be3414/volume-subpaths: remove /var/lib/kubelet/pods/1b3c74bc-088d-4ffa-ae56-933e93be3414/volume-subpaths: no such file or directory Nov 22 08:19:23 crc kubenswrapper[4789]: I1122 08:19:23.489553 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b3c74bc-088d-4ffa-ae56-933e93be3414-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1b3c74bc-088d-4ffa-ae56-933e93be3414" (UID: "1b3c74bc-088d-4ffa-ae56-933e93be3414"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:19:23 crc kubenswrapper[4789]: I1122 08:19:23.562421 4789 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b3c74bc-088d-4ffa-ae56-933e93be3414-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:19:23 crc kubenswrapper[4789]: I1122 08:19:23.562464 4789 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1b3c74bc-088d-4ffa-ae56-933e93be3414-logs\") on node \"crc\" DevicePath \"\"" Nov 22 08:19:23 crc kubenswrapper[4789]: I1122 08:19:23.562477 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q46wb\" (UniqueName: \"kubernetes.io/projected/1b3c74bc-088d-4ffa-ae56-933e93be3414-kube-api-access-q46wb\") on node \"crc\" DevicePath \"\"" Nov 22 08:19:23 crc kubenswrapper[4789]: I1122 08:19:23.869324 4789 generic.go:334] "Generic (PLEG): container finished" podID="1b3c74bc-088d-4ffa-ae56-933e93be3414" containerID="02def008793adfc97f8a0f02c99797d3bc4826e9dbecdf67ddbc48b3937a54cc" exitCode=0 Nov 22 08:19:23 crc kubenswrapper[4789]: I1122 08:19:23.869401 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 22 08:19:23 crc kubenswrapper[4789]: I1122 08:19:23.869427 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1b3c74bc-088d-4ffa-ae56-933e93be3414","Type":"ContainerDied","Data":"02def008793adfc97f8a0f02c99797d3bc4826e9dbecdf67ddbc48b3937a54cc"} Nov 22 08:19:23 crc kubenswrapper[4789]: I1122 08:19:23.869805 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1b3c74bc-088d-4ffa-ae56-933e93be3414","Type":"ContainerDied","Data":"21eba1bbd95d8b47073662a8abea1541b3925bde0c74c6f565d1a23735b140da"} Nov 22 08:19:23 crc kubenswrapper[4789]: I1122 08:19:23.869828 4789 scope.go:117] "RemoveContainer" containerID="02def008793adfc97f8a0f02c99797d3bc4826e9dbecdf67ddbc48b3937a54cc" Nov 22 08:19:23 crc kubenswrapper[4789]: I1122 08:19:23.873998 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b3e0f7a4-cde7-4da6-8339-615054807f46","Type":"ContainerStarted","Data":"a3569e8f73bbb94022be853ca4eca0f31e32f0be828470dce1bedafa7097f19c"} Nov 22 08:19:23 crc kubenswrapper[4789]: I1122 08:19:23.874050 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b3e0f7a4-cde7-4da6-8339-615054807f46","Type":"ContainerStarted","Data":"4c51d2ca1d082219d1fe4a73d0ff639d303d7e5c7e73c3a9868682aade936de1"} Nov 22 08:19:23 crc kubenswrapper[4789]: I1122 08:19:23.893768 4789 scope.go:117] "RemoveContainer" containerID="d1ffc76583baf25f8c868b7ccd9e18ffeb0bf395b6f644e070f422339b0eb730" Nov 22 08:19:23 crc kubenswrapper[4789]: I1122 08:19:23.913284 4789 scope.go:117] "RemoveContainer" containerID="02def008793adfc97f8a0f02c99797d3bc4826e9dbecdf67ddbc48b3937a54cc" Nov 22 08:19:23 crc kubenswrapper[4789]: E1122 08:19:23.913831 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"02def008793adfc97f8a0f02c99797d3bc4826e9dbecdf67ddbc48b3937a54cc\": container with ID starting with 02def008793adfc97f8a0f02c99797d3bc4826e9dbecdf67ddbc48b3937a54cc not found: ID does not exist" containerID="02def008793adfc97f8a0f02c99797d3bc4826e9dbecdf67ddbc48b3937a54cc" Nov 22 08:19:23 crc kubenswrapper[4789]: I1122 08:19:23.913870 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"02def008793adfc97f8a0f02c99797d3bc4826e9dbecdf67ddbc48b3937a54cc"} err="failed to get container status \"02def008793adfc97f8a0f02c99797d3bc4826e9dbecdf67ddbc48b3937a54cc\": rpc error: code = NotFound desc = could not find container \"02def008793adfc97f8a0f02c99797d3bc4826e9dbecdf67ddbc48b3937a54cc\": container with ID starting with 02def008793adfc97f8a0f02c99797d3bc4826e9dbecdf67ddbc48b3937a54cc not found: ID does not exist" Nov 22 08:19:23 crc kubenswrapper[4789]: I1122 08:19:23.913896 4789 scope.go:117] "RemoveContainer" containerID="d1ffc76583baf25f8c868b7ccd9e18ffeb0bf395b6f644e070f422339b0eb730" Nov 22 08:19:23 crc kubenswrapper[4789]: E1122 08:19:23.914305 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d1ffc76583baf25f8c868b7ccd9e18ffeb0bf395b6f644e070f422339b0eb730\": container with ID starting with d1ffc76583baf25f8c868b7ccd9e18ffeb0bf395b6f644e070f422339b0eb730 not found: ID does not exist" containerID="d1ffc76583baf25f8c868b7ccd9e18ffeb0bf395b6f644e070f422339b0eb730" Nov 22 08:19:23 crc kubenswrapper[4789]: I1122 08:19:23.914337 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d1ffc76583baf25f8c868b7ccd9e18ffeb0bf395b6f644e070f422339b0eb730"} err="failed to get container status \"d1ffc76583baf25f8c868b7ccd9e18ffeb0bf395b6f644e070f422339b0eb730\": rpc error: code = NotFound desc = could not find container \"d1ffc76583baf25f8c868b7ccd9e18ffeb0bf395b6f644e070f422339b0eb730\": container with ID starting with d1ffc76583baf25f8c868b7ccd9e18ffeb0bf395b6f644e070f422339b0eb730 not found: ID does not exist" Nov 22 08:19:23 crc kubenswrapper[4789]: I1122 08:19:23.978836 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ff6994a-950e-4aae-8126-1f55702f7eba" path="/var/lib/kubelet/pods/6ff6994a-950e-4aae-8126-1f55702f7eba/volumes" Nov 22 08:19:24 crc kubenswrapper[4789]: I1122 08:19:24.070071 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b3c74bc-088d-4ffa-ae56-933e93be3414-config-data\") pod \"1b3c74bc-088d-4ffa-ae56-933e93be3414\" (UID: \"1b3c74bc-088d-4ffa-ae56-933e93be3414\") " Nov 22 08:19:24 crc kubenswrapper[4789]: I1122 08:19:24.079922 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b3c74bc-088d-4ffa-ae56-933e93be3414-config-data" (OuterVolumeSpecName: "config-data") pod "1b3c74bc-088d-4ffa-ae56-933e93be3414" (UID: "1b3c74bc-088d-4ffa-ae56-933e93be3414"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:19:24 crc kubenswrapper[4789]: I1122 08:19:24.172688 4789 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b3c74bc-088d-4ffa-ae56-933e93be3414-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:19:24 crc kubenswrapper[4789]: I1122 08:19:24.206216 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 22 08:19:24 crc kubenswrapper[4789]: I1122 08:19:24.214387 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 22 08:19:24 crc kubenswrapper[4789]: I1122 08:19:24.227838 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 22 08:19:24 crc kubenswrapper[4789]: E1122 08:19:24.228252 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b3c74bc-088d-4ffa-ae56-933e93be3414" containerName="nova-api-api" Nov 22 08:19:24 crc kubenswrapper[4789]: I1122 08:19:24.228275 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b3c74bc-088d-4ffa-ae56-933e93be3414" containerName="nova-api-api" Nov 22 08:19:24 crc kubenswrapper[4789]: E1122 08:19:24.228295 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b3c74bc-088d-4ffa-ae56-933e93be3414" containerName="nova-api-log" Nov 22 08:19:24 crc kubenswrapper[4789]: I1122 08:19:24.228301 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b3c74bc-088d-4ffa-ae56-933e93be3414" containerName="nova-api-log" Nov 22 08:19:24 crc kubenswrapper[4789]: I1122 08:19:24.228457 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b3c74bc-088d-4ffa-ae56-933e93be3414" containerName="nova-api-api" Nov 22 08:19:24 crc kubenswrapper[4789]: I1122 08:19:24.228473 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b3c74bc-088d-4ffa-ae56-933e93be3414" containerName="nova-api-log" Nov 22 08:19:24 crc kubenswrapper[4789]: I1122 08:19:24.229456 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 22 08:19:24 crc kubenswrapper[4789]: I1122 08:19:24.235148 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Nov 22 08:19:24 crc kubenswrapper[4789]: I1122 08:19:24.235454 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 22 08:19:24 crc kubenswrapper[4789]: I1122 08:19:24.235573 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Nov 22 08:19:24 crc kubenswrapper[4789]: I1122 08:19:24.246601 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 22 08:19:24 crc kubenswrapper[4789]: I1122 08:19:24.376019 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a7ffaa8c-24d2-4d9c-bf31-d231b4f6437e-public-tls-certs\") pod \"nova-api-0\" (UID: \"a7ffaa8c-24d2-4d9c-bf31-d231b4f6437e\") " pod="openstack/nova-api-0" Nov 22 08:19:24 crc kubenswrapper[4789]: I1122 08:19:24.376363 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a7ffaa8c-24d2-4d9c-bf31-d231b4f6437e-internal-tls-certs\") pod \"nova-api-0\" (UID: \"a7ffaa8c-24d2-4d9c-bf31-d231b4f6437e\") " pod="openstack/nova-api-0" Nov 22 08:19:24 crc kubenswrapper[4789]: I1122 08:19:24.376433 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7ffaa8c-24d2-4d9c-bf31-d231b4f6437e-config-data\") pod \"nova-api-0\" (UID: \"a7ffaa8c-24d2-4d9c-bf31-d231b4f6437e\") " pod="openstack/nova-api-0" Nov 22 08:19:24 crc kubenswrapper[4789]: I1122 08:19:24.376669 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2z5dl\" (UniqueName: \"kubernetes.io/projected/a7ffaa8c-24d2-4d9c-bf31-d231b4f6437e-kube-api-access-2z5dl\") pod \"nova-api-0\" (UID: \"a7ffaa8c-24d2-4d9c-bf31-d231b4f6437e\") " pod="openstack/nova-api-0" Nov 22 08:19:24 crc kubenswrapper[4789]: I1122 08:19:24.376777 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a7ffaa8c-24d2-4d9c-bf31-d231b4f6437e-logs\") pod \"nova-api-0\" (UID: \"a7ffaa8c-24d2-4d9c-bf31-d231b4f6437e\") " pod="openstack/nova-api-0" Nov 22 08:19:24 crc kubenswrapper[4789]: I1122 08:19:24.376927 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7ffaa8c-24d2-4d9c-bf31-d231b4f6437e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a7ffaa8c-24d2-4d9c-bf31-d231b4f6437e\") " pod="openstack/nova-api-0" Nov 22 08:19:24 crc kubenswrapper[4789]: I1122 08:19:24.478371 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2z5dl\" (UniqueName: \"kubernetes.io/projected/a7ffaa8c-24d2-4d9c-bf31-d231b4f6437e-kube-api-access-2z5dl\") pod \"nova-api-0\" (UID: \"a7ffaa8c-24d2-4d9c-bf31-d231b4f6437e\") " pod="openstack/nova-api-0" Nov 22 08:19:24 crc kubenswrapper[4789]: I1122 08:19:24.478441 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a7ffaa8c-24d2-4d9c-bf31-d231b4f6437e-logs\") pod \"nova-api-0\" (UID: \"a7ffaa8c-24d2-4d9c-bf31-d231b4f6437e\") " pod="openstack/nova-api-0" Nov 22 08:19:24 crc kubenswrapper[4789]: I1122 08:19:24.478490 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7ffaa8c-24d2-4d9c-bf31-d231b4f6437e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a7ffaa8c-24d2-4d9c-bf31-d231b4f6437e\") " pod="openstack/nova-api-0" Nov 22 08:19:24 crc kubenswrapper[4789]: I1122 08:19:24.478545 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a7ffaa8c-24d2-4d9c-bf31-d231b4f6437e-public-tls-certs\") pod \"nova-api-0\" (UID: \"a7ffaa8c-24d2-4d9c-bf31-d231b4f6437e\") " pod="openstack/nova-api-0" Nov 22 08:19:24 crc kubenswrapper[4789]: I1122 08:19:24.478583 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a7ffaa8c-24d2-4d9c-bf31-d231b4f6437e-internal-tls-certs\") pod \"nova-api-0\" (UID: \"a7ffaa8c-24d2-4d9c-bf31-d231b4f6437e\") " pod="openstack/nova-api-0" Nov 22 08:19:24 crc kubenswrapper[4789]: I1122 08:19:24.478645 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7ffaa8c-24d2-4d9c-bf31-d231b4f6437e-config-data\") pod \"nova-api-0\" (UID: \"a7ffaa8c-24d2-4d9c-bf31-d231b4f6437e\") " pod="openstack/nova-api-0" Nov 22 08:19:24 crc kubenswrapper[4789]: I1122 08:19:24.479922 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a7ffaa8c-24d2-4d9c-bf31-d231b4f6437e-logs\") pod \"nova-api-0\" (UID: \"a7ffaa8c-24d2-4d9c-bf31-d231b4f6437e\") " pod="openstack/nova-api-0" Nov 22 08:19:24 crc kubenswrapper[4789]: I1122 08:19:24.483736 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7ffaa8c-24d2-4d9c-bf31-d231b4f6437e-config-data\") pod \"nova-api-0\" (UID: \"a7ffaa8c-24d2-4d9c-bf31-d231b4f6437e\") " pod="openstack/nova-api-0" Nov 22 08:19:24 crc kubenswrapper[4789]: I1122 08:19:24.485263 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a7ffaa8c-24d2-4d9c-bf31-d231b4f6437e-public-tls-certs\") pod \"nova-api-0\" (UID: \"a7ffaa8c-24d2-4d9c-bf31-d231b4f6437e\") " pod="openstack/nova-api-0" Nov 22 08:19:24 crc kubenswrapper[4789]: I1122 08:19:24.485321 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a7ffaa8c-24d2-4d9c-bf31-d231b4f6437e-internal-tls-certs\") pod \"nova-api-0\" (UID: \"a7ffaa8c-24d2-4d9c-bf31-d231b4f6437e\") " pod="openstack/nova-api-0" Nov 22 08:19:24 crc kubenswrapper[4789]: I1122 08:19:24.485871 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7ffaa8c-24d2-4d9c-bf31-d231b4f6437e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a7ffaa8c-24d2-4d9c-bf31-d231b4f6437e\") " pod="openstack/nova-api-0" Nov 22 08:19:24 crc kubenswrapper[4789]: I1122 08:19:24.501303 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2z5dl\" (UniqueName: \"kubernetes.io/projected/a7ffaa8c-24d2-4d9c-bf31-d231b4f6437e-kube-api-access-2z5dl\") pod \"nova-api-0\" (UID: \"a7ffaa8c-24d2-4d9c-bf31-d231b4f6437e\") " pod="openstack/nova-api-0" Nov 22 08:19:24 crc kubenswrapper[4789]: I1122 08:19:24.550550 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 22 08:19:24 crc kubenswrapper[4789]: I1122 08:19:24.902726 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b3e0f7a4-cde7-4da6-8339-615054807f46","Type":"ContainerStarted","Data":"0ba876bc146a633673594ffc7a55b1ef15f6d31b6f68e177c9d9b989cfe96170"} Nov 22 08:19:25 crc kubenswrapper[4789]: I1122 08:19:25.031365 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 22 08:19:25 crc kubenswrapper[4789]: W1122 08:19:25.031440 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda7ffaa8c_24d2_4d9c_bf31_d231b4f6437e.slice/crio-90367a61cf3c546410a50628442d2d71190a291493c47bfa9f71f589732cacc9 WatchSource:0}: Error finding container 90367a61cf3c546410a50628442d2d71190a291493c47bfa9f71f589732cacc9: Status 404 returned error can't find the container with id 90367a61cf3c546410a50628442d2d71190a291493c47bfa9f71f589732cacc9 Nov 22 08:19:25 crc kubenswrapper[4789]: I1122 08:19:25.918195 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a7ffaa8c-24d2-4d9c-bf31-d231b4f6437e","Type":"ContainerStarted","Data":"58f9f8ae2133e4b6373fb58ea49634210fb379398706244f6b83b567ddb4f221"} Nov 22 08:19:25 crc kubenswrapper[4789]: I1122 08:19:25.918599 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a7ffaa8c-24d2-4d9c-bf31-d231b4f6437e","Type":"ContainerStarted","Data":"fe3a2122b6c2fe085f2a827a9f023bbb1fea1049366684c22dc191b340fffd78"} Nov 22 08:19:25 crc kubenswrapper[4789]: I1122 08:19:25.918616 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a7ffaa8c-24d2-4d9c-bf31-d231b4f6437e","Type":"ContainerStarted","Data":"90367a61cf3c546410a50628442d2d71190a291493c47bfa9f71f589732cacc9"} Nov 22 08:19:25 crc kubenswrapper[4789]: I1122 08:19:25.925920 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b3e0f7a4-cde7-4da6-8339-615054807f46","Type":"ContainerStarted","Data":"e7e7196b9d8de4d32b0220fbdd5f43d6bd1d61ff63e6f25250d131b43d76827d"} Nov 22 08:19:25 crc kubenswrapper[4789]: I1122 08:19:25.940605 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=1.940583367 podStartE2EDuration="1.940583367s" podCreationTimestamp="2025-11-22 08:19:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:19:25.936585467 +0000 UTC m=+1380.170985750" watchObservedRunningTime="2025-11-22 08:19:25.940583367 +0000 UTC m=+1380.174983640" Nov 22 08:19:25 crc kubenswrapper[4789]: I1122 08:19:25.976426 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1b3c74bc-088d-4ffa-ae56-933e93be3414" path="/var/lib/kubelet/pods/1b3c74bc-088d-4ffa-ae56-933e93be3414/volumes" Nov 22 08:19:26 crc kubenswrapper[4789]: I1122 08:19:26.938125 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b3e0f7a4-cde7-4da6-8339-615054807f46","Type":"ContainerStarted","Data":"2fb144347d1d1aad510a6ea7031cee3e906598e8fd06b81b6777431331eafb09"} Nov 22 08:19:26 crc kubenswrapper[4789]: I1122 08:19:26.938866 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 22 08:19:26 crc kubenswrapper[4789]: I1122 08:19:26.972548 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.126238773 podStartE2EDuration="4.972529647s" podCreationTimestamp="2025-11-22 08:19:22 +0000 UTC" firstStartedPulling="2025-11-22 08:19:22.968487741 +0000 UTC m=+1377.202888014" lastFinishedPulling="2025-11-22 08:19:25.814778615 +0000 UTC m=+1380.049178888" observedRunningTime="2025-11-22 08:19:26.965705801 +0000 UTC m=+1381.200106074" watchObservedRunningTime="2025-11-22 08:19:26.972529647 +0000 UTC m=+1381.206929920" Nov 22 08:19:27 crc kubenswrapper[4789]: I1122 08:19:27.146245 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Nov 22 08:19:27 crc kubenswrapper[4789]: I1122 08:19:27.162652 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Nov 22 08:19:27 crc kubenswrapper[4789]: I1122 08:19:27.824367 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-c9b558957-c7k8k" Nov 22 08:19:27 crc kubenswrapper[4789]: I1122 08:19:27.898202 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-69494d9f89-z9hd6"] Nov 22 08:19:27 crc kubenswrapper[4789]: I1122 08:19:27.899144 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-69494d9f89-z9hd6" podUID="75e7c40c-ba8c-45f6-b30c-f654f48956ca" containerName="dnsmasq-dns" containerID="cri-o://209550f4b6c9910b78a3c4910da280b88da86bda2367421bdb30ba3154833e07" gracePeriod=10 Nov 22 08:19:27 crc kubenswrapper[4789]: I1122 08:19:27.963584 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Nov 22 08:19:28 crc kubenswrapper[4789]: I1122 08:19:28.103219 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-c94ng"] Nov 22 08:19:28 crc kubenswrapper[4789]: I1122 08:19:28.108227 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-c94ng" Nov 22 08:19:28 crc kubenswrapper[4789]: I1122 08:19:28.113883 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Nov 22 08:19:28 crc kubenswrapper[4789]: I1122 08:19:28.115075 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Nov 22 08:19:28 crc kubenswrapper[4789]: I1122 08:19:28.145271 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/206a90d7-1c1c-4a17-992a-43bd718fc9ea-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-c94ng\" (UID: \"206a90d7-1c1c-4a17-992a-43bd718fc9ea\") " pod="openstack/nova-cell1-cell-mapping-c94ng" Nov 22 08:19:28 crc kubenswrapper[4789]: I1122 08:19:28.145337 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/206a90d7-1c1c-4a17-992a-43bd718fc9ea-scripts\") pod \"nova-cell1-cell-mapping-c94ng\" (UID: \"206a90d7-1c1c-4a17-992a-43bd718fc9ea\") " pod="openstack/nova-cell1-cell-mapping-c94ng" Nov 22 08:19:28 crc kubenswrapper[4789]: I1122 08:19:28.145387 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4pgnf\" (UniqueName: \"kubernetes.io/projected/206a90d7-1c1c-4a17-992a-43bd718fc9ea-kube-api-access-4pgnf\") pod \"nova-cell1-cell-mapping-c94ng\" (UID: \"206a90d7-1c1c-4a17-992a-43bd718fc9ea\") " pod="openstack/nova-cell1-cell-mapping-c94ng" Nov 22 08:19:28 crc kubenswrapper[4789]: I1122 08:19:28.145495 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/206a90d7-1c1c-4a17-992a-43bd718fc9ea-config-data\") pod \"nova-cell1-cell-mapping-c94ng\" (UID: \"206a90d7-1c1c-4a17-992a-43bd718fc9ea\") " pod="openstack/nova-cell1-cell-mapping-c94ng" Nov 22 08:19:28 crc kubenswrapper[4789]: I1122 08:19:28.146618 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-c94ng"] Nov 22 08:19:28 crc kubenswrapper[4789]: I1122 08:19:28.246555 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/206a90d7-1c1c-4a17-992a-43bd718fc9ea-scripts\") pod \"nova-cell1-cell-mapping-c94ng\" (UID: \"206a90d7-1c1c-4a17-992a-43bd718fc9ea\") " pod="openstack/nova-cell1-cell-mapping-c94ng" Nov 22 08:19:28 crc kubenswrapper[4789]: I1122 08:19:28.246633 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4pgnf\" (UniqueName: \"kubernetes.io/projected/206a90d7-1c1c-4a17-992a-43bd718fc9ea-kube-api-access-4pgnf\") pod \"nova-cell1-cell-mapping-c94ng\" (UID: \"206a90d7-1c1c-4a17-992a-43bd718fc9ea\") " pod="openstack/nova-cell1-cell-mapping-c94ng" Nov 22 08:19:28 crc kubenswrapper[4789]: I1122 08:19:28.246725 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/206a90d7-1c1c-4a17-992a-43bd718fc9ea-config-data\") pod \"nova-cell1-cell-mapping-c94ng\" (UID: \"206a90d7-1c1c-4a17-992a-43bd718fc9ea\") " pod="openstack/nova-cell1-cell-mapping-c94ng" Nov 22 08:19:28 crc kubenswrapper[4789]: I1122 08:19:28.246788 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/206a90d7-1c1c-4a17-992a-43bd718fc9ea-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-c94ng\" (UID: \"206a90d7-1c1c-4a17-992a-43bd718fc9ea\") " pod="openstack/nova-cell1-cell-mapping-c94ng" Nov 22 08:19:28 crc kubenswrapper[4789]: I1122 08:19:28.252351 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/206a90d7-1c1c-4a17-992a-43bd718fc9ea-config-data\") pod \"nova-cell1-cell-mapping-c94ng\" (UID: \"206a90d7-1c1c-4a17-992a-43bd718fc9ea\") " pod="openstack/nova-cell1-cell-mapping-c94ng" Nov 22 08:19:28 crc kubenswrapper[4789]: I1122 08:19:28.252537 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/206a90d7-1c1c-4a17-992a-43bd718fc9ea-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-c94ng\" (UID: \"206a90d7-1c1c-4a17-992a-43bd718fc9ea\") " pod="openstack/nova-cell1-cell-mapping-c94ng" Nov 22 08:19:28 crc kubenswrapper[4789]: I1122 08:19:28.256205 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/206a90d7-1c1c-4a17-992a-43bd718fc9ea-scripts\") pod \"nova-cell1-cell-mapping-c94ng\" (UID: \"206a90d7-1c1c-4a17-992a-43bd718fc9ea\") " pod="openstack/nova-cell1-cell-mapping-c94ng" Nov 22 08:19:28 crc kubenswrapper[4789]: I1122 08:19:28.266831 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4pgnf\" (UniqueName: \"kubernetes.io/projected/206a90d7-1c1c-4a17-992a-43bd718fc9ea-kube-api-access-4pgnf\") pod \"nova-cell1-cell-mapping-c94ng\" (UID: \"206a90d7-1c1c-4a17-992a-43bd718fc9ea\") " pod="openstack/nova-cell1-cell-mapping-c94ng" Nov 22 08:19:28 crc kubenswrapper[4789]: I1122 08:19:28.438571 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-c94ng" Nov 22 08:19:28 crc kubenswrapper[4789]: I1122 08:19:28.476548 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-69494d9f89-z9hd6" Nov 22 08:19:28 crc kubenswrapper[4789]: I1122 08:19:28.653168 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/75e7c40c-ba8c-45f6-b30c-f654f48956ca-config\") pod \"75e7c40c-ba8c-45f6-b30c-f654f48956ca\" (UID: \"75e7c40c-ba8c-45f6-b30c-f654f48956ca\") " Nov 22 08:19:28 crc kubenswrapper[4789]: I1122 08:19:28.653240 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/75e7c40c-ba8c-45f6-b30c-f654f48956ca-ovsdbserver-sb\") pod \"75e7c40c-ba8c-45f6-b30c-f654f48956ca\" (UID: \"75e7c40c-ba8c-45f6-b30c-f654f48956ca\") " Nov 22 08:19:28 crc kubenswrapper[4789]: I1122 08:19:28.653340 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/75e7c40c-ba8c-45f6-b30c-f654f48956ca-ovsdbserver-nb\") pod \"75e7c40c-ba8c-45f6-b30c-f654f48956ca\" (UID: \"75e7c40c-ba8c-45f6-b30c-f654f48956ca\") " Nov 22 08:19:28 crc kubenswrapper[4789]: I1122 08:19:28.653382 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/75e7c40c-ba8c-45f6-b30c-f654f48956ca-dns-svc\") pod \"75e7c40c-ba8c-45f6-b30c-f654f48956ca\" (UID: \"75e7c40c-ba8c-45f6-b30c-f654f48956ca\") " Nov 22 08:19:28 crc kubenswrapper[4789]: I1122 08:19:28.653425 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-57kgk\" (UniqueName: \"kubernetes.io/projected/75e7c40c-ba8c-45f6-b30c-f654f48956ca-kube-api-access-57kgk\") pod \"75e7c40c-ba8c-45f6-b30c-f654f48956ca\" (UID: \"75e7c40c-ba8c-45f6-b30c-f654f48956ca\") " Nov 22 08:19:28 crc kubenswrapper[4789]: I1122 08:19:28.659817 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75e7c40c-ba8c-45f6-b30c-f654f48956ca-kube-api-access-57kgk" (OuterVolumeSpecName: "kube-api-access-57kgk") pod "75e7c40c-ba8c-45f6-b30c-f654f48956ca" (UID: "75e7c40c-ba8c-45f6-b30c-f654f48956ca"). InnerVolumeSpecName "kube-api-access-57kgk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:19:28 crc kubenswrapper[4789]: I1122 08:19:28.707571 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/75e7c40c-ba8c-45f6-b30c-f654f48956ca-config" (OuterVolumeSpecName: "config") pod "75e7c40c-ba8c-45f6-b30c-f654f48956ca" (UID: "75e7c40c-ba8c-45f6-b30c-f654f48956ca"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:19:28 crc kubenswrapper[4789]: I1122 08:19:28.716271 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/75e7c40c-ba8c-45f6-b30c-f654f48956ca-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "75e7c40c-ba8c-45f6-b30c-f654f48956ca" (UID: "75e7c40c-ba8c-45f6-b30c-f654f48956ca"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:19:28 crc kubenswrapper[4789]: I1122 08:19:28.721200 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/75e7c40c-ba8c-45f6-b30c-f654f48956ca-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "75e7c40c-ba8c-45f6-b30c-f654f48956ca" (UID: "75e7c40c-ba8c-45f6-b30c-f654f48956ca"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:19:28 crc kubenswrapper[4789]: I1122 08:19:28.724371 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/75e7c40c-ba8c-45f6-b30c-f654f48956ca-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "75e7c40c-ba8c-45f6-b30c-f654f48956ca" (UID: "75e7c40c-ba8c-45f6-b30c-f654f48956ca"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:19:28 crc kubenswrapper[4789]: I1122 08:19:28.755473 4789 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/75e7c40c-ba8c-45f6-b30c-f654f48956ca-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 22 08:19:28 crc kubenswrapper[4789]: I1122 08:19:28.755514 4789 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/75e7c40c-ba8c-45f6-b30c-f654f48956ca-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 22 08:19:28 crc kubenswrapper[4789]: I1122 08:19:28.755527 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-57kgk\" (UniqueName: \"kubernetes.io/projected/75e7c40c-ba8c-45f6-b30c-f654f48956ca-kube-api-access-57kgk\") on node \"crc\" DevicePath \"\"" Nov 22 08:19:28 crc kubenswrapper[4789]: I1122 08:19:28.755538 4789 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/75e7c40c-ba8c-45f6-b30c-f654f48956ca-config\") on node \"crc\" DevicePath \"\"" Nov 22 08:19:28 crc kubenswrapper[4789]: I1122 08:19:28.755546 4789 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/75e7c40c-ba8c-45f6-b30c-f654f48956ca-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 22 08:19:28 crc kubenswrapper[4789]: I1122 08:19:28.921314 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-c94ng"] Nov 22 08:19:28 crc kubenswrapper[4789]: W1122 08:19:28.923819 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod206a90d7_1c1c_4a17_992a_43bd718fc9ea.slice/crio-f05ab48926a75cf1fcadc03adb3f23832ff5a80acfcca47b04243c458a1da7df WatchSource:0}: Error finding container f05ab48926a75cf1fcadc03adb3f23832ff5a80acfcca47b04243c458a1da7df: Status 404 returned error can't find the container with id f05ab48926a75cf1fcadc03adb3f23832ff5a80acfcca47b04243c458a1da7df Nov 22 08:19:28 crc kubenswrapper[4789]: I1122 08:19:28.966968 4789 generic.go:334] "Generic (PLEG): container finished" podID="75e7c40c-ba8c-45f6-b30c-f654f48956ca" containerID="209550f4b6c9910b78a3c4910da280b88da86bda2367421bdb30ba3154833e07" exitCode=0 Nov 22 08:19:28 crc kubenswrapper[4789]: I1122 08:19:28.967124 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-69494d9f89-z9hd6" event={"ID":"75e7c40c-ba8c-45f6-b30c-f654f48956ca","Type":"ContainerDied","Data":"209550f4b6c9910b78a3c4910da280b88da86bda2367421bdb30ba3154833e07"} Nov 22 08:19:28 crc kubenswrapper[4789]: I1122 08:19:28.967173 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-69494d9f89-z9hd6" event={"ID":"75e7c40c-ba8c-45f6-b30c-f654f48956ca","Type":"ContainerDied","Data":"4c282f547bcbdad803bad222cabb5fcdb560aa8c5982c873cdb4469fca1bc3a9"} Nov 22 08:19:28 crc kubenswrapper[4789]: I1122 08:19:28.967192 4789 scope.go:117] "RemoveContainer" containerID="209550f4b6c9910b78a3c4910da280b88da86bda2367421bdb30ba3154833e07" Nov 22 08:19:28 crc kubenswrapper[4789]: I1122 08:19:28.968859 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-69494d9f89-z9hd6" Nov 22 08:19:28 crc kubenswrapper[4789]: I1122 08:19:28.972083 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-c94ng" event={"ID":"206a90d7-1c1c-4a17-992a-43bd718fc9ea","Type":"ContainerStarted","Data":"f05ab48926a75cf1fcadc03adb3f23832ff5a80acfcca47b04243c458a1da7df"} Nov 22 08:19:29 crc kubenswrapper[4789]: I1122 08:19:29.005696 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-69494d9f89-z9hd6"] Nov 22 08:19:29 crc kubenswrapper[4789]: I1122 08:19:29.007901 4789 scope.go:117] "RemoveContainer" containerID="6783b7130455f20a22183eb8df83c513dd7d6d2621b36957429a75060739d0fb" Nov 22 08:19:29 crc kubenswrapper[4789]: I1122 08:19:29.017223 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-69494d9f89-z9hd6"] Nov 22 08:19:29 crc kubenswrapper[4789]: I1122 08:19:29.086321 4789 scope.go:117] "RemoveContainer" containerID="209550f4b6c9910b78a3c4910da280b88da86bda2367421bdb30ba3154833e07" Nov 22 08:19:29 crc kubenswrapper[4789]: E1122 08:19:29.086835 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"209550f4b6c9910b78a3c4910da280b88da86bda2367421bdb30ba3154833e07\": container with ID starting with 209550f4b6c9910b78a3c4910da280b88da86bda2367421bdb30ba3154833e07 not found: ID does not exist" containerID="209550f4b6c9910b78a3c4910da280b88da86bda2367421bdb30ba3154833e07" Nov 22 08:19:29 crc kubenswrapper[4789]: I1122 08:19:29.086942 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"209550f4b6c9910b78a3c4910da280b88da86bda2367421bdb30ba3154833e07"} err="failed to get container status \"209550f4b6c9910b78a3c4910da280b88da86bda2367421bdb30ba3154833e07\": rpc error: code = NotFound desc = could not find container \"209550f4b6c9910b78a3c4910da280b88da86bda2367421bdb30ba3154833e07\": container with ID starting with 209550f4b6c9910b78a3c4910da280b88da86bda2367421bdb30ba3154833e07 not found: ID does not exist" Nov 22 08:19:29 crc kubenswrapper[4789]: I1122 08:19:29.087038 4789 scope.go:117] "RemoveContainer" containerID="6783b7130455f20a22183eb8df83c513dd7d6d2621b36957429a75060739d0fb" Nov 22 08:19:29 crc kubenswrapper[4789]: E1122 08:19:29.087494 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6783b7130455f20a22183eb8df83c513dd7d6d2621b36957429a75060739d0fb\": container with ID starting with 6783b7130455f20a22183eb8df83c513dd7d6d2621b36957429a75060739d0fb not found: ID does not exist" containerID="6783b7130455f20a22183eb8df83c513dd7d6d2621b36957429a75060739d0fb" Nov 22 08:19:29 crc kubenswrapper[4789]: I1122 08:19:29.087523 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6783b7130455f20a22183eb8df83c513dd7d6d2621b36957429a75060739d0fb"} err="failed to get container status \"6783b7130455f20a22183eb8df83c513dd7d6d2621b36957429a75060739d0fb\": rpc error: code = NotFound desc = could not find container \"6783b7130455f20a22183eb8df83c513dd7d6d2621b36957429a75060739d0fb\": container with ID starting with 6783b7130455f20a22183eb8df83c513dd7d6d2621b36957429a75060739d0fb not found: ID does not exist" Nov 22 08:19:29 crc kubenswrapper[4789]: I1122 08:19:29.990844 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="75e7c40c-ba8c-45f6-b30c-f654f48956ca" path="/var/lib/kubelet/pods/75e7c40c-ba8c-45f6-b30c-f654f48956ca/volumes" Nov 22 08:19:29 crc kubenswrapper[4789]: I1122 08:19:29.993873 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-c94ng" event={"ID":"206a90d7-1c1c-4a17-992a-43bd718fc9ea","Type":"ContainerStarted","Data":"380143d9ea6b3ab19e351a49cb7d0db0752bada11ce7974131318d6c6d340ff3"} Nov 22 08:19:30 crc kubenswrapper[4789]: I1122 08:19:30.015148 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-c94ng" podStartSLOduration=2.015129825 podStartE2EDuration="2.015129825s" podCreationTimestamp="2025-11-22 08:19:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:19:30.009477021 +0000 UTC m=+1384.243877304" watchObservedRunningTime="2025-11-22 08:19:30.015129825 +0000 UTC m=+1384.249530098" Nov 22 08:19:34 crc kubenswrapper[4789]: I1122 08:19:34.551465 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 22 08:19:34 crc kubenswrapper[4789]: I1122 08:19:34.552132 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 22 08:19:35 crc kubenswrapper[4789]: I1122 08:19:35.040059 4789 generic.go:334] "Generic (PLEG): container finished" podID="206a90d7-1c1c-4a17-992a-43bd718fc9ea" containerID="380143d9ea6b3ab19e351a49cb7d0db0752bada11ce7974131318d6c6d340ff3" exitCode=0 Nov 22 08:19:35 crc kubenswrapper[4789]: I1122 08:19:35.040143 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-c94ng" event={"ID":"206a90d7-1c1c-4a17-992a-43bd718fc9ea","Type":"ContainerDied","Data":"380143d9ea6b3ab19e351a49cb7d0db0752bada11ce7974131318d6c6d340ff3"} Nov 22 08:19:35 crc kubenswrapper[4789]: I1122 08:19:35.372494 4789 patch_prober.go:28] interesting pod/machine-config-daemon-pxvrm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 08:19:35 crc kubenswrapper[4789]: I1122 08:19:35.372555 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 08:19:35 crc kubenswrapper[4789]: I1122 08:19:35.372616 4789 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" Nov 22 08:19:35 crc kubenswrapper[4789]: I1122 08:19:35.373296 4789 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3d1ad86d18575ad837bca3f627df48477a6e890e1d7c69736133f6389315bc07"} pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 22 08:19:35 crc kubenswrapper[4789]: I1122 08:19:35.373355 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" containerID="cri-o://3d1ad86d18575ad837bca3f627df48477a6e890e1d7c69736133f6389315bc07" gracePeriod=600 Nov 22 08:19:35 crc kubenswrapper[4789]: I1122 08:19:35.566195 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="a7ffaa8c-24d2-4d9c-bf31-d231b4f6437e" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.187:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 22 08:19:35 crc kubenswrapper[4789]: I1122 08:19:35.566056 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="a7ffaa8c-24d2-4d9c-bf31-d231b4f6437e" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.187:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 22 08:19:36 crc kubenswrapper[4789]: I1122 08:19:36.051720 4789 generic.go:334] "Generic (PLEG): container finished" podID="58aa8071-7ecc-4692-8789-2db04ec70647" containerID="3d1ad86d18575ad837bca3f627df48477a6e890e1d7c69736133f6389315bc07" exitCode=0 Nov 22 08:19:36 crc kubenswrapper[4789]: I1122 08:19:36.051773 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" event={"ID":"58aa8071-7ecc-4692-8789-2db04ec70647","Type":"ContainerDied","Data":"3d1ad86d18575ad837bca3f627df48477a6e890e1d7c69736133f6389315bc07"} Nov 22 08:19:36 crc kubenswrapper[4789]: I1122 08:19:36.052287 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" event={"ID":"58aa8071-7ecc-4692-8789-2db04ec70647","Type":"ContainerStarted","Data":"cd9752d9c3408de592a39b57fc6c7443fb52d47aecefc0deabddc20b20428c47"} Nov 22 08:19:36 crc kubenswrapper[4789]: I1122 08:19:36.052312 4789 scope.go:117] "RemoveContainer" containerID="fe65d4df369fc1d7c708e571722629956fbd341c1d6b8fc8c7f64c8dfbfa797f" Nov 22 08:19:36 crc kubenswrapper[4789]: I1122 08:19:36.417118 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-c94ng" Nov 22 08:19:36 crc kubenswrapper[4789]: I1122 08:19:36.591310 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/206a90d7-1c1c-4a17-992a-43bd718fc9ea-config-data\") pod \"206a90d7-1c1c-4a17-992a-43bd718fc9ea\" (UID: \"206a90d7-1c1c-4a17-992a-43bd718fc9ea\") " Nov 22 08:19:36 crc kubenswrapper[4789]: I1122 08:19:36.591368 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/206a90d7-1c1c-4a17-992a-43bd718fc9ea-combined-ca-bundle\") pod \"206a90d7-1c1c-4a17-992a-43bd718fc9ea\" (UID: \"206a90d7-1c1c-4a17-992a-43bd718fc9ea\") " Nov 22 08:19:36 crc kubenswrapper[4789]: I1122 08:19:36.591492 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/206a90d7-1c1c-4a17-992a-43bd718fc9ea-scripts\") pod \"206a90d7-1c1c-4a17-992a-43bd718fc9ea\" (UID: \"206a90d7-1c1c-4a17-992a-43bd718fc9ea\") " Nov 22 08:19:36 crc kubenswrapper[4789]: I1122 08:19:36.591586 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4pgnf\" (UniqueName: \"kubernetes.io/projected/206a90d7-1c1c-4a17-992a-43bd718fc9ea-kube-api-access-4pgnf\") pod \"206a90d7-1c1c-4a17-992a-43bd718fc9ea\" (UID: \"206a90d7-1c1c-4a17-992a-43bd718fc9ea\") " Nov 22 08:19:36 crc kubenswrapper[4789]: I1122 08:19:36.597568 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/206a90d7-1c1c-4a17-992a-43bd718fc9ea-scripts" (OuterVolumeSpecName: "scripts") pod "206a90d7-1c1c-4a17-992a-43bd718fc9ea" (UID: "206a90d7-1c1c-4a17-992a-43bd718fc9ea"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:19:36 crc kubenswrapper[4789]: I1122 08:19:36.615474 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/206a90d7-1c1c-4a17-992a-43bd718fc9ea-kube-api-access-4pgnf" (OuterVolumeSpecName: "kube-api-access-4pgnf") pod "206a90d7-1c1c-4a17-992a-43bd718fc9ea" (UID: "206a90d7-1c1c-4a17-992a-43bd718fc9ea"). InnerVolumeSpecName "kube-api-access-4pgnf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:19:36 crc kubenswrapper[4789]: I1122 08:19:36.622913 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/206a90d7-1c1c-4a17-992a-43bd718fc9ea-config-data" (OuterVolumeSpecName: "config-data") pod "206a90d7-1c1c-4a17-992a-43bd718fc9ea" (UID: "206a90d7-1c1c-4a17-992a-43bd718fc9ea"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:19:36 crc kubenswrapper[4789]: I1122 08:19:36.626841 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/206a90d7-1c1c-4a17-992a-43bd718fc9ea-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "206a90d7-1c1c-4a17-992a-43bd718fc9ea" (UID: "206a90d7-1c1c-4a17-992a-43bd718fc9ea"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:19:36 crc kubenswrapper[4789]: I1122 08:19:36.694838 4789 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/206a90d7-1c1c-4a17-992a-43bd718fc9ea-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:19:36 crc kubenswrapper[4789]: I1122 08:19:36.694876 4789 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/206a90d7-1c1c-4a17-992a-43bd718fc9ea-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:19:36 crc kubenswrapper[4789]: I1122 08:19:36.694888 4789 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/206a90d7-1c1c-4a17-992a-43bd718fc9ea-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:19:36 crc kubenswrapper[4789]: I1122 08:19:36.694896 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4pgnf\" (UniqueName: \"kubernetes.io/projected/206a90d7-1c1c-4a17-992a-43bd718fc9ea-kube-api-access-4pgnf\") on node \"crc\" DevicePath \"\"" Nov 22 08:19:37 crc kubenswrapper[4789]: I1122 08:19:37.061905 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-c94ng" Nov 22 08:19:37 crc kubenswrapper[4789]: I1122 08:19:37.061898 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-c94ng" event={"ID":"206a90d7-1c1c-4a17-992a-43bd718fc9ea","Type":"ContainerDied","Data":"f05ab48926a75cf1fcadc03adb3f23832ff5a80acfcca47b04243c458a1da7df"} Nov 22 08:19:37 crc kubenswrapper[4789]: I1122 08:19:37.062330 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f05ab48926a75cf1fcadc03adb3f23832ff5a80acfcca47b04243c458a1da7df" Nov 22 08:19:37 crc kubenswrapper[4789]: I1122 08:19:37.240658 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 22 08:19:37 crc kubenswrapper[4789]: I1122 08:19:37.241065 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="a7ffaa8c-24d2-4d9c-bf31-d231b4f6437e" containerName="nova-api-log" containerID="cri-o://fe3a2122b6c2fe085f2a827a9f023bbb1fea1049366684c22dc191b340fffd78" gracePeriod=30 Nov 22 08:19:37 crc kubenswrapper[4789]: I1122 08:19:37.241147 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="a7ffaa8c-24d2-4d9c-bf31-d231b4f6437e" containerName="nova-api-api" containerID="cri-o://58f9f8ae2133e4b6373fb58ea49634210fb379398706244f6b83b567ddb4f221" gracePeriod=30 Nov 22 08:19:37 crc kubenswrapper[4789]: I1122 08:19:37.254542 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 22 08:19:37 crc kubenswrapper[4789]: I1122 08:19:37.254863 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="88a4d40a-bffd-416e-a813-a2dd3ddcf47c" containerName="nova-scheduler-scheduler" containerID="cri-o://f984fc3abdac330f1fc81a8a4c0e461ee37ef7101c674d47b577b6b4b64761f9" gracePeriod=30 Nov 22 08:19:37 crc kubenswrapper[4789]: I1122 08:19:37.266432 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 22 08:19:37 crc kubenswrapper[4789]: I1122 08:19:37.271151 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="08796aa5-e098-4917-8eb3-dca574b4670a" containerName="nova-metadata-log" containerID="cri-o://725dfda767bf2e943a2e5632fba3109331a9370abb6b7cd8ff809b9175abb2d5" gracePeriod=30 Nov 22 08:19:37 crc kubenswrapper[4789]: I1122 08:19:37.271250 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="08796aa5-e098-4917-8eb3-dca574b4670a" containerName="nova-metadata-metadata" containerID="cri-o://e31cb8624b2f37521918d4a58cde08dbb6df0e2b3f29c3df0479fbaf3b71a3e0" gracePeriod=30 Nov 22 08:19:38 crc kubenswrapper[4789]: I1122 08:19:38.075987 4789 generic.go:334] "Generic (PLEG): container finished" podID="a7ffaa8c-24d2-4d9c-bf31-d231b4f6437e" containerID="fe3a2122b6c2fe085f2a827a9f023bbb1fea1049366684c22dc191b340fffd78" exitCode=143 Nov 22 08:19:38 crc kubenswrapper[4789]: I1122 08:19:38.076063 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a7ffaa8c-24d2-4d9c-bf31-d231b4f6437e","Type":"ContainerDied","Data":"fe3a2122b6c2fe085f2a827a9f023bbb1fea1049366684c22dc191b340fffd78"} Nov 22 08:19:38 crc kubenswrapper[4789]: I1122 08:19:38.078851 4789 generic.go:334] "Generic (PLEG): container finished" podID="08796aa5-e098-4917-8eb3-dca574b4670a" containerID="725dfda767bf2e943a2e5632fba3109331a9370abb6b7cd8ff809b9175abb2d5" exitCode=143 Nov 22 08:19:38 crc kubenswrapper[4789]: I1122 08:19:38.078881 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"08796aa5-e098-4917-8eb3-dca574b4670a","Type":"ContainerDied","Data":"725dfda767bf2e943a2e5632fba3109331a9370abb6b7cd8ff809b9175abb2d5"} Nov 22 08:19:40 crc kubenswrapper[4789]: I1122 08:19:40.332668 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="08796aa5-e098-4917-8eb3-dca574b4670a" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.181:8775/\": read tcp 10.217.0.2:33214->10.217.0.181:8775: read: connection reset by peer" Nov 22 08:19:40 crc kubenswrapper[4789]: I1122 08:19:40.332668 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="08796aa5-e098-4917-8eb3-dca574b4670a" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.181:8775/\": read tcp 10.217.0.2:33228->10.217.0.181:8775: read: connection reset by peer" Nov 22 08:19:40 crc kubenswrapper[4789]: I1122 08:19:40.780244 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 22 08:19:40 crc kubenswrapper[4789]: I1122 08:19:40.842486 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 22 08:19:40 crc kubenswrapper[4789]: I1122 08:19:40.869435 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a7ffaa8c-24d2-4d9c-bf31-d231b4f6437e-public-tls-certs\") pod \"a7ffaa8c-24d2-4d9c-bf31-d231b4f6437e\" (UID: \"a7ffaa8c-24d2-4d9c-bf31-d231b4f6437e\") " Nov 22 08:19:40 crc kubenswrapper[4789]: I1122 08:19:40.869513 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2z5dl\" (UniqueName: \"kubernetes.io/projected/a7ffaa8c-24d2-4d9c-bf31-d231b4f6437e-kube-api-access-2z5dl\") pod \"a7ffaa8c-24d2-4d9c-bf31-d231b4f6437e\" (UID: \"a7ffaa8c-24d2-4d9c-bf31-d231b4f6437e\") " Nov 22 08:19:40 crc kubenswrapper[4789]: I1122 08:19:40.869571 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08796aa5-e098-4917-8eb3-dca574b4670a-combined-ca-bundle\") pod \"08796aa5-e098-4917-8eb3-dca574b4670a\" (UID: \"08796aa5-e098-4917-8eb3-dca574b4670a\") " Nov 22 08:19:40 crc kubenswrapper[4789]: I1122 08:19:40.869605 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7ffaa8c-24d2-4d9c-bf31-d231b4f6437e-combined-ca-bundle\") pod \"a7ffaa8c-24d2-4d9c-bf31-d231b4f6437e\" (UID: \"a7ffaa8c-24d2-4d9c-bf31-d231b4f6437e\") " Nov 22 08:19:40 crc kubenswrapper[4789]: I1122 08:19:40.869655 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/08796aa5-e098-4917-8eb3-dca574b4670a-nova-metadata-tls-certs\") pod \"08796aa5-e098-4917-8eb3-dca574b4670a\" (UID: \"08796aa5-e098-4917-8eb3-dca574b4670a\") " Nov 22 08:19:40 crc kubenswrapper[4789]: I1122 08:19:40.869673 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/08796aa5-e098-4917-8eb3-dca574b4670a-logs\") pod \"08796aa5-e098-4917-8eb3-dca574b4670a\" (UID: \"08796aa5-e098-4917-8eb3-dca574b4670a\") " Nov 22 08:19:40 crc kubenswrapper[4789]: I1122 08:19:40.869825 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a7ffaa8c-24d2-4d9c-bf31-d231b4f6437e-logs\") pod \"a7ffaa8c-24d2-4d9c-bf31-d231b4f6437e\" (UID: \"a7ffaa8c-24d2-4d9c-bf31-d231b4f6437e\") " Nov 22 08:19:40 crc kubenswrapper[4789]: I1122 08:19:40.869894 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7ffaa8c-24d2-4d9c-bf31-d231b4f6437e-config-data\") pod \"a7ffaa8c-24d2-4d9c-bf31-d231b4f6437e\" (UID: \"a7ffaa8c-24d2-4d9c-bf31-d231b4f6437e\") " Nov 22 08:19:40 crc kubenswrapper[4789]: I1122 08:19:40.869966 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08796aa5-e098-4917-8eb3-dca574b4670a-config-data\") pod \"08796aa5-e098-4917-8eb3-dca574b4670a\" (UID: \"08796aa5-e098-4917-8eb3-dca574b4670a\") " Nov 22 08:19:40 crc kubenswrapper[4789]: I1122 08:19:40.869991 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a7ffaa8c-24d2-4d9c-bf31-d231b4f6437e-internal-tls-certs\") pod \"a7ffaa8c-24d2-4d9c-bf31-d231b4f6437e\" (UID: \"a7ffaa8c-24d2-4d9c-bf31-d231b4f6437e\") " Nov 22 08:19:40 crc kubenswrapper[4789]: I1122 08:19:40.870548 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a7ffaa8c-24d2-4d9c-bf31-d231b4f6437e-logs" (OuterVolumeSpecName: "logs") pod "a7ffaa8c-24d2-4d9c-bf31-d231b4f6437e" (UID: "a7ffaa8c-24d2-4d9c-bf31-d231b4f6437e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:19:40 crc kubenswrapper[4789]: I1122 08:19:40.871793 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4wzs4\" (UniqueName: \"kubernetes.io/projected/08796aa5-e098-4917-8eb3-dca574b4670a-kube-api-access-4wzs4\") pod \"08796aa5-e098-4917-8eb3-dca574b4670a\" (UID: \"08796aa5-e098-4917-8eb3-dca574b4670a\") " Nov 22 08:19:40 crc kubenswrapper[4789]: I1122 08:19:40.872413 4789 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a7ffaa8c-24d2-4d9c-bf31-d231b4f6437e-logs\") on node \"crc\" DevicePath \"\"" Nov 22 08:19:40 crc kubenswrapper[4789]: I1122 08:19:40.876076 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08796aa5-e098-4917-8eb3-dca574b4670a-kube-api-access-4wzs4" (OuterVolumeSpecName: "kube-api-access-4wzs4") pod "08796aa5-e098-4917-8eb3-dca574b4670a" (UID: "08796aa5-e098-4917-8eb3-dca574b4670a"). InnerVolumeSpecName "kube-api-access-4wzs4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:19:40 crc kubenswrapper[4789]: I1122 08:19:40.877095 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/08796aa5-e098-4917-8eb3-dca574b4670a-logs" (OuterVolumeSpecName: "logs") pod "08796aa5-e098-4917-8eb3-dca574b4670a" (UID: "08796aa5-e098-4917-8eb3-dca574b4670a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:19:40 crc kubenswrapper[4789]: I1122 08:19:40.889883 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7ffaa8c-24d2-4d9c-bf31-d231b4f6437e-kube-api-access-2z5dl" (OuterVolumeSpecName: "kube-api-access-2z5dl") pod "a7ffaa8c-24d2-4d9c-bf31-d231b4f6437e" (UID: "a7ffaa8c-24d2-4d9c-bf31-d231b4f6437e"). InnerVolumeSpecName "kube-api-access-2z5dl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:19:40 crc kubenswrapper[4789]: I1122 08:19:40.903494 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7ffaa8c-24d2-4d9c-bf31-d231b4f6437e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a7ffaa8c-24d2-4d9c-bf31-d231b4f6437e" (UID: "a7ffaa8c-24d2-4d9c-bf31-d231b4f6437e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:19:40 crc kubenswrapper[4789]: I1122 08:19:40.916179 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7ffaa8c-24d2-4d9c-bf31-d231b4f6437e-config-data" (OuterVolumeSpecName: "config-data") pod "a7ffaa8c-24d2-4d9c-bf31-d231b4f6437e" (UID: "a7ffaa8c-24d2-4d9c-bf31-d231b4f6437e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:19:40 crc kubenswrapper[4789]: I1122 08:19:40.925328 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08796aa5-e098-4917-8eb3-dca574b4670a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "08796aa5-e098-4917-8eb3-dca574b4670a" (UID: "08796aa5-e098-4917-8eb3-dca574b4670a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:19:40 crc kubenswrapper[4789]: I1122 08:19:40.926189 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08796aa5-e098-4917-8eb3-dca574b4670a-config-data" (OuterVolumeSpecName: "config-data") pod "08796aa5-e098-4917-8eb3-dca574b4670a" (UID: "08796aa5-e098-4917-8eb3-dca574b4670a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:19:40 crc kubenswrapper[4789]: I1122 08:19:40.933410 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7ffaa8c-24d2-4d9c-bf31-d231b4f6437e-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "a7ffaa8c-24d2-4d9c-bf31-d231b4f6437e" (UID: "a7ffaa8c-24d2-4d9c-bf31-d231b4f6437e"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:19:40 crc kubenswrapper[4789]: I1122 08:19:40.933707 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7ffaa8c-24d2-4d9c-bf31-d231b4f6437e-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "a7ffaa8c-24d2-4d9c-bf31-d231b4f6437e" (UID: "a7ffaa8c-24d2-4d9c-bf31-d231b4f6437e"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:19:40 crc kubenswrapper[4789]: I1122 08:19:40.935263 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08796aa5-e098-4917-8eb3-dca574b4670a-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "08796aa5-e098-4917-8eb3-dca574b4670a" (UID: "08796aa5-e098-4917-8eb3-dca574b4670a"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:19:40 crc kubenswrapper[4789]: I1122 08:19:40.976622 4789 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7ffaa8c-24d2-4d9c-bf31-d231b4f6437e-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:19:40 crc kubenswrapper[4789]: I1122 08:19:40.976673 4789 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08796aa5-e098-4917-8eb3-dca574b4670a-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:19:40 crc kubenswrapper[4789]: I1122 08:19:40.976683 4789 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a7ffaa8c-24d2-4d9c-bf31-d231b4f6437e-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 22 08:19:40 crc kubenswrapper[4789]: I1122 08:19:40.976695 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4wzs4\" (UniqueName: \"kubernetes.io/projected/08796aa5-e098-4917-8eb3-dca574b4670a-kube-api-access-4wzs4\") on node \"crc\" DevicePath \"\"" Nov 22 08:19:40 crc kubenswrapper[4789]: I1122 08:19:40.976707 4789 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a7ffaa8c-24d2-4d9c-bf31-d231b4f6437e-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 22 08:19:40 crc kubenswrapper[4789]: I1122 08:19:40.976717 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2z5dl\" (UniqueName: \"kubernetes.io/projected/a7ffaa8c-24d2-4d9c-bf31-d231b4f6437e-kube-api-access-2z5dl\") on node \"crc\" DevicePath \"\"" Nov 22 08:19:40 crc kubenswrapper[4789]: I1122 08:19:40.976726 4789 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08796aa5-e098-4917-8eb3-dca574b4670a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:19:40 crc kubenswrapper[4789]: I1122 08:19:40.976807 4789 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7ffaa8c-24d2-4d9c-bf31-d231b4f6437e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:19:40 crc kubenswrapper[4789]: I1122 08:19:40.976834 4789 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/08796aa5-e098-4917-8eb3-dca574b4670a-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 22 08:19:40 crc kubenswrapper[4789]: I1122 08:19:40.976847 4789 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/08796aa5-e098-4917-8eb3-dca574b4670a-logs\") on node \"crc\" DevicePath \"\"" Nov 22 08:19:41 crc kubenswrapper[4789]: I1122 08:19:41.107050 4789 generic.go:334] "Generic (PLEG): container finished" podID="88a4d40a-bffd-416e-a813-a2dd3ddcf47c" containerID="f984fc3abdac330f1fc81a8a4c0e461ee37ef7101c674d47b577b6b4b64761f9" exitCode=0 Nov 22 08:19:41 crc kubenswrapper[4789]: I1122 08:19:41.107131 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"88a4d40a-bffd-416e-a813-a2dd3ddcf47c","Type":"ContainerDied","Data":"f984fc3abdac330f1fc81a8a4c0e461ee37ef7101c674d47b577b6b4b64761f9"} Nov 22 08:19:41 crc kubenswrapper[4789]: I1122 08:19:41.112270 4789 generic.go:334] "Generic (PLEG): container finished" podID="08796aa5-e098-4917-8eb3-dca574b4670a" containerID="e31cb8624b2f37521918d4a58cde08dbb6df0e2b3f29c3df0479fbaf3b71a3e0" exitCode=0 Nov 22 08:19:41 crc kubenswrapper[4789]: I1122 08:19:41.112353 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"08796aa5-e098-4917-8eb3-dca574b4670a","Type":"ContainerDied","Data":"e31cb8624b2f37521918d4a58cde08dbb6df0e2b3f29c3df0479fbaf3b71a3e0"} Nov 22 08:19:41 crc kubenswrapper[4789]: I1122 08:19:41.112384 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"08796aa5-e098-4917-8eb3-dca574b4670a","Type":"ContainerDied","Data":"b5d645d57033db6a068d4d4f1f001a1baf3ffa53f8062e3a9af095ad32653cee"} Nov 22 08:19:41 crc kubenswrapper[4789]: I1122 08:19:41.112389 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 22 08:19:41 crc kubenswrapper[4789]: I1122 08:19:41.112404 4789 scope.go:117] "RemoveContainer" containerID="e31cb8624b2f37521918d4a58cde08dbb6df0e2b3f29c3df0479fbaf3b71a3e0" Nov 22 08:19:41 crc kubenswrapper[4789]: I1122 08:19:41.115926 4789 generic.go:334] "Generic (PLEG): container finished" podID="a7ffaa8c-24d2-4d9c-bf31-d231b4f6437e" containerID="58f9f8ae2133e4b6373fb58ea49634210fb379398706244f6b83b567ddb4f221" exitCode=0 Nov 22 08:19:41 crc kubenswrapper[4789]: I1122 08:19:41.115972 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a7ffaa8c-24d2-4d9c-bf31-d231b4f6437e","Type":"ContainerDied","Data":"58f9f8ae2133e4b6373fb58ea49634210fb379398706244f6b83b567ddb4f221"} Nov 22 08:19:41 crc kubenswrapper[4789]: I1122 08:19:41.115998 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a7ffaa8c-24d2-4d9c-bf31-d231b4f6437e","Type":"ContainerDied","Data":"90367a61cf3c546410a50628442d2d71190a291493c47bfa9f71f589732cacc9"} Nov 22 08:19:41 crc kubenswrapper[4789]: I1122 08:19:41.116056 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 22 08:19:41 crc kubenswrapper[4789]: I1122 08:19:41.166294 4789 scope.go:117] "RemoveContainer" containerID="725dfda767bf2e943a2e5632fba3109331a9370abb6b7cd8ff809b9175abb2d5" Nov 22 08:19:41 crc kubenswrapper[4789]: I1122 08:19:41.175563 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 22 08:19:41 crc kubenswrapper[4789]: I1122 08:19:41.201098 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 22 08:19:41 crc kubenswrapper[4789]: I1122 08:19:41.218838 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 22 08:19:41 crc kubenswrapper[4789]: I1122 08:19:41.228725 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 22 08:19:41 crc kubenswrapper[4789]: I1122 08:19:41.238774 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 22 08:19:41 crc kubenswrapper[4789]: E1122 08:19:41.239341 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08796aa5-e098-4917-8eb3-dca574b4670a" containerName="nova-metadata-log" Nov 22 08:19:41 crc kubenswrapper[4789]: I1122 08:19:41.239366 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="08796aa5-e098-4917-8eb3-dca574b4670a" containerName="nova-metadata-log" Nov 22 08:19:41 crc kubenswrapper[4789]: E1122 08:19:41.239405 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7ffaa8c-24d2-4d9c-bf31-d231b4f6437e" containerName="nova-api-log" Nov 22 08:19:41 crc kubenswrapper[4789]: I1122 08:19:41.239413 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7ffaa8c-24d2-4d9c-bf31-d231b4f6437e" containerName="nova-api-log" Nov 22 08:19:41 crc kubenswrapper[4789]: E1122 08:19:41.239430 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7ffaa8c-24d2-4d9c-bf31-d231b4f6437e" containerName="nova-api-api" Nov 22 08:19:41 crc kubenswrapper[4789]: I1122 08:19:41.239439 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7ffaa8c-24d2-4d9c-bf31-d231b4f6437e" containerName="nova-api-api" Nov 22 08:19:41 crc kubenswrapper[4789]: E1122 08:19:41.239457 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="206a90d7-1c1c-4a17-992a-43bd718fc9ea" containerName="nova-manage" Nov 22 08:19:41 crc kubenswrapper[4789]: I1122 08:19:41.239465 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="206a90d7-1c1c-4a17-992a-43bd718fc9ea" containerName="nova-manage" Nov 22 08:19:41 crc kubenswrapper[4789]: E1122 08:19:41.239478 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75e7c40c-ba8c-45f6-b30c-f654f48956ca" containerName="dnsmasq-dns" Nov 22 08:19:41 crc kubenswrapper[4789]: I1122 08:19:41.239485 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="75e7c40c-ba8c-45f6-b30c-f654f48956ca" containerName="dnsmasq-dns" Nov 22 08:19:41 crc kubenswrapper[4789]: E1122 08:19:41.239507 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08796aa5-e098-4917-8eb3-dca574b4670a" containerName="nova-metadata-metadata" Nov 22 08:19:41 crc kubenswrapper[4789]: I1122 08:19:41.239514 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="08796aa5-e098-4917-8eb3-dca574b4670a" containerName="nova-metadata-metadata" Nov 22 08:19:41 crc kubenswrapper[4789]: E1122 08:19:41.239527 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75e7c40c-ba8c-45f6-b30c-f654f48956ca" containerName="init" Nov 22 08:19:41 crc kubenswrapper[4789]: I1122 08:19:41.239534 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="75e7c40c-ba8c-45f6-b30c-f654f48956ca" containerName="init" Nov 22 08:19:41 crc kubenswrapper[4789]: I1122 08:19:41.239828 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7ffaa8c-24d2-4d9c-bf31-d231b4f6437e" containerName="nova-api-log" Nov 22 08:19:41 crc kubenswrapper[4789]: I1122 08:19:41.239855 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="08796aa5-e098-4917-8eb3-dca574b4670a" containerName="nova-metadata-metadata" Nov 22 08:19:41 crc kubenswrapper[4789]: I1122 08:19:41.239876 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7ffaa8c-24d2-4d9c-bf31-d231b4f6437e" containerName="nova-api-api" Nov 22 08:19:41 crc kubenswrapper[4789]: I1122 08:19:41.239897 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="08796aa5-e098-4917-8eb3-dca574b4670a" containerName="nova-metadata-log" Nov 22 08:19:41 crc kubenswrapper[4789]: I1122 08:19:41.239911 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="75e7c40c-ba8c-45f6-b30c-f654f48956ca" containerName="dnsmasq-dns" Nov 22 08:19:41 crc kubenswrapper[4789]: I1122 08:19:41.239923 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="206a90d7-1c1c-4a17-992a-43bd718fc9ea" containerName="nova-manage" Nov 22 08:19:41 crc kubenswrapper[4789]: I1122 08:19:41.241256 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 22 08:19:41 crc kubenswrapper[4789]: I1122 08:19:41.243743 4789 scope.go:117] "RemoveContainer" containerID="e31cb8624b2f37521918d4a58cde08dbb6df0e2b3f29c3df0479fbaf3b71a3e0" Nov 22 08:19:41 crc kubenswrapper[4789]: E1122 08:19:41.244228 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e31cb8624b2f37521918d4a58cde08dbb6df0e2b3f29c3df0479fbaf3b71a3e0\": container with ID starting with e31cb8624b2f37521918d4a58cde08dbb6df0e2b3f29c3df0479fbaf3b71a3e0 not found: ID does not exist" containerID="e31cb8624b2f37521918d4a58cde08dbb6df0e2b3f29c3df0479fbaf3b71a3e0" Nov 22 08:19:41 crc kubenswrapper[4789]: I1122 08:19:41.244295 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e31cb8624b2f37521918d4a58cde08dbb6df0e2b3f29c3df0479fbaf3b71a3e0"} err="failed to get container status \"e31cb8624b2f37521918d4a58cde08dbb6df0e2b3f29c3df0479fbaf3b71a3e0\": rpc error: code = NotFound desc = could not find container \"e31cb8624b2f37521918d4a58cde08dbb6df0e2b3f29c3df0479fbaf3b71a3e0\": container with ID starting with e31cb8624b2f37521918d4a58cde08dbb6df0e2b3f29c3df0479fbaf3b71a3e0 not found: ID does not exist" Nov 22 08:19:41 crc kubenswrapper[4789]: I1122 08:19:41.244322 4789 scope.go:117] "RemoveContainer" containerID="725dfda767bf2e943a2e5632fba3109331a9370abb6b7cd8ff809b9175abb2d5" Nov 22 08:19:41 crc kubenswrapper[4789]: I1122 08:19:41.244247 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 22 08:19:41 crc kubenswrapper[4789]: I1122 08:19:41.245247 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 22 08:19:41 crc kubenswrapper[4789]: E1122 08:19:41.245401 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"725dfda767bf2e943a2e5632fba3109331a9370abb6b7cd8ff809b9175abb2d5\": container with ID starting with 725dfda767bf2e943a2e5632fba3109331a9370abb6b7cd8ff809b9175abb2d5 not found: ID does not exist" containerID="725dfda767bf2e943a2e5632fba3109331a9370abb6b7cd8ff809b9175abb2d5" Nov 22 08:19:41 crc kubenswrapper[4789]: I1122 08:19:41.245439 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"725dfda767bf2e943a2e5632fba3109331a9370abb6b7cd8ff809b9175abb2d5"} err="failed to get container status \"725dfda767bf2e943a2e5632fba3109331a9370abb6b7cd8ff809b9175abb2d5\": rpc error: code = NotFound desc = could not find container \"725dfda767bf2e943a2e5632fba3109331a9370abb6b7cd8ff809b9175abb2d5\": container with ID starting with 725dfda767bf2e943a2e5632fba3109331a9370abb6b7cd8ff809b9175abb2d5 not found: ID does not exist" Nov 22 08:19:41 crc kubenswrapper[4789]: I1122 08:19:41.245467 4789 scope.go:117] "RemoveContainer" containerID="58f9f8ae2133e4b6373fb58ea49634210fb379398706244f6b83b567ddb4f221" Nov 22 08:19:41 crc kubenswrapper[4789]: I1122 08:19:41.248296 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 22 08:19:41 crc kubenswrapper[4789]: I1122 08:19:41.250610 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 22 08:19:41 crc kubenswrapper[4789]: I1122 08:19:41.254072 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 22 08:19:41 crc kubenswrapper[4789]: I1122 08:19:41.254202 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Nov 22 08:19:41 crc kubenswrapper[4789]: I1122 08:19:41.254415 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Nov 22 08:19:41 crc kubenswrapper[4789]: I1122 08:19:41.260564 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 22 08:19:41 crc kubenswrapper[4789]: I1122 08:19:41.273275 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 22 08:19:41 crc kubenswrapper[4789]: I1122 08:19:41.280453 4789 scope.go:117] "RemoveContainer" containerID="fe3a2122b6c2fe085f2a827a9f023bbb1fea1049366684c22dc191b340fffd78" Nov 22 08:19:41 crc kubenswrapper[4789]: I1122 08:19:41.281343 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 22 08:19:41 crc kubenswrapper[4789]: I1122 08:19:41.303928 4789 scope.go:117] "RemoveContainer" containerID="58f9f8ae2133e4b6373fb58ea49634210fb379398706244f6b83b567ddb4f221" Nov 22 08:19:41 crc kubenswrapper[4789]: E1122 08:19:41.304595 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"58f9f8ae2133e4b6373fb58ea49634210fb379398706244f6b83b567ddb4f221\": container with ID starting with 58f9f8ae2133e4b6373fb58ea49634210fb379398706244f6b83b567ddb4f221 not found: ID does not exist" containerID="58f9f8ae2133e4b6373fb58ea49634210fb379398706244f6b83b567ddb4f221" Nov 22 08:19:41 crc kubenswrapper[4789]: I1122 08:19:41.304625 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"58f9f8ae2133e4b6373fb58ea49634210fb379398706244f6b83b567ddb4f221"} err="failed to get container status \"58f9f8ae2133e4b6373fb58ea49634210fb379398706244f6b83b567ddb4f221\": rpc error: code = NotFound desc = could not find container \"58f9f8ae2133e4b6373fb58ea49634210fb379398706244f6b83b567ddb4f221\": container with ID starting with 58f9f8ae2133e4b6373fb58ea49634210fb379398706244f6b83b567ddb4f221 not found: ID does not exist" Nov 22 08:19:41 crc kubenswrapper[4789]: I1122 08:19:41.304645 4789 scope.go:117] "RemoveContainer" containerID="fe3a2122b6c2fe085f2a827a9f023bbb1fea1049366684c22dc191b340fffd78" Nov 22 08:19:41 crc kubenswrapper[4789]: E1122 08:19:41.305718 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fe3a2122b6c2fe085f2a827a9f023bbb1fea1049366684c22dc191b340fffd78\": container with ID starting with fe3a2122b6c2fe085f2a827a9f023bbb1fea1049366684c22dc191b340fffd78 not found: ID does not exist" containerID="fe3a2122b6c2fe085f2a827a9f023bbb1fea1049366684c22dc191b340fffd78" Nov 22 08:19:41 crc kubenswrapper[4789]: I1122 08:19:41.305744 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe3a2122b6c2fe085f2a827a9f023bbb1fea1049366684c22dc191b340fffd78"} err="failed to get container status \"fe3a2122b6c2fe085f2a827a9f023bbb1fea1049366684c22dc191b340fffd78\": rpc error: code = NotFound desc = could not find container \"fe3a2122b6c2fe085f2a827a9f023bbb1fea1049366684c22dc191b340fffd78\": container with ID starting with fe3a2122b6c2fe085f2a827a9f023bbb1fea1049366684c22dc191b340fffd78 not found: ID does not exist" Nov 22 08:19:41 crc kubenswrapper[4789]: I1122 08:19:41.384182 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88a4d40a-bffd-416e-a813-a2dd3ddcf47c-combined-ca-bundle\") pod \"88a4d40a-bffd-416e-a813-a2dd3ddcf47c\" (UID: \"88a4d40a-bffd-416e-a813-a2dd3ddcf47c\") " Nov 22 08:19:41 crc kubenswrapper[4789]: I1122 08:19:41.384251 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tff8g\" (UniqueName: \"kubernetes.io/projected/88a4d40a-bffd-416e-a813-a2dd3ddcf47c-kube-api-access-tff8g\") pod \"88a4d40a-bffd-416e-a813-a2dd3ddcf47c\" (UID: \"88a4d40a-bffd-416e-a813-a2dd3ddcf47c\") " Nov 22 08:19:41 crc kubenswrapper[4789]: I1122 08:19:41.384396 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88a4d40a-bffd-416e-a813-a2dd3ddcf47c-config-data\") pod \"88a4d40a-bffd-416e-a813-a2dd3ddcf47c\" (UID: \"88a4d40a-bffd-416e-a813-a2dd3ddcf47c\") " Nov 22 08:19:41 crc kubenswrapper[4789]: I1122 08:19:41.384674 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/588fc366-e2d8-40bc-b09e-ebcf032e440f-config-data\") pod \"nova-metadata-0\" (UID: \"588fc366-e2d8-40bc-b09e-ebcf032e440f\") " pod="openstack/nova-metadata-0" Nov 22 08:19:41 crc kubenswrapper[4789]: I1122 08:19:41.384702 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/588fc366-e2d8-40bc-b09e-ebcf032e440f-logs\") pod \"nova-metadata-0\" (UID: \"588fc366-e2d8-40bc-b09e-ebcf032e440f\") " pod="openstack/nova-metadata-0" Nov 22 08:19:41 crc kubenswrapper[4789]: I1122 08:19:41.384732 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9585ddf1-5799-43c3-a747-860ef1bf8380-logs\") pod \"nova-api-0\" (UID: \"9585ddf1-5799-43c3-a747-860ef1bf8380\") " pod="openstack/nova-api-0" Nov 22 08:19:41 crc kubenswrapper[4789]: I1122 08:19:41.384752 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9585ddf1-5799-43c3-a747-860ef1bf8380-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"9585ddf1-5799-43c3-a747-860ef1bf8380\") " pod="openstack/nova-api-0" Nov 22 08:19:41 crc kubenswrapper[4789]: I1122 08:19:41.384816 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hfk72\" (UniqueName: \"kubernetes.io/projected/9585ddf1-5799-43c3-a747-860ef1bf8380-kube-api-access-hfk72\") pod \"nova-api-0\" (UID: \"9585ddf1-5799-43c3-a747-860ef1bf8380\") " pod="openstack/nova-api-0" Nov 22 08:19:41 crc kubenswrapper[4789]: I1122 08:19:41.384844 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/588fc366-e2d8-40bc-b09e-ebcf032e440f-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"588fc366-e2d8-40bc-b09e-ebcf032e440f\") " pod="openstack/nova-metadata-0" Nov 22 08:19:41 crc kubenswrapper[4789]: I1122 08:19:41.384869 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9585ddf1-5799-43c3-a747-860ef1bf8380-internal-tls-certs\") pod \"nova-api-0\" (UID: \"9585ddf1-5799-43c3-a747-860ef1bf8380\") " pod="openstack/nova-api-0" Nov 22 08:19:41 crc kubenswrapper[4789]: I1122 08:19:41.384888 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9585ddf1-5799-43c3-a747-860ef1bf8380-public-tls-certs\") pod \"nova-api-0\" (UID: \"9585ddf1-5799-43c3-a747-860ef1bf8380\") " pod="openstack/nova-api-0" Nov 22 08:19:41 crc kubenswrapper[4789]: I1122 08:19:41.385479 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9585ddf1-5799-43c3-a747-860ef1bf8380-config-data\") pod \"nova-api-0\" (UID: \"9585ddf1-5799-43c3-a747-860ef1bf8380\") " pod="openstack/nova-api-0" Nov 22 08:19:41 crc kubenswrapper[4789]: I1122 08:19:41.385512 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/588fc366-e2d8-40bc-b09e-ebcf032e440f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"588fc366-e2d8-40bc-b09e-ebcf032e440f\") " pod="openstack/nova-metadata-0" Nov 22 08:19:41 crc kubenswrapper[4789]: I1122 08:19:41.385540 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z4nm2\" (UniqueName: \"kubernetes.io/projected/588fc366-e2d8-40bc-b09e-ebcf032e440f-kube-api-access-z4nm2\") pod \"nova-metadata-0\" (UID: \"588fc366-e2d8-40bc-b09e-ebcf032e440f\") " pod="openstack/nova-metadata-0" Nov 22 08:19:41 crc kubenswrapper[4789]: I1122 08:19:41.387963 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88a4d40a-bffd-416e-a813-a2dd3ddcf47c-kube-api-access-tff8g" (OuterVolumeSpecName: "kube-api-access-tff8g") pod "88a4d40a-bffd-416e-a813-a2dd3ddcf47c" (UID: "88a4d40a-bffd-416e-a813-a2dd3ddcf47c"). InnerVolumeSpecName "kube-api-access-tff8g". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:19:41 crc kubenswrapper[4789]: I1122 08:19:41.409924 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88a4d40a-bffd-416e-a813-a2dd3ddcf47c-config-data" (OuterVolumeSpecName: "config-data") pod "88a4d40a-bffd-416e-a813-a2dd3ddcf47c" (UID: "88a4d40a-bffd-416e-a813-a2dd3ddcf47c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:19:41 crc kubenswrapper[4789]: I1122 08:19:41.411256 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88a4d40a-bffd-416e-a813-a2dd3ddcf47c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "88a4d40a-bffd-416e-a813-a2dd3ddcf47c" (UID: "88a4d40a-bffd-416e-a813-a2dd3ddcf47c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:19:41 crc kubenswrapper[4789]: I1122 08:19:41.486972 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/588fc366-e2d8-40bc-b09e-ebcf032e440f-config-data\") pod \"nova-metadata-0\" (UID: \"588fc366-e2d8-40bc-b09e-ebcf032e440f\") " pod="openstack/nova-metadata-0" Nov 22 08:19:41 crc kubenswrapper[4789]: I1122 08:19:41.487022 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/588fc366-e2d8-40bc-b09e-ebcf032e440f-logs\") pod \"nova-metadata-0\" (UID: \"588fc366-e2d8-40bc-b09e-ebcf032e440f\") " pod="openstack/nova-metadata-0" Nov 22 08:19:41 crc kubenswrapper[4789]: I1122 08:19:41.487051 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9585ddf1-5799-43c3-a747-860ef1bf8380-logs\") pod \"nova-api-0\" (UID: \"9585ddf1-5799-43c3-a747-860ef1bf8380\") " pod="openstack/nova-api-0" Nov 22 08:19:41 crc kubenswrapper[4789]: I1122 08:19:41.487067 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9585ddf1-5799-43c3-a747-860ef1bf8380-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"9585ddf1-5799-43c3-a747-860ef1bf8380\") " pod="openstack/nova-api-0" Nov 22 08:19:41 crc kubenswrapper[4789]: I1122 08:19:41.487095 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hfk72\" (UniqueName: \"kubernetes.io/projected/9585ddf1-5799-43c3-a747-860ef1bf8380-kube-api-access-hfk72\") pod \"nova-api-0\" (UID: \"9585ddf1-5799-43c3-a747-860ef1bf8380\") " pod="openstack/nova-api-0" Nov 22 08:19:41 crc kubenswrapper[4789]: I1122 08:19:41.487119 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/588fc366-e2d8-40bc-b09e-ebcf032e440f-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"588fc366-e2d8-40bc-b09e-ebcf032e440f\") " pod="openstack/nova-metadata-0" Nov 22 08:19:41 crc kubenswrapper[4789]: I1122 08:19:41.487144 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9585ddf1-5799-43c3-a747-860ef1bf8380-internal-tls-certs\") pod \"nova-api-0\" (UID: \"9585ddf1-5799-43c3-a747-860ef1bf8380\") " pod="openstack/nova-api-0" Nov 22 08:19:41 crc kubenswrapper[4789]: I1122 08:19:41.487159 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9585ddf1-5799-43c3-a747-860ef1bf8380-public-tls-certs\") pod \"nova-api-0\" (UID: \"9585ddf1-5799-43c3-a747-860ef1bf8380\") " pod="openstack/nova-api-0" Nov 22 08:19:41 crc kubenswrapper[4789]: I1122 08:19:41.487186 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9585ddf1-5799-43c3-a747-860ef1bf8380-config-data\") pod \"nova-api-0\" (UID: \"9585ddf1-5799-43c3-a747-860ef1bf8380\") " pod="openstack/nova-api-0" Nov 22 08:19:41 crc kubenswrapper[4789]: I1122 08:19:41.487219 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/588fc366-e2d8-40bc-b09e-ebcf032e440f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"588fc366-e2d8-40bc-b09e-ebcf032e440f\") " pod="openstack/nova-metadata-0" Nov 22 08:19:41 crc kubenswrapper[4789]: I1122 08:19:41.487252 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z4nm2\" (UniqueName: \"kubernetes.io/projected/588fc366-e2d8-40bc-b09e-ebcf032e440f-kube-api-access-z4nm2\") pod \"nova-metadata-0\" (UID: \"588fc366-e2d8-40bc-b09e-ebcf032e440f\") " pod="openstack/nova-metadata-0" Nov 22 08:19:41 crc kubenswrapper[4789]: I1122 08:19:41.487316 4789 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88a4d40a-bffd-416e-a813-a2dd3ddcf47c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:19:41 crc kubenswrapper[4789]: I1122 08:19:41.487327 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tff8g\" (UniqueName: \"kubernetes.io/projected/88a4d40a-bffd-416e-a813-a2dd3ddcf47c-kube-api-access-tff8g\") on node \"crc\" DevicePath \"\"" Nov 22 08:19:41 crc kubenswrapper[4789]: I1122 08:19:41.487339 4789 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88a4d40a-bffd-416e-a813-a2dd3ddcf47c-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:19:41 crc kubenswrapper[4789]: I1122 08:19:41.487531 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/588fc366-e2d8-40bc-b09e-ebcf032e440f-logs\") pod \"nova-metadata-0\" (UID: \"588fc366-e2d8-40bc-b09e-ebcf032e440f\") " pod="openstack/nova-metadata-0" Nov 22 08:19:41 crc kubenswrapper[4789]: I1122 08:19:41.488274 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9585ddf1-5799-43c3-a747-860ef1bf8380-logs\") pod \"nova-api-0\" (UID: \"9585ddf1-5799-43c3-a747-860ef1bf8380\") " pod="openstack/nova-api-0" Nov 22 08:19:41 crc kubenswrapper[4789]: I1122 08:19:41.490907 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9585ddf1-5799-43c3-a747-860ef1bf8380-config-data\") pod \"nova-api-0\" (UID: \"9585ddf1-5799-43c3-a747-860ef1bf8380\") " pod="openstack/nova-api-0" Nov 22 08:19:41 crc kubenswrapper[4789]: I1122 08:19:41.491231 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9585ddf1-5799-43c3-a747-860ef1bf8380-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"9585ddf1-5799-43c3-a747-860ef1bf8380\") " pod="openstack/nova-api-0" Nov 22 08:19:41 crc kubenswrapper[4789]: I1122 08:19:41.491636 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/588fc366-e2d8-40bc-b09e-ebcf032e440f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"588fc366-e2d8-40bc-b09e-ebcf032e440f\") " pod="openstack/nova-metadata-0" Nov 22 08:19:41 crc kubenswrapper[4789]: I1122 08:19:41.492320 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9585ddf1-5799-43c3-a747-860ef1bf8380-internal-tls-certs\") pod \"nova-api-0\" (UID: \"9585ddf1-5799-43c3-a747-860ef1bf8380\") " pod="openstack/nova-api-0" Nov 22 08:19:41 crc kubenswrapper[4789]: I1122 08:19:41.492985 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/588fc366-e2d8-40bc-b09e-ebcf032e440f-config-data\") pod \"nova-metadata-0\" (UID: \"588fc366-e2d8-40bc-b09e-ebcf032e440f\") " pod="openstack/nova-metadata-0" Nov 22 08:19:41 crc kubenswrapper[4789]: I1122 08:19:41.498160 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9585ddf1-5799-43c3-a747-860ef1bf8380-public-tls-certs\") pod \"nova-api-0\" (UID: \"9585ddf1-5799-43c3-a747-860ef1bf8380\") " pod="openstack/nova-api-0" Nov 22 08:19:41 crc kubenswrapper[4789]: I1122 08:19:41.498998 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/588fc366-e2d8-40bc-b09e-ebcf032e440f-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"588fc366-e2d8-40bc-b09e-ebcf032e440f\") " pod="openstack/nova-metadata-0" Nov 22 08:19:41 crc kubenswrapper[4789]: I1122 08:19:41.505696 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z4nm2\" (UniqueName: \"kubernetes.io/projected/588fc366-e2d8-40bc-b09e-ebcf032e440f-kube-api-access-z4nm2\") pod \"nova-metadata-0\" (UID: \"588fc366-e2d8-40bc-b09e-ebcf032e440f\") " pod="openstack/nova-metadata-0" Nov 22 08:19:41 crc kubenswrapper[4789]: I1122 08:19:41.511040 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hfk72\" (UniqueName: \"kubernetes.io/projected/9585ddf1-5799-43c3-a747-860ef1bf8380-kube-api-access-hfk72\") pod \"nova-api-0\" (UID: \"9585ddf1-5799-43c3-a747-860ef1bf8380\") " pod="openstack/nova-api-0" Nov 22 08:19:41 crc kubenswrapper[4789]: I1122 08:19:41.562256 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 22 08:19:41 crc kubenswrapper[4789]: I1122 08:19:41.576014 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 22 08:19:41 crc kubenswrapper[4789]: I1122 08:19:41.974845 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="08796aa5-e098-4917-8eb3-dca574b4670a" path="/var/lib/kubelet/pods/08796aa5-e098-4917-8eb3-dca574b4670a/volumes" Nov 22 08:19:41 crc kubenswrapper[4789]: I1122 08:19:41.975756 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a7ffaa8c-24d2-4d9c-bf31-d231b4f6437e" path="/var/lib/kubelet/pods/a7ffaa8c-24d2-4d9c-bf31-d231b4f6437e/volumes" Nov 22 08:19:42 crc kubenswrapper[4789]: I1122 08:19:42.013096 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 22 08:19:42 crc kubenswrapper[4789]: I1122 08:19:42.070454 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 22 08:19:42 crc kubenswrapper[4789]: W1122 08:19:42.071557 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9585ddf1_5799_43c3_a747_860ef1bf8380.slice/crio-c87dd05fb60eed4b302120f2dce619cf332c3c1af83ea2c29991d1f00cf20865 WatchSource:0}: Error finding container c87dd05fb60eed4b302120f2dce619cf332c3c1af83ea2c29991d1f00cf20865: Status 404 returned error can't find the container with id c87dd05fb60eed4b302120f2dce619cf332c3c1af83ea2c29991d1f00cf20865 Nov 22 08:19:42 crc kubenswrapper[4789]: I1122 08:19:42.127836 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9585ddf1-5799-43c3-a747-860ef1bf8380","Type":"ContainerStarted","Data":"c87dd05fb60eed4b302120f2dce619cf332c3c1af83ea2c29991d1f00cf20865"} Nov 22 08:19:42 crc kubenswrapper[4789]: I1122 08:19:42.130010 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"588fc366-e2d8-40bc-b09e-ebcf032e440f","Type":"ContainerStarted","Data":"de207e909914c4a3b6c070aff17e00b9c1dd5aed0e8b87ef055107725908f8fe"} Nov 22 08:19:42 crc kubenswrapper[4789]: I1122 08:19:42.132211 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"88a4d40a-bffd-416e-a813-a2dd3ddcf47c","Type":"ContainerDied","Data":"aa984d4d1a30af98cecf4aa9492ab6e6e910fc2f1b87761a0cd618d94808551a"} Nov 22 08:19:42 crc kubenswrapper[4789]: I1122 08:19:42.132290 4789 scope.go:117] "RemoveContainer" containerID="f984fc3abdac330f1fc81a8a4c0e461ee37ef7101c674d47b577b6b4b64761f9" Nov 22 08:19:42 crc kubenswrapper[4789]: I1122 08:19:42.132308 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 22 08:19:42 crc kubenswrapper[4789]: I1122 08:19:42.189070 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 22 08:19:42 crc kubenswrapper[4789]: I1122 08:19:42.198647 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Nov 22 08:19:42 crc kubenswrapper[4789]: I1122 08:19:42.206936 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 22 08:19:42 crc kubenswrapper[4789]: E1122 08:19:42.207387 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88a4d40a-bffd-416e-a813-a2dd3ddcf47c" containerName="nova-scheduler-scheduler" Nov 22 08:19:42 crc kubenswrapper[4789]: I1122 08:19:42.207401 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="88a4d40a-bffd-416e-a813-a2dd3ddcf47c" containerName="nova-scheduler-scheduler" Nov 22 08:19:42 crc kubenswrapper[4789]: I1122 08:19:42.207616 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="88a4d40a-bffd-416e-a813-a2dd3ddcf47c" containerName="nova-scheduler-scheduler" Nov 22 08:19:42 crc kubenswrapper[4789]: I1122 08:19:42.208313 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 22 08:19:42 crc kubenswrapper[4789]: I1122 08:19:42.210674 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 22 08:19:42 crc kubenswrapper[4789]: I1122 08:19:42.216015 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 22 08:19:42 crc kubenswrapper[4789]: I1122 08:19:42.304486 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6vbmm\" (UniqueName: \"kubernetes.io/projected/a1cea96e-2c65-4231-a4b4-bf6d197c5981-kube-api-access-6vbmm\") pod \"nova-scheduler-0\" (UID: \"a1cea96e-2c65-4231-a4b4-bf6d197c5981\") " pod="openstack/nova-scheduler-0" Nov 22 08:19:42 crc kubenswrapper[4789]: I1122 08:19:42.304622 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1cea96e-2c65-4231-a4b4-bf6d197c5981-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"a1cea96e-2c65-4231-a4b4-bf6d197c5981\") " pod="openstack/nova-scheduler-0" Nov 22 08:19:42 crc kubenswrapper[4789]: I1122 08:19:42.304878 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1cea96e-2c65-4231-a4b4-bf6d197c5981-config-data\") pod \"nova-scheduler-0\" (UID: \"a1cea96e-2c65-4231-a4b4-bf6d197c5981\") " pod="openstack/nova-scheduler-0" Nov 22 08:19:42 crc kubenswrapper[4789]: I1122 08:19:42.406649 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1cea96e-2c65-4231-a4b4-bf6d197c5981-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"a1cea96e-2c65-4231-a4b4-bf6d197c5981\") " pod="openstack/nova-scheduler-0" Nov 22 08:19:42 crc kubenswrapper[4789]: I1122 08:19:42.406886 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1cea96e-2c65-4231-a4b4-bf6d197c5981-config-data\") pod \"nova-scheduler-0\" (UID: \"a1cea96e-2c65-4231-a4b4-bf6d197c5981\") " pod="openstack/nova-scheduler-0" Nov 22 08:19:42 crc kubenswrapper[4789]: I1122 08:19:42.406918 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6vbmm\" (UniqueName: \"kubernetes.io/projected/a1cea96e-2c65-4231-a4b4-bf6d197c5981-kube-api-access-6vbmm\") pod \"nova-scheduler-0\" (UID: \"a1cea96e-2c65-4231-a4b4-bf6d197c5981\") " pod="openstack/nova-scheduler-0" Nov 22 08:19:42 crc kubenswrapper[4789]: I1122 08:19:42.411166 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1cea96e-2c65-4231-a4b4-bf6d197c5981-config-data\") pod \"nova-scheduler-0\" (UID: \"a1cea96e-2c65-4231-a4b4-bf6d197c5981\") " pod="openstack/nova-scheduler-0" Nov 22 08:19:42 crc kubenswrapper[4789]: I1122 08:19:42.412031 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1cea96e-2c65-4231-a4b4-bf6d197c5981-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"a1cea96e-2c65-4231-a4b4-bf6d197c5981\") " pod="openstack/nova-scheduler-0" Nov 22 08:19:42 crc kubenswrapper[4789]: I1122 08:19:42.427237 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6vbmm\" (UniqueName: \"kubernetes.io/projected/a1cea96e-2c65-4231-a4b4-bf6d197c5981-kube-api-access-6vbmm\") pod \"nova-scheduler-0\" (UID: \"a1cea96e-2c65-4231-a4b4-bf6d197c5981\") " pod="openstack/nova-scheduler-0" Nov 22 08:19:42 crc kubenswrapper[4789]: I1122 08:19:42.532840 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 22 08:19:42 crc kubenswrapper[4789]: I1122 08:19:42.956094 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 22 08:19:42 crc kubenswrapper[4789]: W1122 08:19:42.957491 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda1cea96e_2c65_4231_a4b4_bf6d197c5981.slice/crio-8737ff358c4319aabdd07f96fa6d33752578fb8c0af2363980e41c73b055d548 WatchSource:0}: Error finding container 8737ff358c4319aabdd07f96fa6d33752578fb8c0af2363980e41c73b055d548: Status 404 returned error can't find the container with id 8737ff358c4319aabdd07f96fa6d33752578fb8c0af2363980e41c73b055d548 Nov 22 08:19:43 crc kubenswrapper[4789]: I1122 08:19:43.143381 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"588fc366-e2d8-40bc-b09e-ebcf032e440f","Type":"ContainerStarted","Data":"ef579abe03a36f178f1b07c939e5f624f61894e623bba98c2cf46316b7c3bc99"} Nov 22 08:19:43 crc kubenswrapper[4789]: I1122 08:19:43.143786 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"588fc366-e2d8-40bc-b09e-ebcf032e440f","Type":"ContainerStarted","Data":"b34f3036a3aee8a1669527ff8be956d02f211aedd16b202baccdd4103b40fbe0"} Nov 22 08:19:43 crc kubenswrapper[4789]: I1122 08:19:43.145038 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a1cea96e-2c65-4231-a4b4-bf6d197c5981","Type":"ContainerStarted","Data":"8737ff358c4319aabdd07f96fa6d33752578fb8c0af2363980e41c73b055d548"} Nov 22 08:19:43 crc kubenswrapper[4789]: I1122 08:19:43.158685 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9585ddf1-5799-43c3-a747-860ef1bf8380","Type":"ContainerStarted","Data":"b41d0c7c84cd54f960611daf6bb91ccdebb28596abd1d0401367930241efe539"} Nov 22 08:19:43 crc kubenswrapper[4789]: I1122 08:19:43.162865 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9585ddf1-5799-43c3-a747-860ef1bf8380","Type":"ContainerStarted","Data":"39e0e3e92c9d4ff2d6a56ec395d0a0afa65d4c714e7bcf9f465846bf8cd5da94"} Nov 22 08:19:43 crc kubenswrapper[4789]: I1122 08:19:43.173525 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.173507689 podStartE2EDuration="2.173507689s" podCreationTimestamp="2025-11-22 08:19:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:19:43.165315066 +0000 UTC m=+1397.399715339" watchObservedRunningTime="2025-11-22 08:19:43.173507689 +0000 UTC m=+1397.407907962" Nov 22 08:19:43 crc kubenswrapper[4789]: I1122 08:19:43.186408 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.18637447 podStartE2EDuration="2.18637447s" podCreationTimestamp="2025-11-22 08:19:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:19:43.186189365 +0000 UTC m=+1397.420589648" watchObservedRunningTime="2025-11-22 08:19:43.18637447 +0000 UTC m=+1397.420774743" Nov 22 08:19:43 crc kubenswrapper[4789]: I1122 08:19:43.980598 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="88a4d40a-bffd-416e-a813-a2dd3ddcf47c" path="/var/lib/kubelet/pods/88a4d40a-bffd-416e-a813-a2dd3ddcf47c/volumes" Nov 22 08:19:44 crc kubenswrapper[4789]: I1122 08:19:44.170878 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a1cea96e-2c65-4231-a4b4-bf6d197c5981","Type":"ContainerStarted","Data":"ee6a8e8a02386e188287275dafb781b96973eb7871dfcbf91fda73aca443d1e1"} Nov 22 08:19:44 crc kubenswrapper[4789]: I1122 08:19:44.191837 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.191817647 podStartE2EDuration="2.191817647s" podCreationTimestamp="2025-11-22 08:19:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:19:44.188296512 +0000 UTC m=+1398.422696795" watchObservedRunningTime="2025-11-22 08:19:44.191817647 +0000 UTC m=+1398.426217920" Nov 22 08:19:46 crc kubenswrapper[4789]: I1122 08:19:46.562921 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 22 08:19:46 crc kubenswrapper[4789]: I1122 08:19:46.563542 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 22 08:19:47 crc kubenswrapper[4789]: I1122 08:19:47.532874 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 22 08:19:51 crc kubenswrapper[4789]: I1122 08:19:51.563275 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 22 08:19:51 crc kubenswrapper[4789]: I1122 08:19:51.564128 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 22 08:19:51 crc kubenswrapper[4789]: I1122 08:19:51.576422 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 22 08:19:51 crc kubenswrapper[4789]: I1122 08:19:51.576512 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 22 08:19:52 crc kubenswrapper[4789]: I1122 08:19:52.526155 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 22 08:19:52 crc kubenswrapper[4789]: I1122 08:19:52.533400 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 22 08:19:52 crc kubenswrapper[4789]: I1122 08:19:52.572283 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 22 08:19:52 crc kubenswrapper[4789]: I1122 08:19:52.577969 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="588fc366-e2d8-40bc-b09e-ebcf032e440f" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.189:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 22 08:19:52 crc kubenswrapper[4789]: I1122 08:19:52.578036 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="588fc366-e2d8-40bc-b09e-ebcf032e440f" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.189:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 22 08:19:52 crc kubenswrapper[4789]: I1122 08:19:52.628027 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="9585ddf1-5799-43c3-a747-860ef1bf8380" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.190:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 22 08:19:52 crc kubenswrapper[4789]: I1122 08:19:52.628156 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="9585ddf1-5799-43c3-a747-860ef1bf8380" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.190:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 22 08:19:53 crc kubenswrapper[4789]: I1122 08:19:53.283307 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 22 08:20:01 crc kubenswrapper[4789]: I1122 08:20:01.570832 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 22 08:20:01 crc kubenswrapper[4789]: I1122 08:20:01.571573 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 22 08:20:01 crc kubenswrapper[4789]: I1122 08:20:01.577099 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 22 08:20:01 crc kubenswrapper[4789]: I1122 08:20:01.577516 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 22 08:20:01 crc kubenswrapper[4789]: I1122 08:20:01.586350 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 22 08:20:01 crc kubenswrapper[4789]: I1122 08:20:01.586830 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 22 08:20:01 crc kubenswrapper[4789]: I1122 08:20:01.587136 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 22 08:20:01 crc kubenswrapper[4789]: I1122 08:20:01.598000 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 22 08:20:02 crc kubenswrapper[4789]: I1122 08:20:02.323794 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 22 08:20:02 crc kubenswrapper[4789]: I1122 08:20:02.332217 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 22 08:20:11 crc kubenswrapper[4789]: I1122 08:20:11.019634 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 22 08:20:11 crc kubenswrapper[4789]: I1122 08:20:11.838569 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 22 08:20:14 crc kubenswrapper[4789]: I1122 08:20:14.871138 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="af27dae6-c617-4d16-bb53-19057df7689b" containerName="rabbitmq" containerID="cri-o://5d2637566554df256e507507e52737776868ccf6d6cf34ad3f5ce06caa06a1c8" gracePeriod=604797 Nov 22 08:20:15 crc kubenswrapper[4789]: I1122 08:20:15.226893 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="af27dae6-c617-4d16-bb53-19057df7689b" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.98:5671: connect: connection refused" Nov 22 08:20:15 crc kubenswrapper[4789]: I1122 08:20:15.536503 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="df1ccc5e-c20c-40ce-bb76-17da02c4bc8c" containerName="rabbitmq" containerID="cri-o://2b2f9cf2f918a4e8b984cfe694a19fd982ad7500f88d820af5e9d08ce511ffe4" gracePeriod=604797 Nov 22 08:20:16 crc kubenswrapper[4789]: I1122 08:20:16.933999 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-mbqjc"] Nov 22 08:20:16 crc kubenswrapper[4789]: I1122 08:20:16.936223 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mbqjc" Nov 22 08:20:16 crc kubenswrapper[4789]: I1122 08:20:16.965352 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mbqjc"] Nov 22 08:20:17 crc kubenswrapper[4789]: I1122 08:20:17.108840 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96aba4ce-5859-4803-ac23-919b4bad4f0d-utilities\") pod \"redhat-marketplace-mbqjc\" (UID: \"96aba4ce-5859-4803-ac23-919b4bad4f0d\") " pod="openshift-marketplace/redhat-marketplace-mbqjc" Nov 22 08:20:17 crc kubenswrapper[4789]: I1122 08:20:17.108965 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96aba4ce-5859-4803-ac23-919b4bad4f0d-catalog-content\") pod \"redhat-marketplace-mbqjc\" (UID: \"96aba4ce-5859-4803-ac23-919b4bad4f0d\") " pod="openshift-marketplace/redhat-marketplace-mbqjc" Nov 22 08:20:17 crc kubenswrapper[4789]: I1122 08:20:17.109119 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rtkwh\" (UniqueName: \"kubernetes.io/projected/96aba4ce-5859-4803-ac23-919b4bad4f0d-kube-api-access-rtkwh\") pod \"redhat-marketplace-mbqjc\" (UID: \"96aba4ce-5859-4803-ac23-919b4bad4f0d\") " pod="openshift-marketplace/redhat-marketplace-mbqjc" Nov 22 08:20:17 crc kubenswrapper[4789]: I1122 08:20:17.211004 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96aba4ce-5859-4803-ac23-919b4bad4f0d-utilities\") pod \"redhat-marketplace-mbqjc\" (UID: \"96aba4ce-5859-4803-ac23-919b4bad4f0d\") " pod="openshift-marketplace/redhat-marketplace-mbqjc" Nov 22 08:20:17 crc kubenswrapper[4789]: I1122 08:20:17.211409 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96aba4ce-5859-4803-ac23-919b4bad4f0d-catalog-content\") pod \"redhat-marketplace-mbqjc\" (UID: \"96aba4ce-5859-4803-ac23-919b4bad4f0d\") " pod="openshift-marketplace/redhat-marketplace-mbqjc" Nov 22 08:20:17 crc kubenswrapper[4789]: I1122 08:20:17.211532 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96aba4ce-5859-4803-ac23-919b4bad4f0d-utilities\") pod \"redhat-marketplace-mbqjc\" (UID: \"96aba4ce-5859-4803-ac23-919b4bad4f0d\") " pod="openshift-marketplace/redhat-marketplace-mbqjc" Nov 22 08:20:17 crc kubenswrapper[4789]: I1122 08:20:17.211785 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rtkwh\" (UniqueName: \"kubernetes.io/projected/96aba4ce-5859-4803-ac23-919b4bad4f0d-kube-api-access-rtkwh\") pod \"redhat-marketplace-mbqjc\" (UID: \"96aba4ce-5859-4803-ac23-919b4bad4f0d\") " pod="openshift-marketplace/redhat-marketplace-mbqjc" Nov 22 08:20:17 crc kubenswrapper[4789]: I1122 08:20:17.212151 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96aba4ce-5859-4803-ac23-919b4bad4f0d-catalog-content\") pod \"redhat-marketplace-mbqjc\" (UID: \"96aba4ce-5859-4803-ac23-919b4bad4f0d\") " pod="openshift-marketplace/redhat-marketplace-mbqjc" Nov 22 08:20:17 crc kubenswrapper[4789]: I1122 08:20:17.232981 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rtkwh\" (UniqueName: \"kubernetes.io/projected/96aba4ce-5859-4803-ac23-919b4bad4f0d-kube-api-access-rtkwh\") pod \"redhat-marketplace-mbqjc\" (UID: \"96aba4ce-5859-4803-ac23-919b4bad4f0d\") " pod="openshift-marketplace/redhat-marketplace-mbqjc" Nov 22 08:20:17 crc kubenswrapper[4789]: I1122 08:20:17.262540 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mbqjc" Nov 22 08:20:17 crc kubenswrapper[4789]: I1122 08:20:17.739063 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mbqjc"] Nov 22 08:20:18 crc kubenswrapper[4789]: I1122 08:20:18.462171 4789 generic.go:334] "Generic (PLEG): container finished" podID="96aba4ce-5859-4803-ac23-919b4bad4f0d" containerID="1bf14414f405a586fa8ace6440f7fa4015e58bf9c2db6ad8059273ef684bc82c" exitCode=0 Nov 22 08:20:18 crc kubenswrapper[4789]: I1122 08:20:18.462259 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mbqjc" event={"ID":"96aba4ce-5859-4803-ac23-919b4bad4f0d","Type":"ContainerDied","Data":"1bf14414f405a586fa8ace6440f7fa4015e58bf9c2db6ad8059273ef684bc82c"} Nov 22 08:20:18 crc kubenswrapper[4789]: I1122 08:20:18.462453 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mbqjc" event={"ID":"96aba4ce-5859-4803-ac23-919b4bad4f0d","Type":"ContainerStarted","Data":"6c750f1bf3db12c2532e4fbc74cfdc0e869419c14ab1b9683f46b236ffba6720"} Nov 22 08:20:19 crc kubenswrapper[4789]: I1122 08:20:19.473162 4789 generic.go:334] "Generic (PLEG): container finished" podID="96aba4ce-5859-4803-ac23-919b4bad4f0d" containerID="c6dec650ed5319f6d68f7fadc96076e9baf648b9cfa50a9fb97312709a523024" exitCode=0 Nov 22 08:20:19 crc kubenswrapper[4789]: I1122 08:20:19.473231 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mbqjc" event={"ID":"96aba4ce-5859-4803-ac23-919b4bad4f0d","Type":"ContainerDied","Data":"c6dec650ed5319f6d68f7fadc96076e9baf648b9cfa50a9fb97312709a523024"} Nov 22 08:20:20 crc kubenswrapper[4789]: I1122 08:20:20.485011 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mbqjc" event={"ID":"96aba4ce-5859-4803-ac23-919b4bad4f0d","Type":"ContainerStarted","Data":"4d04a10903c50e595852e4fc8585c46a927df8a685c440af81660580bdb14527"} Nov 22 08:20:20 crc kubenswrapper[4789]: I1122 08:20:20.521140 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-mbqjc" podStartSLOduration=2.8091190729999997 podStartE2EDuration="4.521118844s" podCreationTimestamp="2025-11-22 08:20:16 +0000 UTC" firstStartedPulling="2025-11-22 08:20:18.468243665 +0000 UTC m=+1432.702643938" lastFinishedPulling="2025-11-22 08:20:20.180243436 +0000 UTC m=+1434.414643709" observedRunningTime="2025-11-22 08:20:20.511090921 +0000 UTC m=+1434.745491204" watchObservedRunningTime="2025-11-22 08:20:20.521118844 +0000 UTC m=+1434.755519117" Nov 22 08:20:20 crc kubenswrapper[4789]: I1122 08:20:20.922148 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-ccxrn"] Nov 22 08:20:20 crc kubenswrapper[4789]: I1122 08:20:20.924199 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ccxrn" Nov 22 08:20:20 crc kubenswrapper[4789]: I1122 08:20:20.939251 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ccxrn"] Nov 22 08:20:21 crc kubenswrapper[4789]: I1122 08:20:21.086503 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70a0fabb-bd28-4991-87b7-95236f639ddb-catalog-content\") pod \"redhat-operators-ccxrn\" (UID: \"70a0fabb-bd28-4991-87b7-95236f639ddb\") " pod="openshift-marketplace/redhat-operators-ccxrn" Nov 22 08:20:21 crc kubenswrapper[4789]: I1122 08:20:21.086585 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70a0fabb-bd28-4991-87b7-95236f639ddb-utilities\") pod \"redhat-operators-ccxrn\" (UID: \"70a0fabb-bd28-4991-87b7-95236f639ddb\") " pod="openshift-marketplace/redhat-operators-ccxrn" Nov 22 08:20:21 crc kubenswrapper[4789]: I1122 08:20:21.086912 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2h4tv\" (UniqueName: \"kubernetes.io/projected/70a0fabb-bd28-4991-87b7-95236f639ddb-kube-api-access-2h4tv\") pod \"redhat-operators-ccxrn\" (UID: \"70a0fabb-bd28-4991-87b7-95236f639ddb\") " pod="openshift-marketplace/redhat-operators-ccxrn" Nov 22 08:20:21 crc kubenswrapper[4789]: I1122 08:20:21.188913 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2h4tv\" (UniqueName: \"kubernetes.io/projected/70a0fabb-bd28-4991-87b7-95236f639ddb-kube-api-access-2h4tv\") pod \"redhat-operators-ccxrn\" (UID: \"70a0fabb-bd28-4991-87b7-95236f639ddb\") " pod="openshift-marketplace/redhat-operators-ccxrn" Nov 22 08:20:21 crc kubenswrapper[4789]: I1122 08:20:21.189021 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70a0fabb-bd28-4991-87b7-95236f639ddb-catalog-content\") pod \"redhat-operators-ccxrn\" (UID: \"70a0fabb-bd28-4991-87b7-95236f639ddb\") " pod="openshift-marketplace/redhat-operators-ccxrn" Nov 22 08:20:21 crc kubenswrapper[4789]: I1122 08:20:21.189048 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70a0fabb-bd28-4991-87b7-95236f639ddb-utilities\") pod \"redhat-operators-ccxrn\" (UID: \"70a0fabb-bd28-4991-87b7-95236f639ddb\") " pod="openshift-marketplace/redhat-operators-ccxrn" Nov 22 08:20:21 crc kubenswrapper[4789]: I1122 08:20:21.189607 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70a0fabb-bd28-4991-87b7-95236f639ddb-utilities\") pod \"redhat-operators-ccxrn\" (UID: \"70a0fabb-bd28-4991-87b7-95236f639ddb\") " pod="openshift-marketplace/redhat-operators-ccxrn" Nov 22 08:20:21 crc kubenswrapper[4789]: I1122 08:20:21.189678 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70a0fabb-bd28-4991-87b7-95236f639ddb-catalog-content\") pod \"redhat-operators-ccxrn\" (UID: \"70a0fabb-bd28-4991-87b7-95236f639ddb\") " pod="openshift-marketplace/redhat-operators-ccxrn" Nov 22 08:20:21 crc kubenswrapper[4789]: I1122 08:20:21.208701 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2h4tv\" (UniqueName: \"kubernetes.io/projected/70a0fabb-bd28-4991-87b7-95236f639ddb-kube-api-access-2h4tv\") pod \"redhat-operators-ccxrn\" (UID: \"70a0fabb-bd28-4991-87b7-95236f639ddb\") " pod="openshift-marketplace/redhat-operators-ccxrn" Nov 22 08:20:21 crc kubenswrapper[4789]: I1122 08:20:21.249552 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ccxrn" Nov 22 08:20:21 crc kubenswrapper[4789]: I1122 08:20:21.503389 4789 generic.go:334] "Generic (PLEG): container finished" podID="af27dae6-c617-4d16-bb53-19057df7689b" containerID="5d2637566554df256e507507e52737776868ccf6d6cf34ad3f5ce06caa06a1c8" exitCode=0 Nov 22 08:20:21 crc kubenswrapper[4789]: I1122 08:20:21.503453 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"af27dae6-c617-4d16-bb53-19057df7689b","Type":"ContainerDied","Data":"5d2637566554df256e507507e52737776868ccf6d6cf34ad3f5ce06caa06a1c8"} Nov 22 08:20:21 crc kubenswrapper[4789]: I1122 08:20:21.706374 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ccxrn"] Nov 22 08:20:21 crc kubenswrapper[4789]: W1122 08:20:21.710857 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod70a0fabb_bd28_4991_87b7_95236f639ddb.slice/crio-b48a4dc6ea60edc3412a6182e0aa2e10a90a951f1dfd17ad46a8d8e439963f61 WatchSource:0}: Error finding container b48a4dc6ea60edc3412a6182e0aa2e10a90a951f1dfd17ad46a8d8e439963f61: Status 404 returned error can't find the container with id b48a4dc6ea60edc3412a6182e0aa2e10a90a951f1dfd17ad46a8d8e439963f61 Nov 22 08:20:21 crc kubenswrapper[4789]: I1122 08:20:21.939050 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.107351 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/af27dae6-c617-4d16-bb53-19057df7689b-rabbitmq-tls\") pod \"af27dae6-c617-4d16-bb53-19057df7689b\" (UID: \"af27dae6-c617-4d16-bb53-19057df7689b\") " Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.107634 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/af27dae6-c617-4d16-bb53-19057df7689b-plugins-conf\") pod \"af27dae6-c617-4d16-bb53-19057df7689b\" (UID: \"af27dae6-c617-4d16-bb53-19057df7689b\") " Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.107790 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"af27dae6-c617-4d16-bb53-19057df7689b\" (UID: \"af27dae6-c617-4d16-bb53-19057df7689b\") " Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.108905 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/af27dae6-c617-4d16-bb53-19057df7689b-server-conf\") pod \"af27dae6-c617-4d16-bb53-19057df7689b\" (UID: \"af27dae6-c617-4d16-bb53-19057df7689b\") " Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.109088 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/af27dae6-c617-4d16-bb53-19057df7689b-pod-info\") pod \"af27dae6-c617-4d16-bb53-19057df7689b\" (UID: \"af27dae6-c617-4d16-bb53-19057df7689b\") " Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.109228 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/af27dae6-c617-4d16-bb53-19057df7689b-rabbitmq-confd\") pod \"af27dae6-c617-4d16-bb53-19057df7689b\" (UID: \"af27dae6-c617-4d16-bb53-19057df7689b\") " Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.109466 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/af27dae6-c617-4d16-bb53-19057df7689b-rabbitmq-erlang-cookie\") pod \"af27dae6-c617-4d16-bb53-19057df7689b\" (UID: \"af27dae6-c617-4d16-bb53-19057df7689b\") " Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.109649 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j7wtx\" (UniqueName: \"kubernetes.io/projected/af27dae6-c617-4d16-bb53-19057df7689b-kube-api-access-j7wtx\") pod \"af27dae6-c617-4d16-bb53-19057df7689b\" (UID: \"af27dae6-c617-4d16-bb53-19057df7689b\") " Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.110148 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/af27dae6-c617-4d16-bb53-19057df7689b-rabbitmq-plugins\") pod \"af27dae6-c617-4d16-bb53-19057df7689b\" (UID: \"af27dae6-c617-4d16-bb53-19057df7689b\") " Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.110277 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/af27dae6-c617-4d16-bb53-19057df7689b-erlang-cookie-secret\") pod \"af27dae6-c617-4d16-bb53-19057df7689b\" (UID: \"af27dae6-c617-4d16-bb53-19057df7689b\") " Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.110416 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/af27dae6-c617-4d16-bb53-19057df7689b-config-data\") pod \"af27dae6-c617-4d16-bb53-19057df7689b\" (UID: \"af27dae6-c617-4d16-bb53-19057df7689b\") " Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.111882 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af27dae6-c617-4d16-bb53-19057df7689b-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "af27dae6-c617-4d16-bb53-19057df7689b" (UID: "af27dae6-c617-4d16-bb53-19057df7689b"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.121370 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af27dae6-c617-4d16-bb53-19057df7689b-kube-api-access-j7wtx" (OuterVolumeSpecName: "kube-api-access-j7wtx") pod "af27dae6-c617-4d16-bb53-19057df7689b" (UID: "af27dae6-c617-4d16-bb53-19057df7689b"). InnerVolumeSpecName "kube-api-access-j7wtx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.128943 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/af27dae6-c617-4d16-bb53-19057df7689b-pod-info" (OuterVolumeSpecName: "pod-info") pod "af27dae6-c617-4d16-bb53-19057df7689b" (UID: "af27dae6-c617-4d16-bb53-19057df7689b"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.128950 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af27dae6-c617-4d16-bb53-19057df7689b-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "af27dae6-c617-4d16-bb53-19057df7689b" (UID: "af27dae6-c617-4d16-bb53-19057df7689b"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.131170 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/af27dae6-c617-4d16-bb53-19057df7689b-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "af27dae6-c617-4d16-bb53-19057df7689b" (UID: "af27dae6-c617-4d16-bb53-19057df7689b"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.131579 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af27dae6-c617-4d16-bb53-19057df7689b-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "af27dae6-c617-4d16-bb53-19057df7689b" (UID: "af27dae6-c617-4d16-bb53-19057df7689b"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.142945 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "persistence") pod "af27dae6-c617-4d16-bb53-19057df7689b" (UID: "af27dae6-c617-4d16-bb53-19057df7689b"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.153275 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/af27dae6-c617-4d16-bb53-19057df7689b-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "af27dae6-c617-4d16-bb53-19057df7689b" (UID: "af27dae6-c617-4d16-bb53-19057df7689b"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.203972 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af27dae6-c617-4d16-bb53-19057df7689b-config-data" (OuterVolumeSpecName: "config-data") pod "af27dae6-c617-4d16-bb53-19057df7689b" (UID: "af27dae6-c617-4d16-bb53-19057df7689b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.213333 4789 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/af27dae6-c617-4d16-bb53-19057df7689b-plugins-conf\") on node \"crc\" DevicePath \"\"" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.213408 4789 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.213421 4789 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/af27dae6-c617-4d16-bb53-19057df7689b-pod-info\") on node \"crc\" DevicePath \"\"" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.213430 4789 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/af27dae6-c617-4d16-bb53-19057df7689b-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.213441 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j7wtx\" (UniqueName: \"kubernetes.io/projected/af27dae6-c617-4d16-bb53-19057df7689b-kube-api-access-j7wtx\") on node \"crc\" DevicePath \"\"" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.213450 4789 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/af27dae6-c617-4d16-bb53-19057df7689b-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.213457 4789 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/af27dae6-c617-4d16-bb53-19057df7689b-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.213481 4789 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/af27dae6-c617-4d16-bb53-19057df7689b-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.213489 4789 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/af27dae6-c617-4d16-bb53-19057df7689b-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.236507 4789 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.251366 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af27dae6-c617-4d16-bb53-19057df7689b-server-conf" (OuterVolumeSpecName: "server-conf") pod "af27dae6-c617-4d16-bb53-19057df7689b" (UID: "af27dae6-c617-4d16-bb53-19057df7689b"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.314737 4789 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.314791 4789 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/af27dae6-c617-4d16-bb53-19057df7689b-server-conf\") on node \"crc\" DevicePath \"\"" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.330311 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af27dae6-c617-4d16-bb53-19057df7689b-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "af27dae6-c617-4d16-bb53-19057df7689b" (UID: "af27dae6-c617-4d16-bb53-19057df7689b"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.411686 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.417019 4789 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/af27dae6-c617-4d16-bb53-19057df7689b-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.514339 4789 generic.go:334] "Generic (PLEG): container finished" podID="df1ccc5e-c20c-40ce-bb76-17da02c4bc8c" containerID="2b2f9cf2f918a4e8b984cfe694a19fd982ad7500f88d820af5e9d08ce511ffe4" exitCode=0 Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.514408 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.514435 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"df1ccc5e-c20c-40ce-bb76-17da02c4bc8c","Type":"ContainerDied","Data":"2b2f9cf2f918a4e8b984cfe694a19fd982ad7500f88d820af5e9d08ce511ffe4"} Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.514482 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"df1ccc5e-c20c-40ce-bb76-17da02c4bc8c","Type":"ContainerDied","Data":"9c40add8b63df1edbbdcaf738202f8031af6061b2dda83b3afea044739a87050"} Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.514501 4789 scope.go:117] "RemoveContainer" containerID="2b2f9cf2f918a4e8b984cfe694a19fd982ad7500f88d820af5e9d08ce511ffe4" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.518093 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/df1ccc5e-c20c-40ce-bb76-17da02c4bc8c-erlang-cookie-secret\") pod \"df1ccc5e-c20c-40ce-bb76-17da02c4bc8c\" (UID: \"df1ccc5e-c20c-40ce-bb76-17da02c4bc8c\") " Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.518143 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/df1ccc5e-c20c-40ce-bb76-17da02c4bc8c-pod-info\") pod \"df1ccc5e-c20c-40ce-bb76-17da02c4bc8c\" (UID: \"df1ccc5e-c20c-40ce-bb76-17da02c4bc8c\") " Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.518214 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fv6wj\" (UniqueName: \"kubernetes.io/projected/df1ccc5e-c20c-40ce-bb76-17da02c4bc8c-kube-api-access-fv6wj\") pod \"df1ccc5e-c20c-40ce-bb76-17da02c4bc8c\" (UID: \"df1ccc5e-c20c-40ce-bb76-17da02c4bc8c\") " Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.518243 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/df1ccc5e-c20c-40ce-bb76-17da02c4bc8c-config-data\") pod \"df1ccc5e-c20c-40ce-bb76-17da02c4bc8c\" (UID: \"df1ccc5e-c20c-40ce-bb76-17da02c4bc8c\") " Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.518262 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"df1ccc5e-c20c-40ce-bb76-17da02c4bc8c\" (UID: \"df1ccc5e-c20c-40ce-bb76-17da02c4bc8c\") " Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.518313 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/df1ccc5e-c20c-40ce-bb76-17da02c4bc8c-server-conf\") pod \"df1ccc5e-c20c-40ce-bb76-17da02c4bc8c\" (UID: \"df1ccc5e-c20c-40ce-bb76-17da02c4bc8c\") " Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.518360 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/df1ccc5e-c20c-40ce-bb76-17da02c4bc8c-plugins-conf\") pod \"df1ccc5e-c20c-40ce-bb76-17da02c4bc8c\" (UID: \"df1ccc5e-c20c-40ce-bb76-17da02c4bc8c\") " Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.518394 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/df1ccc5e-c20c-40ce-bb76-17da02c4bc8c-rabbitmq-tls\") pod \"df1ccc5e-c20c-40ce-bb76-17da02c4bc8c\" (UID: \"df1ccc5e-c20c-40ce-bb76-17da02c4bc8c\") " Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.518443 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/df1ccc5e-c20c-40ce-bb76-17da02c4bc8c-rabbitmq-plugins\") pod \"df1ccc5e-c20c-40ce-bb76-17da02c4bc8c\" (UID: \"df1ccc5e-c20c-40ce-bb76-17da02c4bc8c\") " Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.518490 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/df1ccc5e-c20c-40ce-bb76-17da02c4bc8c-rabbitmq-confd\") pod \"df1ccc5e-c20c-40ce-bb76-17da02c4bc8c\" (UID: \"df1ccc5e-c20c-40ce-bb76-17da02c4bc8c\") " Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.518521 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/df1ccc5e-c20c-40ce-bb76-17da02c4bc8c-rabbitmq-erlang-cookie\") pod \"df1ccc5e-c20c-40ce-bb76-17da02c4bc8c\" (UID: \"df1ccc5e-c20c-40ce-bb76-17da02c4bc8c\") " Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.519386 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.519380 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"af27dae6-c617-4d16-bb53-19057df7689b","Type":"ContainerDied","Data":"cc56f8ffb597c855574d458ec8daf6759f5b52074439eea42be55aa3d3b2f5e1"} Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.526189 4789 generic.go:334] "Generic (PLEG): container finished" podID="70a0fabb-bd28-4991-87b7-95236f639ddb" containerID="a5976b507eba98c7547d2fd79706d4b04159ef43f9be684cc22656f65564058a" exitCode=0 Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.526231 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ccxrn" event={"ID":"70a0fabb-bd28-4991-87b7-95236f639ddb","Type":"ContainerDied","Data":"a5976b507eba98c7547d2fd79706d4b04159ef43f9be684cc22656f65564058a"} Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.526253 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ccxrn" event={"ID":"70a0fabb-bd28-4991-87b7-95236f639ddb","Type":"ContainerStarted","Data":"b48a4dc6ea60edc3412a6182e0aa2e10a90a951f1dfd17ad46a8d8e439963f61"} Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.533137 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df1ccc5e-c20c-40ce-bb76-17da02c4bc8c-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "df1ccc5e-c20c-40ce-bb76-17da02c4bc8c" (UID: "df1ccc5e-c20c-40ce-bb76-17da02c4bc8c"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.533178 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "persistence") pod "df1ccc5e-c20c-40ce-bb76-17da02c4bc8c" (UID: "df1ccc5e-c20c-40ce-bb76-17da02c4bc8c"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.534023 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/df1ccc5e-c20c-40ce-bb76-17da02c4bc8c-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "df1ccc5e-c20c-40ce-bb76-17da02c4bc8c" (UID: "df1ccc5e-c20c-40ce-bb76-17da02c4bc8c"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.537874 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/df1ccc5e-c20c-40ce-bb76-17da02c4bc8c-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "df1ccc5e-c20c-40ce-bb76-17da02c4bc8c" (UID: "df1ccc5e-c20c-40ce-bb76-17da02c4bc8c"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.542863 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df1ccc5e-c20c-40ce-bb76-17da02c4bc8c-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "df1ccc5e-c20c-40ce-bb76-17da02c4bc8c" (UID: "df1ccc5e-c20c-40ce-bb76-17da02c4bc8c"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.543204 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/df1ccc5e-c20c-40ce-bb76-17da02c4bc8c-pod-info" (OuterVolumeSpecName: "pod-info") pod "df1ccc5e-c20c-40ce-bb76-17da02c4bc8c" (UID: "df1ccc5e-c20c-40ce-bb76-17da02c4bc8c"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.544896 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df1ccc5e-c20c-40ce-bb76-17da02c4bc8c-kube-api-access-fv6wj" (OuterVolumeSpecName: "kube-api-access-fv6wj") pod "df1ccc5e-c20c-40ce-bb76-17da02c4bc8c" (UID: "df1ccc5e-c20c-40ce-bb76-17da02c4bc8c"). InnerVolumeSpecName "kube-api-access-fv6wj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.555586 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df1ccc5e-c20c-40ce-bb76-17da02c4bc8c-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "df1ccc5e-c20c-40ce-bb76-17da02c4bc8c" (UID: "df1ccc5e-c20c-40ce-bb76-17da02c4bc8c"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.580239 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df1ccc5e-c20c-40ce-bb76-17da02c4bc8c-config-data" (OuterVolumeSpecName: "config-data") pod "df1ccc5e-c20c-40ce-bb76-17da02c4bc8c" (UID: "df1ccc5e-c20c-40ce-bb76-17da02c4bc8c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.623177 4789 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/df1ccc5e-c20c-40ce-bb76-17da02c4bc8c-plugins-conf\") on node \"crc\" DevicePath \"\"" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.623214 4789 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/df1ccc5e-c20c-40ce-bb76-17da02c4bc8c-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.623227 4789 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/df1ccc5e-c20c-40ce-bb76-17da02c4bc8c-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.623242 4789 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/df1ccc5e-c20c-40ce-bb76-17da02c4bc8c-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.623255 4789 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/df1ccc5e-c20c-40ce-bb76-17da02c4bc8c-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.623266 4789 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/df1ccc5e-c20c-40ce-bb76-17da02c4bc8c-pod-info\") on node \"crc\" DevicePath \"\"" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.623279 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fv6wj\" (UniqueName: \"kubernetes.io/projected/df1ccc5e-c20c-40ce-bb76-17da02c4bc8c-kube-api-access-fv6wj\") on node \"crc\" DevicePath \"\"" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.623291 4789 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/df1ccc5e-c20c-40ce-bb76-17da02c4bc8c-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.623314 4789 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.631538 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df1ccc5e-c20c-40ce-bb76-17da02c4bc8c-server-conf" (OuterVolumeSpecName: "server-conf") pod "df1ccc5e-c20c-40ce-bb76-17da02c4bc8c" (UID: "df1ccc5e-c20c-40ce-bb76-17da02c4bc8c"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.659288 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.664380 4789 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.664881 4789 scope.go:117] "RemoveContainer" containerID="b52c35865cafb51202b9dc7fd7ca36b2d1afa82daff2e035973c3e4405cd63b2" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.666213 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.696117 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df1ccc5e-c20c-40ce-bb76-17da02c4bc8c-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "df1ccc5e-c20c-40ce-bb76-17da02c4bc8c" (UID: "df1ccc5e-c20c-40ce-bb76-17da02c4bc8c"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.732061 4789 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.732288 4789 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/df1ccc5e-c20c-40ce-bb76-17da02c4bc8c-server-conf\") on node \"crc\" DevicePath \"\"" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.732303 4789 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/df1ccc5e-c20c-40ce-bb76-17da02c4bc8c-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.733582 4789 scope.go:117] "RemoveContainer" containerID="2b2f9cf2f918a4e8b984cfe694a19fd982ad7500f88d820af5e9d08ce511ffe4" Nov 22 08:20:22 crc kubenswrapper[4789]: E1122 08:20:22.735948 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b2f9cf2f918a4e8b984cfe694a19fd982ad7500f88d820af5e9d08ce511ffe4\": container with ID starting with 2b2f9cf2f918a4e8b984cfe694a19fd982ad7500f88d820af5e9d08ce511ffe4 not found: ID does not exist" containerID="2b2f9cf2f918a4e8b984cfe694a19fd982ad7500f88d820af5e9d08ce511ffe4" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.735989 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b2f9cf2f918a4e8b984cfe694a19fd982ad7500f88d820af5e9d08ce511ffe4"} err="failed to get container status \"2b2f9cf2f918a4e8b984cfe694a19fd982ad7500f88d820af5e9d08ce511ffe4\": rpc error: code = NotFound desc = could not find container \"2b2f9cf2f918a4e8b984cfe694a19fd982ad7500f88d820af5e9d08ce511ffe4\": container with ID starting with 2b2f9cf2f918a4e8b984cfe694a19fd982ad7500f88d820af5e9d08ce511ffe4 not found: ID does not exist" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.736016 4789 scope.go:117] "RemoveContainer" containerID="b52c35865cafb51202b9dc7fd7ca36b2d1afa82daff2e035973c3e4405cd63b2" Nov 22 08:20:22 crc kubenswrapper[4789]: E1122 08:20:22.736692 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b52c35865cafb51202b9dc7fd7ca36b2d1afa82daff2e035973c3e4405cd63b2\": container with ID starting with b52c35865cafb51202b9dc7fd7ca36b2d1afa82daff2e035973c3e4405cd63b2 not found: ID does not exist" containerID="b52c35865cafb51202b9dc7fd7ca36b2d1afa82daff2e035973c3e4405cd63b2" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.736719 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b52c35865cafb51202b9dc7fd7ca36b2d1afa82daff2e035973c3e4405cd63b2"} err="failed to get container status \"b52c35865cafb51202b9dc7fd7ca36b2d1afa82daff2e035973c3e4405cd63b2\": rpc error: code = NotFound desc = could not find container \"b52c35865cafb51202b9dc7fd7ca36b2d1afa82daff2e035973c3e4405cd63b2\": container with ID starting with b52c35865cafb51202b9dc7fd7ca36b2d1afa82daff2e035973c3e4405cd63b2 not found: ID does not exist" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.736733 4789 scope.go:117] "RemoveContainer" containerID="5d2637566554df256e507507e52737776868ccf6d6cf34ad3f5ce06caa06a1c8" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.739887 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Nov 22 08:20:22 crc kubenswrapper[4789]: E1122 08:20:22.740331 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df1ccc5e-c20c-40ce-bb76-17da02c4bc8c" containerName="rabbitmq" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.740350 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="df1ccc5e-c20c-40ce-bb76-17da02c4bc8c" containerName="rabbitmq" Nov 22 08:20:22 crc kubenswrapper[4789]: E1122 08:20:22.740379 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af27dae6-c617-4d16-bb53-19057df7689b" containerName="setup-container" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.740387 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="af27dae6-c617-4d16-bb53-19057df7689b" containerName="setup-container" Nov 22 08:20:22 crc kubenswrapper[4789]: E1122 08:20:22.740406 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df1ccc5e-c20c-40ce-bb76-17da02c4bc8c" containerName="setup-container" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.740414 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="df1ccc5e-c20c-40ce-bb76-17da02c4bc8c" containerName="setup-container" Nov 22 08:20:22 crc kubenswrapper[4789]: E1122 08:20:22.740427 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af27dae6-c617-4d16-bb53-19057df7689b" containerName="rabbitmq" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.740435 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="af27dae6-c617-4d16-bb53-19057df7689b" containerName="rabbitmq" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.740620 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="af27dae6-c617-4d16-bb53-19057df7689b" containerName="rabbitmq" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.740641 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="df1ccc5e-c20c-40ce-bb76-17da02c4bc8c" containerName="rabbitmq" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.741808 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.746646 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.746932 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-ggnlw" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.746959 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.747058 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.747153 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.747244 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.747267 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.753732 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.812528 4789 scope.go:117] "RemoveContainer" containerID="e7ed5333a86c250ae63dba829dea4c481af2d77e17c98c355f0fc1115c16d37c" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.841970 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1f834b47-b9a1-4813-bcac-3b5161eceeac-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"1f834b47-b9a1-4813-bcac-3b5161eceeac\") " pod="openstack/rabbitmq-server-0" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.842020 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1f834b47-b9a1-4813-bcac-3b5161eceeac-pod-info\") pod \"rabbitmq-server-0\" (UID: \"1f834b47-b9a1-4813-bcac-3b5161eceeac\") " pod="openstack/rabbitmq-server-0" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.842054 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-server-0\" (UID: \"1f834b47-b9a1-4813-bcac-3b5161eceeac\") " pod="openstack/rabbitmq-server-0" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.842097 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q4ksd\" (UniqueName: \"kubernetes.io/projected/1f834b47-b9a1-4813-bcac-3b5161eceeac-kube-api-access-q4ksd\") pod \"rabbitmq-server-0\" (UID: \"1f834b47-b9a1-4813-bcac-3b5161eceeac\") " pod="openstack/rabbitmq-server-0" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.842119 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1f834b47-b9a1-4813-bcac-3b5161eceeac-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"1f834b47-b9a1-4813-bcac-3b5161eceeac\") " pod="openstack/rabbitmq-server-0" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.842161 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1f834b47-b9a1-4813-bcac-3b5161eceeac-config-data\") pod \"rabbitmq-server-0\" (UID: \"1f834b47-b9a1-4813-bcac-3b5161eceeac\") " pod="openstack/rabbitmq-server-0" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.842197 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1f834b47-b9a1-4813-bcac-3b5161eceeac-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"1f834b47-b9a1-4813-bcac-3b5161eceeac\") " pod="openstack/rabbitmq-server-0" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.842226 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1f834b47-b9a1-4813-bcac-3b5161eceeac-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"1f834b47-b9a1-4813-bcac-3b5161eceeac\") " pod="openstack/rabbitmq-server-0" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.842259 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1f834b47-b9a1-4813-bcac-3b5161eceeac-server-conf\") pod \"rabbitmq-server-0\" (UID: \"1f834b47-b9a1-4813-bcac-3b5161eceeac\") " pod="openstack/rabbitmq-server-0" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.842290 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1f834b47-b9a1-4813-bcac-3b5161eceeac-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"1f834b47-b9a1-4813-bcac-3b5161eceeac\") " pod="openstack/rabbitmq-server-0" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.842380 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1f834b47-b9a1-4813-bcac-3b5161eceeac-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"1f834b47-b9a1-4813-bcac-3b5161eceeac\") " pod="openstack/rabbitmq-server-0" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.900869 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.923145 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.945674 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1f834b47-b9a1-4813-bcac-3b5161eceeac-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"1f834b47-b9a1-4813-bcac-3b5161eceeac\") " pod="openstack/rabbitmq-server-0" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.945777 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1f834b47-b9a1-4813-bcac-3b5161eceeac-pod-info\") pod \"rabbitmq-server-0\" (UID: \"1f834b47-b9a1-4813-bcac-3b5161eceeac\") " pod="openstack/rabbitmq-server-0" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.945849 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-server-0\" (UID: \"1f834b47-b9a1-4813-bcac-3b5161eceeac\") " pod="openstack/rabbitmq-server-0" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.946259 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1f834b47-b9a1-4813-bcac-3b5161eceeac-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"1f834b47-b9a1-4813-bcac-3b5161eceeac\") " pod="openstack/rabbitmq-server-0" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.946373 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q4ksd\" (UniqueName: \"kubernetes.io/projected/1f834b47-b9a1-4813-bcac-3b5161eceeac-kube-api-access-q4ksd\") pod \"rabbitmq-server-0\" (UID: \"1f834b47-b9a1-4813-bcac-3b5161eceeac\") " pod="openstack/rabbitmq-server-0" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.946410 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1f834b47-b9a1-4813-bcac-3b5161eceeac-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"1f834b47-b9a1-4813-bcac-3b5161eceeac\") " pod="openstack/rabbitmq-server-0" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.946445 4789 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-server-0\" (UID: \"1f834b47-b9a1-4813-bcac-3b5161eceeac\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/rabbitmq-server-0" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.946993 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1f834b47-b9a1-4813-bcac-3b5161eceeac-config-data\") pod \"rabbitmq-server-0\" (UID: \"1f834b47-b9a1-4813-bcac-3b5161eceeac\") " pod="openstack/rabbitmq-server-0" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.947045 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1f834b47-b9a1-4813-bcac-3b5161eceeac-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"1f834b47-b9a1-4813-bcac-3b5161eceeac\") " pod="openstack/rabbitmq-server-0" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.947074 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1f834b47-b9a1-4813-bcac-3b5161eceeac-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"1f834b47-b9a1-4813-bcac-3b5161eceeac\") " pod="openstack/rabbitmq-server-0" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.947101 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1f834b47-b9a1-4813-bcac-3b5161eceeac-server-conf\") pod \"rabbitmq-server-0\" (UID: \"1f834b47-b9a1-4813-bcac-3b5161eceeac\") " pod="openstack/rabbitmq-server-0" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.947132 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1f834b47-b9a1-4813-bcac-3b5161eceeac-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"1f834b47-b9a1-4813-bcac-3b5161eceeac\") " pod="openstack/rabbitmq-server-0" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.947199 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1f834b47-b9a1-4813-bcac-3b5161eceeac-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"1f834b47-b9a1-4813-bcac-3b5161eceeac\") " pod="openstack/rabbitmq-server-0" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.947481 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1f834b47-b9a1-4813-bcac-3b5161eceeac-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"1f834b47-b9a1-4813-bcac-3b5161eceeac\") " pod="openstack/rabbitmq-server-0" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.948121 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1f834b47-b9a1-4813-bcac-3b5161eceeac-config-data\") pod \"rabbitmq-server-0\" (UID: \"1f834b47-b9a1-4813-bcac-3b5161eceeac\") " pod="openstack/rabbitmq-server-0" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.950058 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1f834b47-b9a1-4813-bcac-3b5161eceeac-pod-info\") pod \"rabbitmq-server-0\" (UID: \"1f834b47-b9a1-4813-bcac-3b5161eceeac\") " pod="openstack/rabbitmq-server-0" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.950796 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1f834b47-b9a1-4813-bcac-3b5161eceeac-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"1f834b47-b9a1-4813-bcac-3b5161eceeac\") " pod="openstack/rabbitmq-server-0" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.951011 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1f834b47-b9a1-4813-bcac-3b5161eceeac-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"1f834b47-b9a1-4813-bcac-3b5161eceeac\") " pod="openstack/rabbitmq-server-0" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.951646 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1f834b47-b9a1-4813-bcac-3b5161eceeac-server-conf\") pod \"rabbitmq-server-0\" (UID: \"1f834b47-b9a1-4813-bcac-3b5161eceeac\") " pod="openstack/rabbitmq-server-0" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.952577 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1f834b47-b9a1-4813-bcac-3b5161eceeac-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"1f834b47-b9a1-4813-bcac-3b5161eceeac\") " pod="openstack/rabbitmq-server-0" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.959141 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1f834b47-b9a1-4813-bcac-3b5161eceeac-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"1f834b47-b9a1-4813-bcac-3b5161eceeac\") " pod="openstack/rabbitmq-server-0" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.959231 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.971768 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q4ksd\" (UniqueName: \"kubernetes.io/projected/1f834b47-b9a1-4813-bcac-3b5161eceeac-kube-api-access-q4ksd\") pod \"rabbitmq-server-0\" (UID: \"1f834b47-b9a1-4813-bcac-3b5161eceeac\") " pod="openstack/rabbitmq-server-0" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.978422 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.978454 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.981601 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.981668 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.981981 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-4hvjr" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.982146 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.982329 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.982475 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.982500 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Nov 22 08:20:22 crc kubenswrapper[4789]: I1122 08:20:22.996348 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-server-0\" (UID: \"1f834b47-b9a1-4813-bcac-3b5161eceeac\") " pod="openstack/rabbitmq-server-0" Nov 22 08:20:23 crc kubenswrapper[4789]: I1122 08:20:23.048599 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/7ab63a4d-2e35-4b93-803b-13a38a43b8d6-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"7ab63a4d-2e35-4b93-803b-13a38a43b8d6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:20:23 crc kubenswrapper[4789]: I1122 08:20:23.048674 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/7ab63a4d-2e35-4b93-803b-13a38a43b8d6-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"7ab63a4d-2e35-4b93-803b-13a38a43b8d6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:20:23 crc kubenswrapper[4789]: I1122 08:20:23.048723 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/7ab63a4d-2e35-4b93-803b-13a38a43b8d6-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"7ab63a4d-2e35-4b93-803b-13a38a43b8d6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:20:23 crc kubenswrapper[4789]: I1122 08:20:23.048743 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/7ab63a4d-2e35-4b93-803b-13a38a43b8d6-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"7ab63a4d-2e35-4b93-803b-13a38a43b8d6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:20:23 crc kubenswrapper[4789]: I1122 08:20:23.048783 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/7ab63a4d-2e35-4b93-803b-13a38a43b8d6-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"7ab63a4d-2e35-4b93-803b-13a38a43b8d6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:20:23 crc kubenswrapper[4789]: I1122 08:20:23.048814 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"7ab63a4d-2e35-4b93-803b-13a38a43b8d6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:20:23 crc kubenswrapper[4789]: I1122 08:20:23.049124 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/7ab63a4d-2e35-4b93-803b-13a38a43b8d6-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"7ab63a4d-2e35-4b93-803b-13a38a43b8d6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:20:23 crc kubenswrapper[4789]: I1122 08:20:23.049185 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/7ab63a4d-2e35-4b93-803b-13a38a43b8d6-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"7ab63a4d-2e35-4b93-803b-13a38a43b8d6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:20:23 crc kubenswrapper[4789]: I1122 08:20:23.049225 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q7zwn\" (UniqueName: \"kubernetes.io/projected/7ab63a4d-2e35-4b93-803b-13a38a43b8d6-kube-api-access-q7zwn\") pod \"rabbitmq-cell1-server-0\" (UID: \"7ab63a4d-2e35-4b93-803b-13a38a43b8d6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:20:23 crc kubenswrapper[4789]: I1122 08:20:23.049357 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7ab63a4d-2e35-4b93-803b-13a38a43b8d6-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"7ab63a4d-2e35-4b93-803b-13a38a43b8d6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:20:23 crc kubenswrapper[4789]: I1122 08:20:23.049465 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/7ab63a4d-2e35-4b93-803b-13a38a43b8d6-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"7ab63a4d-2e35-4b93-803b-13a38a43b8d6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:20:23 crc kubenswrapper[4789]: I1122 08:20:23.071260 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 22 08:20:23 crc kubenswrapper[4789]: I1122 08:20:23.151199 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/7ab63a4d-2e35-4b93-803b-13a38a43b8d6-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"7ab63a4d-2e35-4b93-803b-13a38a43b8d6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:20:23 crc kubenswrapper[4789]: I1122 08:20:23.151486 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/7ab63a4d-2e35-4b93-803b-13a38a43b8d6-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"7ab63a4d-2e35-4b93-803b-13a38a43b8d6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:20:23 crc kubenswrapper[4789]: I1122 08:20:23.151524 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/7ab63a4d-2e35-4b93-803b-13a38a43b8d6-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"7ab63a4d-2e35-4b93-803b-13a38a43b8d6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:20:23 crc kubenswrapper[4789]: I1122 08:20:23.151544 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/7ab63a4d-2e35-4b93-803b-13a38a43b8d6-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"7ab63a4d-2e35-4b93-803b-13a38a43b8d6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:20:23 crc kubenswrapper[4789]: I1122 08:20:23.151575 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"7ab63a4d-2e35-4b93-803b-13a38a43b8d6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:20:23 crc kubenswrapper[4789]: I1122 08:20:23.151654 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/7ab63a4d-2e35-4b93-803b-13a38a43b8d6-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"7ab63a4d-2e35-4b93-803b-13a38a43b8d6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:20:23 crc kubenswrapper[4789]: I1122 08:20:23.151680 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/7ab63a4d-2e35-4b93-803b-13a38a43b8d6-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"7ab63a4d-2e35-4b93-803b-13a38a43b8d6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:20:23 crc kubenswrapper[4789]: I1122 08:20:23.151706 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q7zwn\" (UniqueName: \"kubernetes.io/projected/7ab63a4d-2e35-4b93-803b-13a38a43b8d6-kube-api-access-q7zwn\") pod \"rabbitmq-cell1-server-0\" (UID: \"7ab63a4d-2e35-4b93-803b-13a38a43b8d6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:20:23 crc kubenswrapper[4789]: I1122 08:20:23.151764 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7ab63a4d-2e35-4b93-803b-13a38a43b8d6-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"7ab63a4d-2e35-4b93-803b-13a38a43b8d6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:20:23 crc kubenswrapper[4789]: I1122 08:20:23.151810 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/7ab63a4d-2e35-4b93-803b-13a38a43b8d6-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"7ab63a4d-2e35-4b93-803b-13a38a43b8d6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:20:23 crc kubenswrapper[4789]: I1122 08:20:23.151868 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/7ab63a4d-2e35-4b93-803b-13a38a43b8d6-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"7ab63a4d-2e35-4b93-803b-13a38a43b8d6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:20:23 crc kubenswrapper[4789]: I1122 08:20:23.153123 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/7ab63a4d-2e35-4b93-803b-13a38a43b8d6-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"7ab63a4d-2e35-4b93-803b-13a38a43b8d6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:20:23 crc kubenswrapper[4789]: I1122 08:20:23.156603 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7ab63a4d-2e35-4b93-803b-13a38a43b8d6-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"7ab63a4d-2e35-4b93-803b-13a38a43b8d6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:20:23 crc kubenswrapper[4789]: I1122 08:20:23.157157 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/7ab63a4d-2e35-4b93-803b-13a38a43b8d6-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"7ab63a4d-2e35-4b93-803b-13a38a43b8d6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:20:23 crc kubenswrapper[4789]: I1122 08:20:23.157219 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/7ab63a4d-2e35-4b93-803b-13a38a43b8d6-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"7ab63a4d-2e35-4b93-803b-13a38a43b8d6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:20:23 crc kubenswrapper[4789]: I1122 08:20:23.157278 4789 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"7ab63a4d-2e35-4b93-803b-13a38a43b8d6\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:20:23 crc kubenswrapper[4789]: I1122 08:20:23.158105 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/7ab63a4d-2e35-4b93-803b-13a38a43b8d6-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"7ab63a4d-2e35-4b93-803b-13a38a43b8d6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:20:23 crc kubenswrapper[4789]: I1122 08:20:23.158484 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/7ab63a4d-2e35-4b93-803b-13a38a43b8d6-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"7ab63a4d-2e35-4b93-803b-13a38a43b8d6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:20:23 crc kubenswrapper[4789]: I1122 08:20:23.158976 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/7ab63a4d-2e35-4b93-803b-13a38a43b8d6-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"7ab63a4d-2e35-4b93-803b-13a38a43b8d6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:20:23 crc kubenswrapper[4789]: I1122 08:20:23.169613 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/7ab63a4d-2e35-4b93-803b-13a38a43b8d6-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"7ab63a4d-2e35-4b93-803b-13a38a43b8d6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:20:23 crc kubenswrapper[4789]: I1122 08:20:23.169692 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/7ab63a4d-2e35-4b93-803b-13a38a43b8d6-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"7ab63a4d-2e35-4b93-803b-13a38a43b8d6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:20:23 crc kubenswrapper[4789]: I1122 08:20:23.179488 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q7zwn\" (UniqueName: \"kubernetes.io/projected/7ab63a4d-2e35-4b93-803b-13a38a43b8d6-kube-api-access-q7zwn\") pod \"rabbitmq-cell1-server-0\" (UID: \"7ab63a4d-2e35-4b93-803b-13a38a43b8d6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:20:23 crc kubenswrapper[4789]: I1122 08:20:23.201500 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"7ab63a4d-2e35-4b93-803b-13a38a43b8d6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:20:23 crc kubenswrapper[4789]: I1122 08:20:23.297197 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:20:23 crc kubenswrapper[4789]: I1122 08:20:23.511400 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 22 08:20:23 crc kubenswrapper[4789]: W1122 08:20:23.526918 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1f834b47_b9a1_4813_bcac_3b5161eceeac.slice/crio-14453ac0f1cd33a5db9463dcb750367514a8f82b5832c3657f06bc2bbca2f1cc WatchSource:0}: Error finding container 14453ac0f1cd33a5db9463dcb750367514a8f82b5832c3657f06bc2bbca2f1cc: Status 404 returned error can't find the container with id 14453ac0f1cd33a5db9463dcb750367514a8f82b5832c3657f06bc2bbca2f1cc Nov 22 08:20:23 crc kubenswrapper[4789]: I1122 08:20:23.561086 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ccxrn" event={"ID":"70a0fabb-bd28-4991-87b7-95236f639ddb","Type":"ContainerStarted","Data":"c7ec533ad1968148cca44a814c0fb8cecc5ba6f1805ce8d5609fd0de736207a6"} Nov 22 08:20:23 crc kubenswrapper[4789]: I1122 08:20:23.563242 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"1f834b47-b9a1-4813-bcac-3b5161eceeac","Type":"ContainerStarted","Data":"14453ac0f1cd33a5db9463dcb750367514a8f82b5832c3657f06bc2bbca2f1cc"} Nov 22 08:20:23 crc kubenswrapper[4789]: W1122 08:20:23.760871 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7ab63a4d_2e35_4b93_803b_13a38a43b8d6.slice/crio-89a1f4b1761d8e55dd7dfbef498b9c8228cb1325b845a4cd92ade18013e840b8 WatchSource:0}: Error finding container 89a1f4b1761d8e55dd7dfbef498b9c8228cb1325b845a4cd92ade18013e840b8: Status 404 returned error can't find the container with id 89a1f4b1761d8e55dd7dfbef498b9c8228cb1325b845a4cd92ade18013e840b8 Nov 22 08:20:23 crc kubenswrapper[4789]: I1122 08:20:23.765327 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 22 08:20:23 crc kubenswrapper[4789]: I1122 08:20:23.975887 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af27dae6-c617-4d16-bb53-19057df7689b" path="/var/lib/kubelet/pods/af27dae6-c617-4d16-bb53-19057df7689b/volumes" Nov 22 08:20:23 crc kubenswrapper[4789]: I1122 08:20:23.979228 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="df1ccc5e-c20c-40ce-bb76-17da02c4bc8c" path="/var/lib/kubelet/pods/df1ccc5e-c20c-40ce-bb76-17da02c4bc8c/volumes" Nov 22 08:20:24 crc kubenswrapper[4789]: I1122 08:20:24.572856 4789 generic.go:334] "Generic (PLEG): container finished" podID="70a0fabb-bd28-4991-87b7-95236f639ddb" containerID="c7ec533ad1968148cca44a814c0fb8cecc5ba6f1805ce8d5609fd0de736207a6" exitCode=0 Nov 22 08:20:24 crc kubenswrapper[4789]: I1122 08:20:24.572921 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ccxrn" event={"ID":"70a0fabb-bd28-4991-87b7-95236f639ddb","Type":"ContainerDied","Data":"c7ec533ad1968148cca44a814c0fb8cecc5ba6f1805ce8d5609fd0de736207a6"} Nov 22 08:20:24 crc kubenswrapper[4789]: I1122 08:20:24.574834 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"7ab63a4d-2e35-4b93-803b-13a38a43b8d6","Type":"ContainerStarted","Data":"89a1f4b1761d8e55dd7dfbef498b9c8228cb1325b845a4cd92ade18013e840b8"} Nov 22 08:20:25 crc kubenswrapper[4789]: I1122 08:20:25.585901 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ccxrn" event={"ID":"70a0fabb-bd28-4991-87b7-95236f639ddb","Type":"ContainerStarted","Data":"bf24bad4cf9f730c08c10af9dcc81cda4c4d4644510e1624d12b515e8bde48e6"} Nov 22 08:20:25 crc kubenswrapper[4789]: I1122 08:20:25.588344 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"1f834b47-b9a1-4813-bcac-3b5161eceeac","Type":"ContainerStarted","Data":"e6918ce58dc9f9b564d498294a7689318093d05b1c5621c30d7ed281dee0220c"} Nov 22 08:20:25 crc kubenswrapper[4789]: I1122 08:20:25.590531 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"7ab63a4d-2e35-4b93-803b-13a38a43b8d6","Type":"ContainerStarted","Data":"bb525204bedb37c4d25f2a2754f8c9a7fac03f11f8e707bc850991ebc78b8d46"} Nov 22 08:20:25 crc kubenswrapper[4789]: I1122 08:20:25.604728 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-ccxrn" podStartSLOduration=3.075083893 podStartE2EDuration="5.604707829s" podCreationTimestamp="2025-11-22 08:20:20 +0000 UTC" firstStartedPulling="2025-11-22 08:20:22.530533828 +0000 UTC m=+1436.764934101" lastFinishedPulling="2025-11-22 08:20:25.060157754 +0000 UTC m=+1439.294558037" observedRunningTime="2025-11-22 08:20:25.601303926 +0000 UTC m=+1439.835704219" watchObservedRunningTime="2025-11-22 08:20:25.604707829 +0000 UTC m=+1439.839108102" Nov 22 08:20:26 crc kubenswrapper[4789]: I1122 08:20:26.289259 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-568675b579-28s9m"] Nov 22 08:20:26 crc kubenswrapper[4789]: I1122 08:20:26.291268 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-568675b579-28s9m" Nov 22 08:20:26 crc kubenswrapper[4789]: I1122 08:20:26.294435 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Nov 22 08:20:26 crc kubenswrapper[4789]: I1122 08:20:26.301107 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-568675b579-28s9m"] Nov 22 08:20:26 crc kubenswrapper[4789]: I1122 08:20:26.415434 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ftjjp\" (UniqueName: \"kubernetes.io/projected/93590b86-d27a-43f3-8b03-761b7cbf3c4c-kube-api-access-ftjjp\") pod \"dnsmasq-dns-568675b579-28s9m\" (UID: \"93590b86-d27a-43f3-8b03-761b7cbf3c4c\") " pod="openstack/dnsmasq-dns-568675b579-28s9m" Nov 22 08:20:26 crc kubenswrapper[4789]: I1122 08:20:26.415852 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/93590b86-d27a-43f3-8b03-761b7cbf3c4c-ovsdbserver-sb\") pod \"dnsmasq-dns-568675b579-28s9m\" (UID: \"93590b86-d27a-43f3-8b03-761b7cbf3c4c\") " pod="openstack/dnsmasq-dns-568675b579-28s9m" Nov 22 08:20:26 crc kubenswrapper[4789]: I1122 08:20:26.415879 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/93590b86-d27a-43f3-8b03-761b7cbf3c4c-ovsdbserver-nb\") pod \"dnsmasq-dns-568675b579-28s9m\" (UID: \"93590b86-d27a-43f3-8b03-761b7cbf3c4c\") " pod="openstack/dnsmasq-dns-568675b579-28s9m" Nov 22 08:20:26 crc kubenswrapper[4789]: I1122 08:20:26.415937 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/93590b86-d27a-43f3-8b03-761b7cbf3c4c-config\") pod \"dnsmasq-dns-568675b579-28s9m\" (UID: \"93590b86-d27a-43f3-8b03-761b7cbf3c4c\") " pod="openstack/dnsmasq-dns-568675b579-28s9m" Nov 22 08:20:26 crc kubenswrapper[4789]: I1122 08:20:26.416026 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/93590b86-d27a-43f3-8b03-761b7cbf3c4c-openstack-edpm-ipam\") pod \"dnsmasq-dns-568675b579-28s9m\" (UID: \"93590b86-d27a-43f3-8b03-761b7cbf3c4c\") " pod="openstack/dnsmasq-dns-568675b579-28s9m" Nov 22 08:20:26 crc kubenswrapper[4789]: I1122 08:20:26.416174 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/93590b86-d27a-43f3-8b03-761b7cbf3c4c-dns-svc\") pod \"dnsmasq-dns-568675b579-28s9m\" (UID: \"93590b86-d27a-43f3-8b03-761b7cbf3c4c\") " pod="openstack/dnsmasq-dns-568675b579-28s9m" Nov 22 08:20:26 crc kubenswrapper[4789]: I1122 08:20:26.518032 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ftjjp\" (UniqueName: \"kubernetes.io/projected/93590b86-d27a-43f3-8b03-761b7cbf3c4c-kube-api-access-ftjjp\") pod \"dnsmasq-dns-568675b579-28s9m\" (UID: \"93590b86-d27a-43f3-8b03-761b7cbf3c4c\") " pod="openstack/dnsmasq-dns-568675b579-28s9m" Nov 22 08:20:26 crc kubenswrapper[4789]: I1122 08:20:26.518101 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/93590b86-d27a-43f3-8b03-761b7cbf3c4c-ovsdbserver-sb\") pod \"dnsmasq-dns-568675b579-28s9m\" (UID: \"93590b86-d27a-43f3-8b03-761b7cbf3c4c\") " pod="openstack/dnsmasq-dns-568675b579-28s9m" Nov 22 08:20:26 crc kubenswrapper[4789]: I1122 08:20:26.518117 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/93590b86-d27a-43f3-8b03-761b7cbf3c4c-ovsdbserver-nb\") pod \"dnsmasq-dns-568675b579-28s9m\" (UID: \"93590b86-d27a-43f3-8b03-761b7cbf3c4c\") " pod="openstack/dnsmasq-dns-568675b579-28s9m" Nov 22 08:20:26 crc kubenswrapper[4789]: I1122 08:20:26.518144 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/93590b86-d27a-43f3-8b03-761b7cbf3c4c-config\") pod \"dnsmasq-dns-568675b579-28s9m\" (UID: \"93590b86-d27a-43f3-8b03-761b7cbf3c4c\") " pod="openstack/dnsmasq-dns-568675b579-28s9m" Nov 22 08:20:26 crc kubenswrapper[4789]: I1122 08:20:26.518164 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/93590b86-d27a-43f3-8b03-761b7cbf3c4c-openstack-edpm-ipam\") pod \"dnsmasq-dns-568675b579-28s9m\" (UID: \"93590b86-d27a-43f3-8b03-761b7cbf3c4c\") " pod="openstack/dnsmasq-dns-568675b579-28s9m" Nov 22 08:20:26 crc kubenswrapper[4789]: I1122 08:20:26.518206 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/93590b86-d27a-43f3-8b03-761b7cbf3c4c-dns-svc\") pod \"dnsmasq-dns-568675b579-28s9m\" (UID: \"93590b86-d27a-43f3-8b03-761b7cbf3c4c\") " pod="openstack/dnsmasq-dns-568675b579-28s9m" Nov 22 08:20:26 crc kubenswrapper[4789]: I1122 08:20:26.519133 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/93590b86-d27a-43f3-8b03-761b7cbf3c4c-dns-svc\") pod \"dnsmasq-dns-568675b579-28s9m\" (UID: \"93590b86-d27a-43f3-8b03-761b7cbf3c4c\") " pod="openstack/dnsmasq-dns-568675b579-28s9m" Nov 22 08:20:26 crc kubenswrapper[4789]: I1122 08:20:26.519170 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/93590b86-d27a-43f3-8b03-761b7cbf3c4c-config\") pod \"dnsmasq-dns-568675b579-28s9m\" (UID: \"93590b86-d27a-43f3-8b03-761b7cbf3c4c\") " pod="openstack/dnsmasq-dns-568675b579-28s9m" Nov 22 08:20:26 crc kubenswrapper[4789]: I1122 08:20:26.519253 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/93590b86-d27a-43f3-8b03-761b7cbf3c4c-ovsdbserver-sb\") pod \"dnsmasq-dns-568675b579-28s9m\" (UID: \"93590b86-d27a-43f3-8b03-761b7cbf3c4c\") " pod="openstack/dnsmasq-dns-568675b579-28s9m" Nov 22 08:20:26 crc kubenswrapper[4789]: I1122 08:20:26.519364 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/93590b86-d27a-43f3-8b03-761b7cbf3c4c-openstack-edpm-ipam\") pod \"dnsmasq-dns-568675b579-28s9m\" (UID: \"93590b86-d27a-43f3-8b03-761b7cbf3c4c\") " pod="openstack/dnsmasq-dns-568675b579-28s9m" Nov 22 08:20:26 crc kubenswrapper[4789]: I1122 08:20:26.519386 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/93590b86-d27a-43f3-8b03-761b7cbf3c4c-ovsdbserver-nb\") pod \"dnsmasq-dns-568675b579-28s9m\" (UID: \"93590b86-d27a-43f3-8b03-761b7cbf3c4c\") " pod="openstack/dnsmasq-dns-568675b579-28s9m" Nov 22 08:20:26 crc kubenswrapper[4789]: I1122 08:20:26.536469 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ftjjp\" (UniqueName: \"kubernetes.io/projected/93590b86-d27a-43f3-8b03-761b7cbf3c4c-kube-api-access-ftjjp\") pod \"dnsmasq-dns-568675b579-28s9m\" (UID: \"93590b86-d27a-43f3-8b03-761b7cbf3c4c\") " pod="openstack/dnsmasq-dns-568675b579-28s9m" Nov 22 08:20:26 crc kubenswrapper[4789]: I1122 08:20:26.618494 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-568675b579-28s9m" Nov 22 08:20:27 crc kubenswrapper[4789]: I1122 08:20:27.064089 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-568675b579-28s9m"] Nov 22 08:20:27 crc kubenswrapper[4789]: I1122 08:20:27.264050 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-mbqjc" Nov 22 08:20:27 crc kubenswrapper[4789]: I1122 08:20:27.264332 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-mbqjc" Nov 22 08:20:27 crc kubenswrapper[4789]: I1122 08:20:27.313097 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-mbqjc" Nov 22 08:20:27 crc kubenswrapper[4789]: I1122 08:20:27.606794 4789 generic.go:334] "Generic (PLEG): container finished" podID="93590b86-d27a-43f3-8b03-761b7cbf3c4c" containerID="396e7e5f01f96600a936c1c13abf483b28356df97c2da01c918f9a5f6f757fbf" exitCode=0 Nov 22 08:20:27 crc kubenswrapper[4789]: I1122 08:20:27.606860 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-568675b579-28s9m" event={"ID":"93590b86-d27a-43f3-8b03-761b7cbf3c4c","Type":"ContainerDied","Data":"396e7e5f01f96600a936c1c13abf483b28356df97c2da01c918f9a5f6f757fbf"} Nov 22 08:20:27 crc kubenswrapper[4789]: I1122 08:20:27.606906 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-568675b579-28s9m" event={"ID":"93590b86-d27a-43f3-8b03-761b7cbf3c4c","Type":"ContainerStarted","Data":"20ea6dc88ee07fc9c4aa16e905f553c0199da3d7a366f29216a883dc17a35e46"} Nov 22 08:20:27 crc kubenswrapper[4789]: I1122 08:20:27.657152 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-mbqjc" Nov 22 08:20:28 crc kubenswrapper[4789]: I1122 08:20:28.509127 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mbqjc"] Nov 22 08:20:28 crc kubenswrapper[4789]: I1122 08:20:28.616230 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-568675b579-28s9m" event={"ID":"93590b86-d27a-43f3-8b03-761b7cbf3c4c","Type":"ContainerStarted","Data":"80a3f643f8e497908ffd57ba12413d1beda4d63347ba30a88f88378b5a0be4cb"} Nov 22 08:20:28 crc kubenswrapper[4789]: I1122 08:20:28.616691 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-568675b579-28s9m" Nov 22 08:20:28 crc kubenswrapper[4789]: I1122 08:20:28.634802 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-568675b579-28s9m" podStartSLOduration=2.634781405 podStartE2EDuration="2.634781405s" podCreationTimestamp="2025-11-22 08:20:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:20:28.633786988 +0000 UTC m=+1442.868187261" watchObservedRunningTime="2025-11-22 08:20:28.634781405 +0000 UTC m=+1442.869181678" Nov 22 08:20:29 crc kubenswrapper[4789]: I1122 08:20:29.625204 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-mbqjc" podUID="96aba4ce-5859-4803-ac23-919b4bad4f0d" containerName="registry-server" containerID="cri-o://4d04a10903c50e595852e4fc8585c46a927df8a685c440af81660580bdb14527" gracePeriod=2 Nov 22 08:20:31 crc kubenswrapper[4789]: I1122 08:20:31.250350 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-ccxrn" Nov 22 08:20:31 crc kubenswrapper[4789]: I1122 08:20:31.251336 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-ccxrn" Nov 22 08:20:31 crc kubenswrapper[4789]: I1122 08:20:31.309413 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-ccxrn" Nov 22 08:20:31 crc kubenswrapper[4789]: I1122 08:20:31.684221 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-ccxrn" Nov 22 08:20:32 crc kubenswrapper[4789]: I1122 08:20:32.113447 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-ccxrn"] Nov 22 08:20:32 crc kubenswrapper[4789]: I1122 08:20:32.652316 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mbqjc" event={"ID":"96aba4ce-5859-4803-ac23-919b4bad4f0d","Type":"ContainerDied","Data":"4d04a10903c50e595852e4fc8585c46a927df8a685c440af81660580bdb14527"} Nov 22 08:20:32 crc kubenswrapper[4789]: I1122 08:20:32.652319 4789 generic.go:334] "Generic (PLEG): container finished" podID="96aba4ce-5859-4803-ac23-919b4bad4f0d" containerID="4d04a10903c50e595852e4fc8585c46a927df8a685c440af81660580bdb14527" exitCode=0 Nov 22 08:20:32 crc kubenswrapper[4789]: I1122 08:20:32.735323 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mbqjc" Nov 22 08:20:32 crc kubenswrapper[4789]: I1122 08:20:32.831244 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96aba4ce-5859-4803-ac23-919b4bad4f0d-catalog-content\") pod \"96aba4ce-5859-4803-ac23-919b4bad4f0d\" (UID: \"96aba4ce-5859-4803-ac23-919b4bad4f0d\") " Nov 22 08:20:32 crc kubenswrapper[4789]: I1122 08:20:32.831301 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96aba4ce-5859-4803-ac23-919b4bad4f0d-utilities\") pod \"96aba4ce-5859-4803-ac23-919b4bad4f0d\" (UID: \"96aba4ce-5859-4803-ac23-919b4bad4f0d\") " Nov 22 08:20:32 crc kubenswrapper[4789]: I1122 08:20:32.831374 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rtkwh\" (UniqueName: \"kubernetes.io/projected/96aba4ce-5859-4803-ac23-919b4bad4f0d-kube-api-access-rtkwh\") pod \"96aba4ce-5859-4803-ac23-919b4bad4f0d\" (UID: \"96aba4ce-5859-4803-ac23-919b4bad4f0d\") " Nov 22 08:20:32 crc kubenswrapper[4789]: I1122 08:20:32.832811 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/96aba4ce-5859-4803-ac23-919b4bad4f0d-utilities" (OuterVolumeSpecName: "utilities") pod "96aba4ce-5859-4803-ac23-919b4bad4f0d" (UID: "96aba4ce-5859-4803-ac23-919b4bad4f0d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:20:32 crc kubenswrapper[4789]: I1122 08:20:32.837415 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96aba4ce-5859-4803-ac23-919b4bad4f0d-kube-api-access-rtkwh" (OuterVolumeSpecName: "kube-api-access-rtkwh") pod "96aba4ce-5859-4803-ac23-919b4bad4f0d" (UID: "96aba4ce-5859-4803-ac23-919b4bad4f0d"). InnerVolumeSpecName "kube-api-access-rtkwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:20:32 crc kubenswrapper[4789]: I1122 08:20:32.848332 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/96aba4ce-5859-4803-ac23-919b4bad4f0d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "96aba4ce-5859-4803-ac23-919b4bad4f0d" (UID: "96aba4ce-5859-4803-ac23-919b4bad4f0d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:20:32 crc kubenswrapper[4789]: I1122 08:20:32.932806 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96aba4ce-5859-4803-ac23-919b4bad4f0d-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 08:20:32 crc kubenswrapper[4789]: I1122 08:20:32.932837 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96aba4ce-5859-4803-ac23-919b4bad4f0d-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 08:20:32 crc kubenswrapper[4789]: I1122 08:20:32.932846 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rtkwh\" (UniqueName: \"kubernetes.io/projected/96aba4ce-5859-4803-ac23-919b4bad4f0d-kube-api-access-rtkwh\") on node \"crc\" DevicePath \"\"" Nov 22 08:20:33 crc kubenswrapper[4789]: I1122 08:20:33.664593 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mbqjc" Nov 22 08:20:33 crc kubenswrapper[4789]: I1122 08:20:33.664666 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mbqjc" event={"ID":"96aba4ce-5859-4803-ac23-919b4bad4f0d","Type":"ContainerDied","Data":"6c750f1bf3db12c2532e4fbc74cfdc0e869419c14ab1b9683f46b236ffba6720"} Nov 22 08:20:33 crc kubenswrapper[4789]: I1122 08:20:33.664733 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-ccxrn" podUID="70a0fabb-bd28-4991-87b7-95236f639ddb" containerName="registry-server" containerID="cri-o://bf24bad4cf9f730c08c10af9dcc81cda4c4d4644510e1624d12b515e8bde48e6" gracePeriod=2 Nov 22 08:20:33 crc kubenswrapper[4789]: I1122 08:20:33.665206 4789 scope.go:117] "RemoveContainer" containerID="4d04a10903c50e595852e4fc8585c46a927df8a685c440af81660580bdb14527" Nov 22 08:20:33 crc kubenswrapper[4789]: I1122 08:20:33.692852 4789 scope.go:117] "RemoveContainer" containerID="c6dec650ed5319f6d68f7fadc96076e9baf648b9cfa50a9fb97312709a523024" Nov 22 08:20:33 crc kubenswrapper[4789]: I1122 08:20:33.700606 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mbqjc"] Nov 22 08:20:33 crc kubenswrapper[4789]: I1122 08:20:33.707285 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-mbqjc"] Nov 22 08:20:33 crc kubenswrapper[4789]: I1122 08:20:33.735712 4789 scope.go:117] "RemoveContainer" containerID="1bf14414f405a586fa8ace6440f7fa4015e58bf9c2db6ad8059273ef684bc82c" Nov 22 08:20:33 crc kubenswrapper[4789]: I1122 08:20:33.975789 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96aba4ce-5859-4803-ac23-919b4bad4f0d" path="/var/lib/kubelet/pods/96aba4ce-5859-4803-ac23-919b4bad4f0d/volumes" Nov 22 08:20:34 crc kubenswrapper[4789]: I1122 08:20:34.679618 4789 generic.go:334] "Generic (PLEG): container finished" podID="70a0fabb-bd28-4991-87b7-95236f639ddb" containerID="bf24bad4cf9f730c08c10af9dcc81cda4c4d4644510e1624d12b515e8bde48e6" exitCode=0 Nov 22 08:20:34 crc kubenswrapper[4789]: I1122 08:20:34.679699 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ccxrn" event={"ID":"70a0fabb-bd28-4991-87b7-95236f639ddb","Type":"ContainerDied","Data":"bf24bad4cf9f730c08c10af9dcc81cda4c4d4644510e1624d12b515e8bde48e6"} Nov 22 08:20:34 crc kubenswrapper[4789]: I1122 08:20:34.680132 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ccxrn" event={"ID":"70a0fabb-bd28-4991-87b7-95236f639ddb","Type":"ContainerDied","Data":"b48a4dc6ea60edc3412a6182e0aa2e10a90a951f1dfd17ad46a8d8e439963f61"} Nov 22 08:20:34 crc kubenswrapper[4789]: I1122 08:20:34.680149 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b48a4dc6ea60edc3412a6182e0aa2e10a90a951f1dfd17ad46a8d8e439963f61" Nov 22 08:20:34 crc kubenswrapper[4789]: I1122 08:20:34.741424 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ccxrn" Nov 22 08:20:34 crc kubenswrapper[4789]: I1122 08:20:34.867049 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70a0fabb-bd28-4991-87b7-95236f639ddb-utilities\") pod \"70a0fabb-bd28-4991-87b7-95236f639ddb\" (UID: \"70a0fabb-bd28-4991-87b7-95236f639ddb\") " Nov 22 08:20:34 crc kubenswrapper[4789]: I1122 08:20:34.867148 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70a0fabb-bd28-4991-87b7-95236f639ddb-catalog-content\") pod \"70a0fabb-bd28-4991-87b7-95236f639ddb\" (UID: \"70a0fabb-bd28-4991-87b7-95236f639ddb\") " Nov 22 08:20:34 crc kubenswrapper[4789]: I1122 08:20:34.867234 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2h4tv\" (UniqueName: \"kubernetes.io/projected/70a0fabb-bd28-4991-87b7-95236f639ddb-kube-api-access-2h4tv\") pod \"70a0fabb-bd28-4991-87b7-95236f639ddb\" (UID: \"70a0fabb-bd28-4991-87b7-95236f639ddb\") " Nov 22 08:20:34 crc kubenswrapper[4789]: I1122 08:20:34.868370 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/70a0fabb-bd28-4991-87b7-95236f639ddb-utilities" (OuterVolumeSpecName: "utilities") pod "70a0fabb-bd28-4991-87b7-95236f639ddb" (UID: "70a0fabb-bd28-4991-87b7-95236f639ddb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:20:34 crc kubenswrapper[4789]: I1122 08:20:34.872737 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70a0fabb-bd28-4991-87b7-95236f639ddb-kube-api-access-2h4tv" (OuterVolumeSpecName: "kube-api-access-2h4tv") pod "70a0fabb-bd28-4991-87b7-95236f639ddb" (UID: "70a0fabb-bd28-4991-87b7-95236f639ddb"). InnerVolumeSpecName "kube-api-access-2h4tv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:20:34 crc kubenswrapper[4789]: I1122 08:20:34.963008 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/70a0fabb-bd28-4991-87b7-95236f639ddb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "70a0fabb-bd28-4991-87b7-95236f639ddb" (UID: "70a0fabb-bd28-4991-87b7-95236f639ddb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:20:34 crc kubenswrapper[4789]: I1122 08:20:34.969168 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70a0fabb-bd28-4991-87b7-95236f639ddb-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 08:20:34 crc kubenswrapper[4789]: I1122 08:20:34.969215 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2h4tv\" (UniqueName: \"kubernetes.io/projected/70a0fabb-bd28-4991-87b7-95236f639ddb-kube-api-access-2h4tv\") on node \"crc\" DevicePath \"\"" Nov 22 08:20:34 crc kubenswrapper[4789]: I1122 08:20:34.969227 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70a0fabb-bd28-4991-87b7-95236f639ddb-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 08:20:35 crc kubenswrapper[4789]: I1122 08:20:35.689743 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ccxrn" Nov 22 08:20:35 crc kubenswrapper[4789]: I1122 08:20:35.721080 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-ccxrn"] Nov 22 08:20:35 crc kubenswrapper[4789]: I1122 08:20:35.728656 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-ccxrn"] Nov 22 08:20:35 crc kubenswrapper[4789]: I1122 08:20:35.978627 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="70a0fabb-bd28-4991-87b7-95236f639ddb" path="/var/lib/kubelet/pods/70a0fabb-bd28-4991-87b7-95236f639ddb/volumes" Nov 22 08:20:36 crc kubenswrapper[4789]: I1122 08:20:36.621153 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-568675b579-28s9m" Nov 22 08:20:36 crc kubenswrapper[4789]: I1122 08:20:36.690360 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-c9b558957-c7k8k"] Nov 22 08:20:36 crc kubenswrapper[4789]: I1122 08:20:36.690605 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-c9b558957-c7k8k" podUID="1899d9cc-6fb5-499f-8a8a-c7fe3b923e96" containerName="dnsmasq-dns" containerID="cri-o://56819ef5f9657982c799cca0b86e2b25a4b3e1c3c10a456999068c3f20196c16" gracePeriod=10 Nov 22 08:20:36 crc kubenswrapper[4789]: I1122 08:20:36.798257 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6dc44c56c-bclr4"] Nov 22 08:20:36 crc kubenswrapper[4789]: E1122 08:20:36.798627 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96aba4ce-5859-4803-ac23-919b4bad4f0d" containerName="extract-utilities" Nov 22 08:20:36 crc kubenswrapper[4789]: I1122 08:20:36.798645 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="96aba4ce-5859-4803-ac23-919b4bad4f0d" containerName="extract-utilities" Nov 22 08:20:36 crc kubenswrapper[4789]: E1122 08:20:36.798658 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70a0fabb-bd28-4991-87b7-95236f639ddb" containerName="extract-content" Nov 22 08:20:36 crc kubenswrapper[4789]: I1122 08:20:36.798665 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="70a0fabb-bd28-4991-87b7-95236f639ddb" containerName="extract-content" Nov 22 08:20:36 crc kubenswrapper[4789]: E1122 08:20:36.798693 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96aba4ce-5859-4803-ac23-919b4bad4f0d" containerName="registry-server" Nov 22 08:20:36 crc kubenswrapper[4789]: I1122 08:20:36.798700 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="96aba4ce-5859-4803-ac23-919b4bad4f0d" containerName="registry-server" Nov 22 08:20:36 crc kubenswrapper[4789]: E1122 08:20:36.798710 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70a0fabb-bd28-4991-87b7-95236f639ddb" containerName="registry-server" Nov 22 08:20:36 crc kubenswrapper[4789]: I1122 08:20:36.798716 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="70a0fabb-bd28-4991-87b7-95236f639ddb" containerName="registry-server" Nov 22 08:20:36 crc kubenswrapper[4789]: E1122 08:20:36.798725 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96aba4ce-5859-4803-ac23-919b4bad4f0d" containerName="extract-content" Nov 22 08:20:36 crc kubenswrapper[4789]: I1122 08:20:36.798730 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="96aba4ce-5859-4803-ac23-919b4bad4f0d" containerName="extract-content" Nov 22 08:20:36 crc kubenswrapper[4789]: E1122 08:20:36.798762 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70a0fabb-bd28-4991-87b7-95236f639ddb" containerName="extract-utilities" Nov 22 08:20:36 crc kubenswrapper[4789]: I1122 08:20:36.798768 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="70a0fabb-bd28-4991-87b7-95236f639ddb" containerName="extract-utilities" Nov 22 08:20:36 crc kubenswrapper[4789]: I1122 08:20:36.798918 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="96aba4ce-5859-4803-ac23-919b4bad4f0d" containerName="registry-server" Nov 22 08:20:36 crc kubenswrapper[4789]: I1122 08:20:36.798938 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="70a0fabb-bd28-4991-87b7-95236f639ddb" containerName="registry-server" Nov 22 08:20:36 crc kubenswrapper[4789]: I1122 08:20:36.799864 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6dc44c56c-bclr4" Nov 22 08:20:36 crc kubenswrapper[4789]: I1122 08:20:36.825515 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6dc44c56c-bclr4"] Nov 22 08:20:36 crc kubenswrapper[4789]: I1122 08:20:36.901875 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7785105e-179b-4b0b-9457-cd513d0b762d-ovsdbserver-nb\") pod \"dnsmasq-dns-6dc44c56c-bclr4\" (UID: \"7785105e-179b-4b0b-9457-cd513d0b762d\") " pod="openstack/dnsmasq-dns-6dc44c56c-bclr4" Nov 22 08:20:36 crc kubenswrapper[4789]: I1122 08:20:36.901930 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qgzxc\" (UniqueName: \"kubernetes.io/projected/7785105e-179b-4b0b-9457-cd513d0b762d-kube-api-access-qgzxc\") pod \"dnsmasq-dns-6dc44c56c-bclr4\" (UID: \"7785105e-179b-4b0b-9457-cd513d0b762d\") " pod="openstack/dnsmasq-dns-6dc44c56c-bclr4" Nov 22 08:20:36 crc kubenswrapper[4789]: I1122 08:20:36.902033 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/7785105e-179b-4b0b-9457-cd513d0b762d-openstack-edpm-ipam\") pod \"dnsmasq-dns-6dc44c56c-bclr4\" (UID: \"7785105e-179b-4b0b-9457-cd513d0b762d\") " pod="openstack/dnsmasq-dns-6dc44c56c-bclr4" Nov 22 08:20:36 crc kubenswrapper[4789]: I1122 08:20:36.902073 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7785105e-179b-4b0b-9457-cd513d0b762d-ovsdbserver-sb\") pod \"dnsmasq-dns-6dc44c56c-bclr4\" (UID: \"7785105e-179b-4b0b-9457-cd513d0b762d\") " pod="openstack/dnsmasq-dns-6dc44c56c-bclr4" Nov 22 08:20:36 crc kubenswrapper[4789]: I1122 08:20:36.902170 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7785105e-179b-4b0b-9457-cd513d0b762d-dns-svc\") pod \"dnsmasq-dns-6dc44c56c-bclr4\" (UID: \"7785105e-179b-4b0b-9457-cd513d0b762d\") " pod="openstack/dnsmasq-dns-6dc44c56c-bclr4" Nov 22 08:20:36 crc kubenswrapper[4789]: I1122 08:20:36.902241 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7785105e-179b-4b0b-9457-cd513d0b762d-config\") pod \"dnsmasq-dns-6dc44c56c-bclr4\" (UID: \"7785105e-179b-4b0b-9457-cd513d0b762d\") " pod="openstack/dnsmasq-dns-6dc44c56c-bclr4" Nov 22 08:20:37 crc kubenswrapper[4789]: I1122 08:20:37.003713 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/7785105e-179b-4b0b-9457-cd513d0b762d-openstack-edpm-ipam\") pod \"dnsmasq-dns-6dc44c56c-bclr4\" (UID: \"7785105e-179b-4b0b-9457-cd513d0b762d\") " pod="openstack/dnsmasq-dns-6dc44c56c-bclr4" Nov 22 08:20:37 crc kubenswrapper[4789]: I1122 08:20:37.004113 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7785105e-179b-4b0b-9457-cd513d0b762d-ovsdbserver-sb\") pod \"dnsmasq-dns-6dc44c56c-bclr4\" (UID: \"7785105e-179b-4b0b-9457-cd513d0b762d\") " pod="openstack/dnsmasq-dns-6dc44c56c-bclr4" Nov 22 08:20:37 crc kubenswrapper[4789]: I1122 08:20:37.004185 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7785105e-179b-4b0b-9457-cd513d0b762d-dns-svc\") pod \"dnsmasq-dns-6dc44c56c-bclr4\" (UID: \"7785105e-179b-4b0b-9457-cd513d0b762d\") " pod="openstack/dnsmasq-dns-6dc44c56c-bclr4" Nov 22 08:20:37 crc kubenswrapper[4789]: I1122 08:20:37.004240 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7785105e-179b-4b0b-9457-cd513d0b762d-config\") pod \"dnsmasq-dns-6dc44c56c-bclr4\" (UID: \"7785105e-179b-4b0b-9457-cd513d0b762d\") " pod="openstack/dnsmasq-dns-6dc44c56c-bclr4" Nov 22 08:20:37 crc kubenswrapper[4789]: I1122 08:20:37.004283 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7785105e-179b-4b0b-9457-cd513d0b762d-ovsdbserver-nb\") pod \"dnsmasq-dns-6dc44c56c-bclr4\" (UID: \"7785105e-179b-4b0b-9457-cd513d0b762d\") " pod="openstack/dnsmasq-dns-6dc44c56c-bclr4" Nov 22 08:20:37 crc kubenswrapper[4789]: I1122 08:20:37.004309 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qgzxc\" (UniqueName: \"kubernetes.io/projected/7785105e-179b-4b0b-9457-cd513d0b762d-kube-api-access-qgzxc\") pod \"dnsmasq-dns-6dc44c56c-bclr4\" (UID: \"7785105e-179b-4b0b-9457-cd513d0b762d\") " pod="openstack/dnsmasq-dns-6dc44c56c-bclr4" Nov 22 08:20:37 crc kubenswrapper[4789]: I1122 08:20:37.005595 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/7785105e-179b-4b0b-9457-cd513d0b762d-openstack-edpm-ipam\") pod \"dnsmasq-dns-6dc44c56c-bclr4\" (UID: \"7785105e-179b-4b0b-9457-cd513d0b762d\") " pod="openstack/dnsmasq-dns-6dc44c56c-bclr4" Nov 22 08:20:37 crc kubenswrapper[4789]: I1122 08:20:37.005819 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7785105e-179b-4b0b-9457-cd513d0b762d-dns-svc\") pod \"dnsmasq-dns-6dc44c56c-bclr4\" (UID: \"7785105e-179b-4b0b-9457-cd513d0b762d\") " pod="openstack/dnsmasq-dns-6dc44c56c-bclr4" Nov 22 08:20:37 crc kubenswrapper[4789]: I1122 08:20:37.006260 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7785105e-179b-4b0b-9457-cd513d0b762d-config\") pod \"dnsmasq-dns-6dc44c56c-bclr4\" (UID: \"7785105e-179b-4b0b-9457-cd513d0b762d\") " pod="openstack/dnsmasq-dns-6dc44c56c-bclr4" Nov 22 08:20:37 crc kubenswrapper[4789]: I1122 08:20:37.006452 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7785105e-179b-4b0b-9457-cd513d0b762d-ovsdbserver-nb\") pod \"dnsmasq-dns-6dc44c56c-bclr4\" (UID: \"7785105e-179b-4b0b-9457-cd513d0b762d\") " pod="openstack/dnsmasq-dns-6dc44c56c-bclr4" Nov 22 08:20:37 crc kubenswrapper[4789]: I1122 08:20:37.007473 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7785105e-179b-4b0b-9457-cd513d0b762d-ovsdbserver-sb\") pod \"dnsmasq-dns-6dc44c56c-bclr4\" (UID: \"7785105e-179b-4b0b-9457-cd513d0b762d\") " pod="openstack/dnsmasq-dns-6dc44c56c-bclr4" Nov 22 08:20:37 crc kubenswrapper[4789]: I1122 08:20:37.026471 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qgzxc\" (UniqueName: \"kubernetes.io/projected/7785105e-179b-4b0b-9457-cd513d0b762d-kube-api-access-qgzxc\") pod \"dnsmasq-dns-6dc44c56c-bclr4\" (UID: \"7785105e-179b-4b0b-9457-cd513d0b762d\") " pod="openstack/dnsmasq-dns-6dc44c56c-bclr4" Nov 22 08:20:37 crc kubenswrapper[4789]: I1122 08:20:37.168884 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-c9b558957-c7k8k" Nov 22 08:20:37 crc kubenswrapper[4789]: I1122 08:20:37.170790 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6dc44c56c-bclr4" Nov 22 08:20:37 crc kubenswrapper[4789]: I1122 08:20:37.209972 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1899d9cc-6fb5-499f-8a8a-c7fe3b923e96-config\") pod \"1899d9cc-6fb5-499f-8a8a-c7fe3b923e96\" (UID: \"1899d9cc-6fb5-499f-8a8a-c7fe3b923e96\") " Nov 22 08:20:37 crc kubenswrapper[4789]: I1122 08:20:37.210056 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fg72r\" (UniqueName: \"kubernetes.io/projected/1899d9cc-6fb5-499f-8a8a-c7fe3b923e96-kube-api-access-fg72r\") pod \"1899d9cc-6fb5-499f-8a8a-c7fe3b923e96\" (UID: \"1899d9cc-6fb5-499f-8a8a-c7fe3b923e96\") " Nov 22 08:20:37 crc kubenswrapper[4789]: I1122 08:20:37.210132 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1899d9cc-6fb5-499f-8a8a-c7fe3b923e96-ovsdbserver-sb\") pod \"1899d9cc-6fb5-499f-8a8a-c7fe3b923e96\" (UID: \"1899d9cc-6fb5-499f-8a8a-c7fe3b923e96\") " Nov 22 08:20:37 crc kubenswrapper[4789]: I1122 08:20:37.210190 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1899d9cc-6fb5-499f-8a8a-c7fe3b923e96-dns-svc\") pod \"1899d9cc-6fb5-499f-8a8a-c7fe3b923e96\" (UID: \"1899d9cc-6fb5-499f-8a8a-c7fe3b923e96\") " Nov 22 08:20:37 crc kubenswrapper[4789]: I1122 08:20:37.210283 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1899d9cc-6fb5-499f-8a8a-c7fe3b923e96-ovsdbserver-nb\") pod \"1899d9cc-6fb5-499f-8a8a-c7fe3b923e96\" (UID: \"1899d9cc-6fb5-499f-8a8a-c7fe3b923e96\") " Nov 22 08:20:37 crc kubenswrapper[4789]: I1122 08:20:37.214358 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1899d9cc-6fb5-499f-8a8a-c7fe3b923e96-kube-api-access-fg72r" (OuterVolumeSpecName: "kube-api-access-fg72r") pod "1899d9cc-6fb5-499f-8a8a-c7fe3b923e96" (UID: "1899d9cc-6fb5-499f-8a8a-c7fe3b923e96"). InnerVolumeSpecName "kube-api-access-fg72r". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:20:37 crc kubenswrapper[4789]: I1122 08:20:37.266534 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1899d9cc-6fb5-499f-8a8a-c7fe3b923e96-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "1899d9cc-6fb5-499f-8a8a-c7fe3b923e96" (UID: "1899d9cc-6fb5-499f-8a8a-c7fe3b923e96"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:20:37 crc kubenswrapper[4789]: I1122 08:20:37.266607 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1899d9cc-6fb5-499f-8a8a-c7fe3b923e96-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "1899d9cc-6fb5-499f-8a8a-c7fe3b923e96" (UID: "1899d9cc-6fb5-499f-8a8a-c7fe3b923e96"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:20:37 crc kubenswrapper[4789]: I1122 08:20:37.266668 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1899d9cc-6fb5-499f-8a8a-c7fe3b923e96-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "1899d9cc-6fb5-499f-8a8a-c7fe3b923e96" (UID: "1899d9cc-6fb5-499f-8a8a-c7fe3b923e96"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:20:37 crc kubenswrapper[4789]: I1122 08:20:37.288679 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1899d9cc-6fb5-499f-8a8a-c7fe3b923e96-config" (OuterVolumeSpecName: "config") pod "1899d9cc-6fb5-499f-8a8a-c7fe3b923e96" (UID: "1899d9cc-6fb5-499f-8a8a-c7fe3b923e96"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:20:37 crc kubenswrapper[4789]: I1122 08:20:37.312584 4789 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1899d9cc-6fb5-499f-8a8a-c7fe3b923e96-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 22 08:20:37 crc kubenswrapper[4789]: I1122 08:20:37.312618 4789 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1899d9cc-6fb5-499f-8a8a-c7fe3b923e96-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 22 08:20:37 crc kubenswrapper[4789]: I1122 08:20:37.312631 4789 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1899d9cc-6fb5-499f-8a8a-c7fe3b923e96-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 22 08:20:37 crc kubenswrapper[4789]: I1122 08:20:37.312643 4789 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1899d9cc-6fb5-499f-8a8a-c7fe3b923e96-config\") on node \"crc\" DevicePath \"\"" Nov 22 08:20:37 crc kubenswrapper[4789]: I1122 08:20:37.312656 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fg72r\" (UniqueName: \"kubernetes.io/projected/1899d9cc-6fb5-499f-8a8a-c7fe3b923e96-kube-api-access-fg72r\") on node \"crc\" DevicePath \"\"" Nov 22 08:20:37 crc kubenswrapper[4789]: I1122 08:20:37.625394 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6dc44c56c-bclr4"] Nov 22 08:20:37 crc kubenswrapper[4789]: W1122 08:20:37.627301 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7785105e_179b_4b0b_9457_cd513d0b762d.slice/crio-c2e34834666a28e734d9e59dfa0e4cf2b249b55fab28d0800cc4646e97c30629 WatchSource:0}: Error finding container c2e34834666a28e734d9e59dfa0e4cf2b249b55fab28d0800cc4646e97c30629: Status 404 returned error can't find the container with id c2e34834666a28e734d9e59dfa0e4cf2b249b55fab28d0800cc4646e97c30629 Nov 22 08:20:37 crc kubenswrapper[4789]: I1122 08:20:37.711832 4789 generic.go:334] "Generic (PLEG): container finished" podID="1899d9cc-6fb5-499f-8a8a-c7fe3b923e96" containerID="56819ef5f9657982c799cca0b86e2b25a4b3e1c3c10a456999068c3f20196c16" exitCode=0 Nov 22 08:20:37 crc kubenswrapper[4789]: I1122 08:20:37.711919 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-c9b558957-c7k8k" Nov 22 08:20:37 crc kubenswrapper[4789]: I1122 08:20:37.711934 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-c9b558957-c7k8k" event={"ID":"1899d9cc-6fb5-499f-8a8a-c7fe3b923e96","Type":"ContainerDied","Data":"56819ef5f9657982c799cca0b86e2b25a4b3e1c3c10a456999068c3f20196c16"} Nov 22 08:20:37 crc kubenswrapper[4789]: I1122 08:20:37.711961 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-c9b558957-c7k8k" event={"ID":"1899d9cc-6fb5-499f-8a8a-c7fe3b923e96","Type":"ContainerDied","Data":"9ee8362ad000dde85bad9761a8221d23863d0602d9bdf5f0ca85557aedb2fedc"} Nov 22 08:20:37 crc kubenswrapper[4789]: I1122 08:20:37.711978 4789 scope.go:117] "RemoveContainer" containerID="56819ef5f9657982c799cca0b86e2b25a4b3e1c3c10a456999068c3f20196c16" Nov 22 08:20:37 crc kubenswrapper[4789]: I1122 08:20:37.715392 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6dc44c56c-bclr4" event={"ID":"7785105e-179b-4b0b-9457-cd513d0b762d","Type":"ContainerStarted","Data":"c2e34834666a28e734d9e59dfa0e4cf2b249b55fab28d0800cc4646e97c30629"} Nov 22 08:20:37 crc kubenswrapper[4789]: I1122 08:20:37.742900 4789 scope.go:117] "RemoveContainer" containerID="298646ec1930685846c3079c5df5819831f8d41c2dcf045279ed940d562894d6" Nov 22 08:20:37 crc kubenswrapper[4789]: I1122 08:20:37.753899 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-c9b558957-c7k8k"] Nov 22 08:20:37 crc kubenswrapper[4789]: I1122 08:20:37.762271 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-c9b558957-c7k8k"] Nov 22 08:20:37 crc kubenswrapper[4789]: I1122 08:20:37.777266 4789 scope.go:117] "RemoveContainer" containerID="56819ef5f9657982c799cca0b86e2b25a4b3e1c3c10a456999068c3f20196c16" Nov 22 08:20:37 crc kubenswrapper[4789]: E1122 08:20:37.777795 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"56819ef5f9657982c799cca0b86e2b25a4b3e1c3c10a456999068c3f20196c16\": container with ID starting with 56819ef5f9657982c799cca0b86e2b25a4b3e1c3c10a456999068c3f20196c16 not found: ID does not exist" containerID="56819ef5f9657982c799cca0b86e2b25a4b3e1c3c10a456999068c3f20196c16" Nov 22 08:20:37 crc kubenswrapper[4789]: I1122 08:20:37.777845 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"56819ef5f9657982c799cca0b86e2b25a4b3e1c3c10a456999068c3f20196c16"} err="failed to get container status \"56819ef5f9657982c799cca0b86e2b25a4b3e1c3c10a456999068c3f20196c16\": rpc error: code = NotFound desc = could not find container \"56819ef5f9657982c799cca0b86e2b25a4b3e1c3c10a456999068c3f20196c16\": container with ID starting with 56819ef5f9657982c799cca0b86e2b25a4b3e1c3c10a456999068c3f20196c16 not found: ID does not exist" Nov 22 08:20:37 crc kubenswrapper[4789]: I1122 08:20:37.777877 4789 scope.go:117] "RemoveContainer" containerID="298646ec1930685846c3079c5df5819831f8d41c2dcf045279ed940d562894d6" Nov 22 08:20:37 crc kubenswrapper[4789]: E1122 08:20:37.778257 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"298646ec1930685846c3079c5df5819831f8d41c2dcf045279ed940d562894d6\": container with ID starting with 298646ec1930685846c3079c5df5819831f8d41c2dcf045279ed940d562894d6 not found: ID does not exist" containerID="298646ec1930685846c3079c5df5819831f8d41c2dcf045279ed940d562894d6" Nov 22 08:20:37 crc kubenswrapper[4789]: I1122 08:20:37.778291 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"298646ec1930685846c3079c5df5819831f8d41c2dcf045279ed940d562894d6"} err="failed to get container status \"298646ec1930685846c3079c5df5819831f8d41c2dcf045279ed940d562894d6\": rpc error: code = NotFound desc = could not find container \"298646ec1930685846c3079c5df5819831f8d41c2dcf045279ed940d562894d6\": container with ID starting with 298646ec1930685846c3079c5df5819831f8d41c2dcf045279ed940d562894d6 not found: ID does not exist" Nov 22 08:20:37 crc kubenswrapper[4789]: I1122 08:20:37.976950 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1899d9cc-6fb5-499f-8a8a-c7fe3b923e96" path="/var/lib/kubelet/pods/1899d9cc-6fb5-499f-8a8a-c7fe3b923e96/volumes" Nov 22 08:20:38 crc kubenswrapper[4789]: I1122 08:20:38.726462 4789 generic.go:334] "Generic (PLEG): container finished" podID="7785105e-179b-4b0b-9457-cd513d0b762d" containerID="7102520ed46ba5732ae374650631275c786a0e7dab1acfd1c3ee85a2dc51b901" exitCode=0 Nov 22 08:20:38 crc kubenswrapper[4789]: I1122 08:20:38.726519 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6dc44c56c-bclr4" event={"ID":"7785105e-179b-4b0b-9457-cd513d0b762d","Type":"ContainerDied","Data":"7102520ed46ba5732ae374650631275c786a0e7dab1acfd1c3ee85a2dc51b901"} Nov 22 08:20:39 crc kubenswrapper[4789]: I1122 08:20:39.737522 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6dc44c56c-bclr4" event={"ID":"7785105e-179b-4b0b-9457-cd513d0b762d","Type":"ContainerStarted","Data":"ec54bb2c6c2fd3993a4998368dd12ab72c475858fe45b965cb565d3a6b79af1f"} Nov 22 08:20:39 crc kubenswrapper[4789]: I1122 08:20:39.737913 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6dc44c56c-bclr4" Nov 22 08:20:39 crc kubenswrapper[4789]: I1122 08:20:39.764695 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6dc44c56c-bclr4" podStartSLOduration=3.7646772349999997 podStartE2EDuration="3.764677235s" podCreationTimestamp="2025-11-22 08:20:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:20:39.755274928 +0000 UTC m=+1453.989675211" watchObservedRunningTime="2025-11-22 08:20:39.764677235 +0000 UTC m=+1453.999077508" Nov 22 08:20:47 crc kubenswrapper[4789]: I1122 08:20:47.173363 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6dc44c56c-bclr4" Nov 22 08:20:47 crc kubenswrapper[4789]: I1122 08:20:47.232144 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-568675b579-28s9m"] Nov 22 08:20:47 crc kubenswrapper[4789]: I1122 08:20:47.232658 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-568675b579-28s9m" podUID="93590b86-d27a-43f3-8b03-761b7cbf3c4c" containerName="dnsmasq-dns" containerID="cri-o://80a3f643f8e497908ffd57ba12413d1beda4d63347ba30a88f88378b5a0be4cb" gracePeriod=10 Nov 22 08:20:47 crc kubenswrapper[4789]: I1122 08:20:47.680798 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-568675b579-28s9m" Nov 22 08:20:47 crc kubenswrapper[4789]: I1122 08:20:47.799655 4789 generic.go:334] "Generic (PLEG): container finished" podID="93590b86-d27a-43f3-8b03-761b7cbf3c4c" containerID="80a3f643f8e497908ffd57ba12413d1beda4d63347ba30a88f88378b5a0be4cb" exitCode=0 Nov 22 08:20:47 crc kubenswrapper[4789]: I1122 08:20:47.799704 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-568675b579-28s9m" Nov 22 08:20:47 crc kubenswrapper[4789]: I1122 08:20:47.799723 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-568675b579-28s9m" event={"ID":"93590b86-d27a-43f3-8b03-761b7cbf3c4c","Type":"ContainerDied","Data":"80a3f643f8e497908ffd57ba12413d1beda4d63347ba30a88f88378b5a0be4cb"} Nov 22 08:20:47 crc kubenswrapper[4789]: I1122 08:20:47.800140 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-568675b579-28s9m" event={"ID":"93590b86-d27a-43f3-8b03-761b7cbf3c4c","Type":"ContainerDied","Data":"20ea6dc88ee07fc9c4aa16e905f553c0199da3d7a366f29216a883dc17a35e46"} Nov 22 08:20:47 crc kubenswrapper[4789]: I1122 08:20:47.800306 4789 scope.go:117] "RemoveContainer" containerID="80a3f643f8e497908ffd57ba12413d1beda4d63347ba30a88f88378b5a0be4cb" Nov 22 08:20:47 crc kubenswrapper[4789]: I1122 08:20:47.800504 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/93590b86-d27a-43f3-8b03-761b7cbf3c4c-dns-svc\") pod \"93590b86-d27a-43f3-8b03-761b7cbf3c4c\" (UID: \"93590b86-d27a-43f3-8b03-761b7cbf3c4c\") " Nov 22 08:20:47 crc kubenswrapper[4789]: I1122 08:20:47.800533 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/93590b86-d27a-43f3-8b03-761b7cbf3c4c-config\") pod \"93590b86-d27a-43f3-8b03-761b7cbf3c4c\" (UID: \"93590b86-d27a-43f3-8b03-761b7cbf3c4c\") " Nov 22 08:20:47 crc kubenswrapper[4789]: I1122 08:20:47.800727 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/93590b86-d27a-43f3-8b03-761b7cbf3c4c-ovsdbserver-sb\") pod \"93590b86-d27a-43f3-8b03-761b7cbf3c4c\" (UID: \"93590b86-d27a-43f3-8b03-761b7cbf3c4c\") " Nov 22 08:20:47 crc kubenswrapper[4789]: I1122 08:20:47.800897 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ftjjp\" (UniqueName: \"kubernetes.io/projected/93590b86-d27a-43f3-8b03-761b7cbf3c4c-kube-api-access-ftjjp\") pod \"93590b86-d27a-43f3-8b03-761b7cbf3c4c\" (UID: \"93590b86-d27a-43f3-8b03-761b7cbf3c4c\") " Nov 22 08:20:47 crc kubenswrapper[4789]: I1122 08:20:47.800956 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/93590b86-d27a-43f3-8b03-761b7cbf3c4c-ovsdbserver-nb\") pod \"93590b86-d27a-43f3-8b03-761b7cbf3c4c\" (UID: \"93590b86-d27a-43f3-8b03-761b7cbf3c4c\") " Nov 22 08:20:47 crc kubenswrapper[4789]: I1122 08:20:47.801123 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/93590b86-d27a-43f3-8b03-761b7cbf3c4c-openstack-edpm-ipam\") pod \"93590b86-d27a-43f3-8b03-761b7cbf3c4c\" (UID: \"93590b86-d27a-43f3-8b03-761b7cbf3c4c\") " Nov 22 08:20:47 crc kubenswrapper[4789]: I1122 08:20:47.806021 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93590b86-d27a-43f3-8b03-761b7cbf3c4c-kube-api-access-ftjjp" (OuterVolumeSpecName: "kube-api-access-ftjjp") pod "93590b86-d27a-43f3-8b03-761b7cbf3c4c" (UID: "93590b86-d27a-43f3-8b03-761b7cbf3c4c"). InnerVolumeSpecName "kube-api-access-ftjjp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:20:47 crc kubenswrapper[4789]: I1122 08:20:47.845517 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/93590b86-d27a-43f3-8b03-761b7cbf3c4c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "93590b86-d27a-43f3-8b03-761b7cbf3c4c" (UID: "93590b86-d27a-43f3-8b03-761b7cbf3c4c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:20:47 crc kubenswrapper[4789]: I1122 08:20:47.849329 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/93590b86-d27a-43f3-8b03-761b7cbf3c4c-config" (OuterVolumeSpecName: "config") pod "93590b86-d27a-43f3-8b03-761b7cbf3c4c" (UID: "93590b86-d27a-43f3-8b03-761b7cbf3c4c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:20:47 crc kubenswrapper[4789]: I1122 08:20:47.855265 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/93590b86-d27a-43f3-8b03-761b7cbf3c4c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "93590b86-d27a-43f3-8b03-761b7cbf3c4c" (UID: "93590b86-d27a-43f3-8b03-761b7cbf3c4c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:20:47 crc kubenswrapper[4789]: I1122 08:20:47.858636 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/93590b86-d27a-43f3-8b03-761b7cbf3c4c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "93590b86-d27a-43f3-8b03-761b7cbf3c4c" (UID: "93590b86-d27a-43f3-8b03-761b7cbf3c4c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:20:47 crc kubenswrapper[4789]: I1122 08:20:47.861848 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/93590b86-d27a-43f3-8b03-761b7cbf3c4c-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "93590b86-d27a-43f3-8b03-761b7cbf3c4c" (UID: "93590b86-d27a-43f3-8b03-761b7cbf3c4c"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:20:47 crc kubenswrapper[4789]: I1122 08:20:47.905201 4789 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/93590b86-d27a-43f3-8b03-761b7cbf3c4c-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 22 08:20:47 crc kubenswrapper[4789]: I1122 08:20:47.905255 4789 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/93590b86-d27a-43f3-8b03-761b7cbf3c4c-config\") on node \"crc\" DevicePath \"\"" Nov 22 08:20:47 crc kubenswrapper[4789]: I1122 08:20:47.905270 4789 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/93590b86-d27a-43f3-8b03-761b7cbf3c4c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 22 08:20:47 crc kubenswrapper[4789]: I1122 08:20:47.905283 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ftjjp\" (UniqueName: \"kubernetes.io/projected/93590b86-d27a-43f3-8b03-761b7cbf3c4c-kube-api-access-ftjjp\") on node \"crc\" DevicePath \"\"" Nov 22 08:20:47 crc kubenswrapper[4789]: I1122 08:20:47.905296 4789 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/93590b86-d27a-43f3-8b03-761b7cbf3c4c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 22 08:20:47 crc kubenswrapper[4789]: I1122 08:20:47.905307 4789 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/93590b86-d27a-43f3-8b03-761b7cbf3c4c-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Nov 22 08:20:47 crc kubenswrapper[4789]: I1122 08:20:47.933628 4789 scope.go:117] "RemoveContainer" containerID="396e7e5f01f96600a936c1c13abf483b28356df97c2da01c918f9a5f6f757fbf" Nov 22 08:20:47 crc kubenswrapper[4789]: I1122 08:20:47.951882 4789 scope.go:117] "RemoveContainer" containerID="80a3f643f8e497908ffd57ba12413d1beda4d63347ba30a88f88378b5a0be4cb" Nov 22 08:20:47 crc kubenswrapper[4789]: E1122 08:20:47.952359 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"80a3f643f8e497908ffd57ba12413d1beda4d63347ba30a88f88378b5a0be4cb\": container with ID starting with 80a3f643f8e497908ffd57ba12413d1beda4d63347ba30a88f88378b5a0be4cb not found: ID does not exist" containerID="80a3f643f8e497908ffd57ba12413d1beda4d63347ba30a88f88378b5a0be4cb" Nov 22 08:20:47 crc kubenswrapper[4789]: I1122 08:20:47.952396 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80a3f643f8e497908ffd57ba12413d1beda4d63347ba30a88f88378b5a0be4cb"} err="failed to get container status \"80a3f643f8e497908ffd57ba12413d1beda4d63347ba30a88f88378b5a0be4cb\": rpc error: code = NotFound desc = could not find container \"80a3f643f8e497908ffd57ba12413d1beda4d63347ba30a88f88378b5a0be4cb\": container with ID starting with 80a3f643f8e497908ffd57ba12413d1beda4d63347ba30a88f88378b5a0be4cb not found: ID does not exist" Nov 22 08:20:47 crc kubenswrapper[4789]: I1122 08:20:47.952418 4789 scope.go:117] "RemoveContainer" containerID="396e7e5f01f96600a936c1c13abf483b28356df97c2da01c918f9a5f6f757fbf" Nov 22 08:20:47 crc kubenswrapper[4789]: E1122 08:20:47.952812 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"396e7e5f01f96600a936c1c13abf483b28356df97c2da01c918f9a5f6f757fbf\": container with ID starting with 396e7e5f01f96600a936c1c13abf483b28356df97c2da01c918f9a5f6f757fbf not found: ID does not exist" containerID="396e7e5f01f96600a936c1c13abf483b28356df97c2da01c918f9a5f6f757fbf" Nov 22 08:20:47 crc kubenswrapper[4789]: I1122 08:20:47.952855 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"396e7e5f01f96600a936c1c13abf483b28356df97c2da01c918f9a5f6f757fbf"} err="failed to get container status \"396e7e5f01f96600a936c1c13abf483b28356df97c2da01c918f9a5f6f757fbf\": rpc error: code = NotFound desc = could not find container \"396e7e5f01f96600a936c1c13abf483b28356df97c2da01c918f9a5f6f757fbf\": container with ID starting with 396e7e5f01f96600a936c1c13abf483b28356df97c2da01c918f9a5f6f757fbf not found: ID does not exist" Nov 22 08:20:48 crc kubenswrapper[4789]: I1122 08:20:48.118813 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-568675b579-28s9m"] Nov 22 08:20:48 crc kubenswrapper[4789]: I1122 08:20:48.126399 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-568675b579-28s9m"] Nov 22 08:20:49 crc kubenswrapper[4789]: I1122 08:20:49.975935 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="93590b86-d27a-43f3-8b03-761b7cbf3c4c" path="/var/lib/kubelet/pods/93590b86-d27a-43f3-8b03-761b7cbf3c4c/volumes" Nov 22 08:20:50 crc kubenswrapper[4789]: I1122 08:20:50.685929 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-8697r"] Nov 22 08:20:50 crc kubenswrapper[4789]: E1122 08:20:50.686335 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1899d9cc-6fb5-499f-8a8a-c7fe3b923e96" containerName="init" Nov 22 08:20:50 crc kubenswrapper[4789]: I1122 08:20:50.686355 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="1899d9cc-6fb5-499f-8a8a-c7fe3b923e96" containerName="init" Nov 22 08:20:50 crc kubenswrapper[4789]: E1122 08:20:50.686373 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93590b86-d27a-43f3-8b03-761b7cbf3c4c" containerName="dnsmasq-dns" Nov 22 08:20:50 crc kubenswrapper[4789]: I1122 08:20:50.686381 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="93590b86-d27a-43f3-8b03-761b7cbf3c4c" containerName="dnsmasq-dns" Nov 22 08:20:50 crc kubenswrapper[4789]: E1122 08:20:50.686411 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93590b86-d27a-43f3-8b03-761b7cbf3c4c" containerName="init" Nov 22 08:20:50 crc kubenswrapper[4789]: I1122 08:20:50.686420 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="93590b86-d27a-43f3-8b03-761b7cbf3c4c" containerName="init" Nov 22 08:20:50 crc kubenswrapper[4789]: E1122 08:20:50.686446 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1899d9cc-6fb5-499f-8a8a-c7fe3b923e96" containerName="dnsmasq-dns" Nov 22 08:20:50 crc kubenswrapper[4789]: I1122 08:20:50.686474 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="1899d9cc-6fb5-499f-8a8a-c7fe3b923e96" containerName="dnsmasq-dns" Nov 22 08:20:50 crc kubenswrapper[4789]: I1122 08:20:50.686685 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="93590b86-d27a-43f3-8b03-761b7cbf3c4c" containerName="dnsmasq-dns" Nov 22 08:20:50 crc kubenswrapper[4789]: I1122 08:20:50.686718 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="1899d9cc-6fb5-499f-8a8a-c7fe3b923e96" containerName="dnsmasq-dns" Nov 22 08:20:50 crc kubenswrapper[4789]: I1122 08:20:50.688386 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8697r" Nov 22 08:20:50 crc kubenswrapper[4789]: I1122 08:20:50.705076 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8697r"] Nov 22 08:20:50 crc kubenswrapper[4789]: I1122 08:20:50.752240 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/66210910-8ddb-4910-a188-49e6989efc1c-utilities\") pod \"community-operators-8697r\" (UID: \"66210910-8ddb-4910-a188-49e6989efc1c\") " pod="openshift-marketplace/community-operators-8697r" Nov 22 08:20:50 crc kubenswrapper[4789]: I1122 08:20:50.752455 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/66210910-8ddb-4910-a188-49e6989efc1c-catalog-content\") pod \"community-operators-8697r\" (UID: \"66210910-8ddb-4910-a188-49e6989efc1c\") " pod="openshift-marketplace/community-operators-8697r" Nov 22 08:20:50 crc kubenswrapper[4789]: I1122 08:20:50.753027 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sddw2\" (UniqueName: \"kubernetes.io/projected/66210910-8ddb-4910-a188-49e6989efc1c-kube-api-access-sddw2\") pod \"community-operators-8697r\" (UID: \"66210910-8ddb-4910-a188-49e6989efc1c\") " pod="openshift-marketplace/community-operators-8697r" Nov 22 08:20:50 crc kubenswrapper[4789]: I1122 08:20:50.855216 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/66210910-8ddb-4910-a188-49e6989efc1c-catalog-content\") pod \"community-operators-8697r\" (UID: \"66210910-8ddb-4910-a188-49e6989efc1c\") " pod="openshift-marketplace/community-operators-8697r" Nov 22 08:20:50 crc kubenswrapper[4789]: I1122 08:20:50.855388 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sddw2\" (UniqueName: \"kubernetes.io/projected/66210910-8ddb-4910-a188-49e6989efc1c-kube-api-access-sddw2\") pod \"community-operators-8697r\" (UID: \"66210910-8ddb-4910-a188-49e6989efc1c\") " pod="openshift-marketplace/community-operators-8697r" Nov 22 08:20:50 crc kubenswrapper[4789]: I1122 08:20:50.855436 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/66210910-8ddb-4910-a188-49e6989efc1c-utilities\") pod \"community-operators-8697r\" (UID: \"66210910-8ddb-4910-a188-49e6989efc1c\") " pod="openshift-marketplace/community-operators-8697r" Nov 22 08:20:50 crc kubenswrapper[4789]: I1122 08:20:50.855848 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/66210910-8ddb-4910-a188-49e6989efc1c-catalog-content\") pod \"community-operators-8697r\" (UID: \"66210910-8ddb-4910-a188-49e6989efc1c\") " pod="openshift-marketplace/community-operators-8697r" Nov 22 08:20:50 crc kubenswrapper[4789]: I1122 08:20:50.855889 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/66210910-8ddb-4910-a188-49e6989efc1c-utilities\") pod \"community-operators-8697r\" (UID: \"66210910-8ddb-4910-a188-49e6989efc1c\") " pod="openshift-marketplace/community-operators-8697r" Nov 22 08:20:50 crc kubenswrapper[4789]: I1122 08:20:50.874673 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sddw2\" (UniqueName: \"kubernetes.io/projected/66210910-8ddb-4910-a188-49e6989efc1c-kube-api-access-sddw2\") pod \"community-operators-8697r\" (UID: \"66210910-8ddb-4910-a188-49e6989efc1c\") " pod="openshift-marketplace/community-operators-8697r" Nov 22 08:20:51 crc kubenswrapper[4789]: I1122 08:20:51.005044 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8697r" Nov 22 08:20:51 crc kubenswrapper[4789]: I1122 08:20:51.575956 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8697r"] Nov 22 08:20:51 crc kubenswrapper[4789]: W1122 08:20:51.577358 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod66210910_8ddb_4910_a188_49e6989efc1c.slice/crio-18277eecd612ce18fc8e804a12b343fc022720ce3c6e5c1b093cd4008826746d WatchSource:0}: Error finding container 18277eecd612ce18fc8e804a12b343fc022720ce3c6e5c1b093cd4008826746d: Status 404 returned error can't find the container with id 18277eecd612ce18fc8e804a12b343fc022720ce3c6e5c1b093cd4008826746d Nov 22 08:20:51 crc kubenswrapper[4789]: I1122 08:20:51.839800 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8697r" event={"ID":"66210910-8ddb-4910-a188-49e6989efc1c","Type":"ContainerStarted","Data":"76d5424c3d40a0d7e328cee8597e9fd22844786a4c40323eacca722b6638d892"} Nov 22 08:20:51 crc kubenswrapper[4789]: I1122 08:20:51.840167 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8697r" event={"ID":"66210910-8ddb-4910-a188-49e6989efc1c","Type":"ContainerStarted","Data":"18277eecd612ce18fc8e804a12b343fc022720ce3c6e5c1b093cd4008826746d"} Nov 22 08:20:52 crc kubenswrapper[4789]: I1122 08:20:52.858243 4789 generic.go:334] "Generic (PLEG): container finished" podID="66210910-8ddb-4910-a188-49e6989efc1c" containerID="76d5424c3d40a0d7e328cee8597e9fd22844786a4c40323eacca722b6638d892" exitCode=0 Nov 22 08:20:52 crc kubenswrapper[4789]: I1122 08:20:52.858430 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8697r" event={"ID":"66210910-8ddb-4910-a188-49e6989efc1c","Type":"ContainerDied","Data":"76d5424c3d40a0d7e328cee8597e9fd22844786a4c40323eacca722b6638d892"} Nov 22 08:20:53 crc kubenswrapper[4789]: I1122 08:20:53.869070 4789 generic.go:334] "Generic (PLEG): container finished" podID="66210910-8ddb-4910-a188-49e6989efc1c" containerID="8eb992dc052ba421ac5baa6e94cdef4d373dee460e1b28d39fe4af94a1e37497" exitCode=0 Nov 22 08:20:53 crc kubenswrapper[4789]: I1122 08:20:53.869119 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8697r" event={"ID":"66210910-8ddb-4910-a188-49e6989efc1c","Type":"ContainerDied","Data":"8eb992dc052ba421ac5baa6e94cdef4d373dee460e1b28d39fe4af94a1e37497"} Nov 22 08:20:54 crc kubenswrapper[4789]: I1122 08:20:54.879681 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8697r" event={"ID":"66210910-8ddb-4910-a188-49e6989efc1c","Type":"ContainerStarted","Data":"5ffb225af0d71924394a848395cb7d79eaa110d2041bafe946a667e045e5e228"} Nov 22 08:20:54 crc kubenswrapper[4789]: I1122 08:20:54.900541 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-8697r" podStartSLOduration=2.35782639 podStartE2EDuration="4.900517681s" podCreationTimestamp="2025-11-22 08:20:50 +0000 UTC" firstStartedPulling="2025-11-22 08:20:51.841437815 +0000 UTC m=+1466.075838088" lastFinishedPulling="2025-11-22 08:20:54.384129096 +0000 UTC m=+1468.618529379" observedRunningTime="2025-11-22 08:20:54.894909578 +0000 UTC m=+1469.129309871" watchObservedRunningTime="2025-11-22 08:20:54.900517681 +0000 UTC m=+1469.134917954" Nov 22 08:20:57 crc kubenswrapper[4789]: I1122 08:20:57.559051 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8q4rs"] Nov 22 08:20:57 crc kubenswrapper[4789]: I1122 08:20:57.560514 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8q4rs" Nov 22 08:20:57 crc kubenswrapper[4789]: I1122 08:20:57.564157 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 22 08:20:57 crc kubenswrapper[4789]: I1122 08:20:57.564353 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 22 08:20:57 crc kubenswrapper[4789]: I1122 08:20:57.564565 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 22 08:20:57 crc kubenswrapper[4789]: I1122 08:20:57.564608 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-glmr7" Nov 22 08:20:57 crc kubenswrapper[4789]: I1122 08:20:57.573254 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8q4rs"] Nov 22 08:20:57 crc kubenswrapper[4789]: I1122 08:20:57.679216 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4852655b-677f-4933-b297-f7e78ec4c037-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-8q4rs\" (UID: \"4852655b-677f-4933-b297-f7e78ec4c037\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8q4rs" Nov 22 08:20:57 crc kubenswrapper[4789]: I1122 08:20:57.679315 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4852655b-677f-4933-b297-f7e78ec4c037-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-8q4rs\" (UID: \"4852655b-677f-4933-b297-f7e78ec4c037\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8q4rs" Nov 22 08:20:57 crc kubenswrapper[4789]: I1122 08:20:57.679359 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4852655b-677f-4933-b297-f7e78ec4c037-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-8q4rs\" (UID: \"4852655b-677f-4933-b297-f7e78ec4c037\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8q4rs" Nov 22 08:20:57 crc kubenswrapper[4789]: I1122 08:20:57.679440 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lqgzz\" (UniqueName: \"kubernetes.io/projected/4852655b-677f-4933-b297-f7e78ec4c037-kube-api-access-lqgzz\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-8q4rs\" (UID: \"4852655b-677f-4933-b297-f7e78ec4c037\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8q4rs" Nov 22 08:20:57 crc kubenswrapper[4789]: I1122 08:20:57.781296 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lqgzz\" (UniqueName: \"kubernetes.io/projected/4852655b-677f-4933-b297-f7e78ec4c037-kube-api-access-lqgzz\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-8q4rs\" (UID: \"4852655b-677f-4933-b297-f7e78ec4c037\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8q4rs" Nov 22 08:20:57 crc kubenswrapper[4789]: I1122 08:20:57.781411 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4852655b-677f-4933-b297-f7e78ec4c037-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-8q4rs\" (UID: \"4852655b-677f-4933-b297-f7e78ec4c037\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8q4rs" Nov 22 08:20:57 crc kubenswrapper[4789]: I1122 08:20:57.781454 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4852655b-677f-4933-b297-f7e78ec4c037-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-8q4rs\" (UID: \"4852655b-677f-4933-b297-f7e78ec4c037\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8q4rs" Nov 22 08:20:57 crc kubenswrapper[4789]: I1122 08:20:57.781487 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4852655b-677f-4933-b297-f7e78ec4c037-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-8q4rs\" (UID: \"4852655b-677f-4933-b297-f7e78ec4c037\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8q4rs" Nov 22 08:20:57 crc kubenswrapper[4789]: I1122 08:20:57.787337 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4852655b-677f-4933-b297-f7e78ec4c037-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-8q4rs\" (UID: \"4852655b-677f-4933-b297-f7e78ec4c037\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8q4rs" Nov 22 08:20:57 crc kubenswrapper[4789]: I1122 08:20:57.787435 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4852655b-677f-4933-b297-f7e78ec4c037-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-8q4rs\" (UID: \"4852655b-677f-4933-b297-f7e78ec4c037\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8q4rs" Nov 22 08:20:57 crc kubenswrapper[4789]: I1122 08:20:57.788354 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4852655b-677f-4933-b297-f7e78ec4c037-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-8q4rs\" (UID: \"4852655b-677f-4933-b297-f7e78ec4c037\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8q4rs" Nov 22 08:20:57 crc kubenswrapper[4789]: I1122 08:20:57.796828 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lqgzz\" (UniqueName: \"kubernetes.io/projected/4852655b-677f-4933-b297-f7e78ec4c037-kube-api-access-lqgzz\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-8q4rs\" (UID: \"4852655b-677f-4933-b297-f7e78ec4c037\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8q4rs" Nov 22 08:20:57 crc kubenswrapper[4789]: I1122 08:20:57.878513 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8q4rs" Nov 22 08:20:57 crc kubenswrapper[4789]: I1122 08:20:57.907775 4789 generic.go:334] "Generic (PLEG): container finished" podID="1f834b47-b9a1-4813-bcac-3b5161eceeac" containerID="e6918ce58dc9f9b564d498294a7689318093d05b1c5621c30d7ed281dee0220c" exitCode=0 Nov 22 08:20:57 crc kubenswrapper[4789]: I1122 08:20:57.907855 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"1f834b47-b9a1-4813-bcac-3b5161eceeac","Type":"ContainerDied","Data":"e6918ce58dc9f9b564d498294a7689318093d05b1c5621c30d7ed281dee0220c"} Nov 22 08:20:57 crc kubenswrapper[4789]: I1122 08:20:57.909672 4789 generic.go:334] "Generic (PLEG): container finished" podID="7ab63a4d-2e35-4b93-803b-13a38a43b8d6" containerID="bb525204bedb37c4d25f2a2754f8c9a7fac03f11f8e707bc850991ebc78b8d46" exitCode=0 Nov 22 08:20:57 crc kubenswrapper[4789]: I1122 08:20:57.909718 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"7ab63a4d-2e35-4b93-803b-13a38a43b8d6","Type":"ContainerDied","Data":"bb525204bedb37c4d25f2a2754f8c9a7fac03f11f8e707bc850991ebc78b8d46"} Nov 22 08:20:58 crc kubenswrapper[4789]: W1122 08:20:58.411489 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4852655b_677f_4933_b297_f7e78ec4c037.slice/crio-e3ed6f6867d3509c58270b9191dadc20d9d669d64c9f2bc99a65a2ccfb6c46be WatchSource:0}: Error finding container e3ed6f6867d3509c58270b9191dadc20d9d669d64c9f2bc99a65a2ccfb6c46be: Status 404 returned error can't find the container with id e3ed6f6867d3509c58270b9191dadc20d9d669d64c9f2bc99a65a2ccfb6c46be Nov 22 08:20:58 crc kubenswrapper[4789]: I1122 08:20:58.411862 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8q4rs"] Nov 22 08:20:58 crc kubenswrapper[4789]: I1122 08:20:58.922466 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"1f834b47-b9a1-4813-bcac-3b5161eceeac","Type":"ContainerStarted","Data":"aa4bb5e01ab60986d3d23d1ec88e9dde8947fa82e1e3b1251f6688607a043932"} Nov 22 08:20:58 crc kubenswrapper[4789]: I1122 08:20:58.922784 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Nov 22 08:20:58 crc kubenswrapper[4789]: I1122 08:20:58.925139 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"7ab63a4d-2e35-4b93-803b-13a38a43b8d6","Type":"ContainerStarted","Data":"80c4f828001cf830413b04d741b3fc1eaea02ebab435ccdf84d2730266260bda"} Nov 22 08:20:58 crc kubenswrapper[4789]: I1122 08:20:58.925402 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:20:58 crc kubenswrapper[4789]: I1122 08:20:58.926835 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8q4rs" event={"ID":"4852655b-677f-4933-b297-f7e78ec4c037","Type":"ContainerStarted","Data":"e3ed6f6867d3509c58270b9191dadc20d9d669d64c9f2bc99a65a2ccfb6c46be"} Nov 22 08:20:58 crc kubenswrapper[4789]: I1122 08:20:58.953242 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=36.953222784 podStartE2EDuration="36.953222784s" podCreationTimestamp="2025-11-22 08:20:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:20:58.944602819 +0000 UTC m=+1473.179003102" watchObservedRunningTime="2025-11-22 08:20:58.953222784 +0000 UTC m=+1473.187623057" Nov 22 08:20:58 crc kubenswrapper[4789]: I1122 08:20:58.982347 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=36.982325499 podStartE2EDuration="36.982325499s" podCreationTimestamp="2025-11-22 08:20:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:20:58.971823222 +0000 UTC m=+1473.206223495" watchObservedRunningTime="2025-11-22 08:20:58.982325499 +0000 UTC m=+1473.216725782" Nov 22 08:21:01 crc kubenswrapper[4789]: I1122 08:21:01.006030 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-8697r" Nov 22 08:21:01 crc kubenswrapper[4789]: I1122 08:21:01.006560 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-8697r" Nov 22 08:21:01 crc kubenswrapper[4789]: I1122 08:21:01.072812 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-8697r" Nov 22 08:21:02 crc kubenswrapper[4789]: I1122 08:21:02.021424 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-8697r" Nov 22 08:21:02 crc kubenswrapper[4789]: I1122 08:21:02.073927 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8697r"] Nov 22 08:21:03 crc kubenswrapper[4789]: I1122 08:21:03.968463 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-8697r" podUID="66210910-8ddb-4910-a188-49e6989efc1c" containerName="registry-server" containerID="cri-o://5ffb225af0d71924394a848395cb7d79eaa110d2041bafe946a667e045e5e228" gracePeriod=2 Nov 22 08:21:07 crc kubenswrapper[4789]: I1122 08:21:06.934937 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-8697r_66210910-8ddb-4910-a188-49e6989efc1c/registry-server/0.log" Nov 22 08:21:07 crc kubenswrapper[4789]: I1122 08:21:06.936353 4789 generic.go:334] "Generic (PLEG): container finished" podID="66210910-8ddb-4910-a188-49e6989efc1c" containerID="5ffb225af0d71924394a848395cb7d79eaa110d2041bafe946a667e045e5e228" exitCode=137 Nov 22 08:21:07 crc kubenswrapper[4789]: I1122 08:21:06.936395 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8697r" event={"ID":"66210910-8ddb-4910-a188-49e6989efc1c","Type":"ContainerDied","Data":"5ffb225af0d71924394a848395cb7d79eaa110d2041bafe946a667e045e5e228"} Nov 22 08:21:11 crc kubenswrapper[4789]: E1122 08:21:11.006170 4789 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 5ffb225af0d71924394a848395cb7d79eaa110d2041bafe946a667e045e5e228 is running failed: container process not found" containerID="5ffb225af0d71924394a848395cb7d79eaa110d2041bafe946a667e045e5e228" cmd=["grpc_health_probe","-addr=:50051"] Nov 22 08:21:11 crc kubenswrapper[4789]: E1122 08:21:11.007891 4789 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 5ffb225af0d71924394a848395cb7d79eaa110d2041bafe946a667e045e5e228 is running failed: container process not found" containerID="5ffb225af0d71924394a848395cb7d79eaa110d2041bafe946a667e045e5e228" cmd=["grpc_health_probe","-addr=:50051"] Nov 22 08:21:11 crc kubenswrapper[4789]: E1122 08:21:11.008372 4789 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 5ffb225af0d71924394a848395cb7d79eaa110d2041bafe946a667e045e5e228 is running failed: container process not found" containerID="5ffb225af0d71924394a848395cb7d79eaa110d2041bafe946a667e045e5e228" cmd=["grpc_health_probe","-addr=:50051"] Nov 22 08:21:11 crc kubenswrapper[4789]: E1122 08:21:11.008482 4789 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 5ffb225af0d71924394a848395cb7d79eaa110d2041bafe946a667e045e5e228 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/community-operators-8697r" podUID="66210910-8ddb-4910-a188-49e6989efc1c" containerName="registry-server" Nov 22 08:21:12 crc kubenswrapper[4789]: I1122 08:21:12.254639 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-8697r_66210910-8ddb-4910-a188-49e6989efc1c/registry-server/0.log" Nov 22 08:21:12 crc kubenswrapper[4789]: I1122 08:21:12.255917 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8697r" Nov 22 08:21:12 crc kubenswrapper[4789]: I1122 08:21:12.351978 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/66210910-8ddb-4910-a188-49e6989efc1c-catalog-content\") pod \"66210910-8ddb-4910-a188-49e6989efc1c\" (UID: \"66210910-8ddb-4910-a188-49e6989efc1c\") " Nov 22 08:21:12 crc kubenswrapper[4789]: I1122 08:21:12.352398 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/66210910-8ddb-4910-a188-49e6989efc1c-utilities\") pod \"66210910-8ddb-4910-a188-49e6989efc1c\" (UID: \"66210910-8ddb-4910-a188-49e6989efc1c\") " Nov 22 08:21:12 crc kubenswrapper[4789]: I1122 08:21:12.352485 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sddw2\" (UniqueName: \"kubernetes.io/projected/66210910-8ddb-4910-a188-49e6989efc1c-kube-api-access-sddw2\") pod \"66210910-8ddb-4910-a188-49e6989efc1c\" (UID: \"66210910-8ddb-4910-a188-49e6989efc1c\") " Nov 22 08:21:12 crc kubenswrapper[4789]: I1122 08:21:12.355325 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/66210910-8ddb-4910-a188-49e6989efc1c-utilities" (OuterVolumeSpecName: "utilities") pod "66210910-8ddb-4910-a188-49e6989efc1c" (UID: "66210910-8ddb-4910-a188-49e6989efc1c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:21:12 crc kubenswrapper[4789]: I1122 08:21:12.360349 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66210910-8ddb-4910-a188-49e6989efc1c-kube-api-access-sddw2" (OuterVolumeSpecName: "kube-api-access-sddw2") pod "66210910-8ddb-4910-a188-49e6989efc1c" (UID: "66210910-8ddb-4910-a188-49e6989efc1c"). InnerVolumeSpecName "kube-api-access-sddw2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:21:12 crc kubenswrapper[4789]: I1122 08:21:12.412172 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/66210910-8ddb-4910-a188-49e6989efc1c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "66210910-8ddb-4910-a188-49e6989efc1c" (UID: "66210910-8ddb-4910-a188-49e6989efc1c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:21:12 crc kubenswrapper[4789]: I1122 08:21:12.455832 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/66210910-8ddb-4910-a188-49e6989efc1c-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 08:21:12 crc kubenswrapper[4789]: I1122 08:21:12.456062 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/66210910-8ddb-4910-a188-49e6989efc1c-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 08:21:12 crc kubenswrapper[4789]: I1122 08:21:12.456074 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sddw2\" (UniqueName: \"kubernetes.io/projected/66210910-8ddb-4910-a188-49e6989efc1c-kube-api-access-sddw2\") on node \"crc\" DevicePath \"\"" Nov 22 08:21:12 crc kubenswrapper[4789]: I1122 08:21:12.992288 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8q4rs" event={"ID":"4852655b-677f-4933-b297-f7e78ec4c037","Type":"ContainerStarted","Data":"a029cef305ad2ebf2b3aef186e8ffd588cb9295debda093f4487b22fe6bc9363"} Nov 22 08:21:12 crc kubenswrapper[4789]: I1122 08:21:12.994456 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-8697r_66210910-8ddb-4910-a188-49e6989efc1c/registry-server/0.log" Nov 22 08:21:12 crc kubenswrapper[4789]: I1122 08:21:12.995245 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8697r" event={"ID":"66210910-8ddb-4910-a188-49e6989efc1c","Type":"ContainerDied","Data":"18277eecd612ce18fc8e804a12b343fc022720ce3c6e5c1b093cd4008826746d"} Nov 22 08:21:12 crc kubenswrapper[4789]: I1122 08:21:12.995290 4789 scope.go:117] "RemoveContainer" containerID="5ffb225af0d71924394a848395cb7d79eaa110d2041bafe946a667e045e5e228" Nov 22 08:21:12 crc kubenswrapper[4789]: I1122 08:21:12.995325 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8697r" Nov 22 08:21:13 crc kubenswrapper[4789]: I1122 08:21:13.019270 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8q4rs" podStartSLOduration=2.384990642 podStartE2EDuration="16.019251778s" podCreationTimestamp="2025-11-22 08:20:57 +0000 UTC" firstStartedPulling="2025-11-22 08:20:58.414243052 +0000 UTC m=+1472.648643325" lastFinishedPulling="2025-11-22 08:21:12.048504188 +0000 UTC m=+1486.282904461" observedRunningTime="2025-11-22 08:21:13.013185593 +0000 UTC m=+1487.247585866" watchObservedRunningTime="2025-11-22 08:21:13.019251778 +0000 UTC m=+1487.253652051" Nov 22 08:21:13 crc kubenswrapper[4789]: I1122 08:21:13.027400 4789 scope.go:117] "RemoveContainer" containerID="8eb992dc052ba421ac5baa6e94cdef4d373dee460e1b28d39fe4af94a1e37497" Nov 22 08:21:13 crc kubenswrapper[4789]: I1122 08:21:13.031000 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8697r"] Nov 22 08:21:13 crc kubenswrapper[4789]: I1122 08:21:13.040436 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-8697r"] Nov 22 08:21:13 crc kubenswrapper[4789]: I1122 08:21:13.056441 4789 scope.go:117] "RemoveContainer" containerID="76d5424c3d40a0d7e328cee8597e9fd22844786a4c40323eacca722b6638d892" Nov 22 08:21:13 crc kubenswrapper[4789]: I1122 08:21:13.075596 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Nov 22 08:21:13 crc kubenswrapper[4789]: I1122 08:21:13.300369 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Nov 22 08:21:13 crc kubenswrapper[4789]: I1122 08:21:13.976210 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="66210910-8ddb-4910-a188-49e6989efc1c" path="/var/lib/kubelet/pods/66210910-8ddb-4910-a188-49e6989efc1c/volumes" Nov 22 08:21:24 crc kubenswrapper[4789]: I1122 08:21:24.096236 4789 generic.go:334] "Generic (PLEG): container finished" podID="4852655b-677f-4933-b297-f7e78ec4c037" containerID="a029cef305ad2ebf2b3aef186e8ffd588cb9295debda093f4487b22fe6bc9363" exitCode=0 Nov 22 08:21:24 crc kubenswrapper[4789]: I1122 08:21:24.096316 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8q4rs" event={"ID":"4852655b-677f-4933-b297-f7e78ec4c037","Type":"ContainerDied","Data":"a029cef305ad2ebf2b3aef186e8ffd588cb9295debda093f4487b22fe6bc9363"} Nov 22 08:21:25 crc kubenswrapper[4789]: I1122 08:21:25.480417 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8q4rs" Nov 22 08:21:25 crc kubenswrapper[4789]: I1122 08:21:25.596651 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4852655b-677f-4933-b297-f7e78ec4c037-inventory\") pod \"4852655b-677f-4933-b297-f7e78ec4c037\" (UID: \"4852655b-677f-4933-b297-f7e78ec4c037\") " Nov 22 08:21:25 crc kubenswrapper[4789]: I1122 08:21:25.596735 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lqgzz\" (UniqueName: \"kubernetes.io/projected/4852655b-677f-4933-b297-f7e78ec4c037-kube-api-access-lqgzz\") pod \"4852655b-677f-4933-b297-f7e78ec4c037\" (UID: \"4852655b-677f-4933-b297-f7e78ec4c037\") " Nov 22 08:21:25 crc kubenswrapper[4789]: I1122 08:21:25.596866 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4852655b-677f-4933-b297-f7e78ec4c037-ssh-key\") pod \"4852655b-677f-4933-b297-f7e78ec4c037\" (UID: \"4852655b-677f-4933-b297-f7e78ec4c037\") " Nov 22 08:21:25 crc kubenswrapper[4789]: I1122 08:21:25.597023 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4852655b-677f-4933-b297-f7e78ec4c037-repo-setup-combined-ca-bundle\") pod \"4852655b-677f-4933-b297-f7e78ec4c037\" (UID: \"4852655b-677f-4933-b297-f7e78ec4c037\") " Nov 22 08:21:25 crc kubenswrapper[4789]: I1122 08:21:25.603229 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4852655b-677f-4933-b297-f7e78ec4c037-kube-api-access-lqgzz" (OuterVolumeSpecName: "kube-api-access-lqgzz") pod "4852655b-677f-4933-b297-f7e78ec4c037" (UID: "4852655b-677f-4933-b297-f7e78ec4c037"). InnerVolumeSpecName "kube-api-access-lqgzz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:21:25 crc kubenswrapper[4789]: I1122 08:21:25.603582 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4852655b-677f-4933-b297-f7e78ec4c037-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "4852655b-677f-4933-b297-f7e78ec4c037" (UID: "4852655b-677f-4933-b297-f7e78ec4c037"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:21:25 crc kubenswrapper[4789]: I1122 08:21:25.624534 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4852655b-677f-4933-b297-f7e78ec4c037-inventory" (OuterVolumeSpecName: "inventory") pod "4852655b-677f-4933-b297-f7e78ec4c037" (UID: "4852655b-677f-4933-b297-f7e78ec4c037"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:21:25 crc kubenswrapper[4789]: I1122 08:21:25.625786 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4852655b-677f-4933-b297-f7e78ec4c037-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4852655b-677f-4933-b297-f7e78ec4c037" (UID: "4852655b-677f-4933-b297-f7e78ec4c037"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:21:25 crc kubenswrapper[4789]: I1122 08:21:25.698923 4789 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4852655b-677f-4933-b297-f7e78ec4c037-inventory\") on node \"crc\" DevicePath \"\"" Nov 22 08:21:25 crc kubenswrapper[4789]: I1122 08:21:25.698963 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lqgzz\" (UniqueName: \"kubernetes.io/projected/4852655b-677f-4933-b297-f7e78ec4c037-kube-api-access-lqgzz\") on node \"crc\" DevicePath \"\"" Nov 22 08:21:25 crc kubenswrapper[4789]: I1122 08:21:25.698974 4789 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4852655b-677f-4933-b297-f7e78ec4c037-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 22 08:21:25 crc kubenswrapper[4789]: I1122 08:21:25.698983 4789 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4852655b-677f-4933-b297-f7e78ec4c037-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:21:26 crc kubenswrapper[4789]: I1122 08:21:26.116854 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8q4rs" event={"ID":"4852655b-677f-4933-b297-f7e78ec4c037","Type":"ContainerDied","Data":"e3ed6f6867d3509c58270b9191dadc20d9d669d64c9f2bc99a65a2ccfb6c46be"} Nov 22 08:21:26 crc kubenswrapper[4789]: I1122 08:21:26.116896 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e3ed6f6867d3509c58270b9191dadc20d9d669d64c9f2bc99a65a2ccfb6c46be" Nov 22 08:21:26 crc kubenswrapper[4789]: I1122 08:21:26.116929 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8q4rs" Nov 22 08:21:26 crc kubenswrapper[4789]: I1122 08:21:26.180647 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xqcfq"] Nov 22 08:21:26 crc kubenswrapper[4789]: E1122 08:21:26.181198 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4852655b-677f-4933-b297-f7e78ec4c037" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 22 08:21:26 crc kubenswrapper[4789]: I1122 08:21:26.181224 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="4852655b-677f-4933-b297-f7e78ec4c037" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 22 08:21:26 crc kubenswrapper[4789]: E1122 08:21:26.181249 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66210910-8ddb-4910-a188-49e6989efc1c" containerName="extract-content" Nov 22 08:21:26 crc kubenswrapper[4789]: I1122 08:21:26.181258 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="66210910-8ddb-4910-a188-49e6989efc1c" containerName="extract-content" Nov 22 08:21:26 crc kubenswrapper[4789]: E1122 08:21:26.181293 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66210910-8ddb-4910-a188-49e6989efc1c" containerName="extract-utilities" Nov 22 08:21:26 crc kubenswrapper[4789]: I1122 08:21:26.181304 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="66210910-8ddb-4910-a188-49e6989efc1c" containerName="extract-utilities" Nov 22 08:21:26 crc kubenswrapper[4789]: E1122 08:21:26.181327 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66210910-8ddb-4910-a188-49e6989efc1c" containerName="registry-server" Nov 22 08:21:26 crc kubenswrapper[4789]: I1122 08:21:26.181334 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="66210910-8ddb-4910-a188-49e6989efc1c" containerName="registry-server" Nov 22 08:21:26 crc kubenswrapper[4789]: I1122 08:21:26.181552 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="4852655b-677f-4933-b297-f7e78ec4c037" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 22 08:21:26 crc kubenswrapper[4789]: I1122 08:21:26.181567 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="66210910-8ddb-4910-a188-49e6989efc1c" containerName="registry-server" Nov 22 08:21:26 crc kubenswrapper[4789]: I1122 08:21:26.182306 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xqcfq" Nov 22 08:21:26 crc kubenswrapper[4789]: I1122 08:21:26.184322 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 22 08:21:26 crc kubenswrapper[4789]: I1122 08:21:26.184482 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-glmr7" Nov 22 08:21:26 crc kubenswrapper[4789]: I1122 08:21:26.184659 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 22 08:21:26 crc kubenswrapper[4789]: I1122 08:21:26.184780 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 22 08:21:26 crc kubenswrapper[4789]: I1122 08:21:26.189040 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xqcfq"] Nov 22 08:21:26 crc kubenswrapper[4789]: I1122 08:21:26.310510 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bcf20d51-f782-4c2a-b093-489e9cbd300e-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-xqcfq\" (UID: \"bcf20d51-f782-4c2a-b093-489e9cbd300e\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xqcfq" Nov 22 08:21:26 crc kubenswrapper[4789]: I1122 08:21:26.310872 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bcf20d51-f782-4c2a-b093-489e9cbd300e-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-xqcfq\" (UID: \"bcf20d51-f782-4c2a-b093-489e9cbd300e\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xqcfq" Nov 22 08:21:26 crc kubenswrapper[4789]: I1122 08:21:26.311241 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5xbbw\" (UniqueName: \"kubernetes.io/projected/bcf20d51-f782-4c2a-b093-489e9cbd300e-kube-api-access-5xbbw\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-xqcfq\" (UID: \"bcf20d51-f782-4c2a-b093-489e9cbd300e\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xqcfq" Nov 22 08:21:26 crc kubenswrapper[4789]: I1122 08:21:26.311306 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bcf20d51-f782-4c2a-b093-489e9cbd300e-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-xqcfq\" (UID: \"bcf20d51-f782-4c2a-b093-489e9cbd300e\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xqcfq" Nov 22 08:21:26 crc kubenswrapper[4789]: I1122 08:21:26.413505 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5xbbw\" (UniqueName: \"kubernetes.io/projected/bcf20d51-f782-4c2a-b093-489e9cbd300e-kube-api-access-5xbbw\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-xqcfq\" (UID: \"bcf20d51-f782-4c2a-b093-489e9cbd300e\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xqcfq" Nov 22 08:21:26 crc kubenswrapper[4789]: I1122 08:21:26.413570 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bcf20d51-f782-4c2a-b093-489e9cbd300e-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-xqcfq\" (UID: \"bcf20d51-f782-4c2a-b093-489e9cbd300e\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xqcfq" Nov 22 08:21:26 crc kubenswrapper[4789]: I1122 08:21:26.413654 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bcf20d51-f782-4c2a-b093-489e9cbd300e-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-xqcfq\" (UID: \"bcf20d51-f782-4c2a-b093-489e9cbd300e\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xqcfq" Nov 22 08:21:26 crc kubenswrapper[4789]: I1122 08:21:26.413698 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bcf20d51-f782-4c2a-b093-489e9cbd300e-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-xqcfq\" (UID: \"bcf20d51-f782-4c2a-b093-489e9cbd300e\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xqcfq" Nov 22 08:21:26 crc kubenswrapper[4789]: I1122 08:21:26.418977 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bcf20d51-f782-4c2a-b093-489e9cbd300e-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-xqcfq\" (UID: \"bcf20d51-f782-4c2a-b093-489e9cbd300e\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xqcfq" Nov 22 08:21:26 crc kubenswrapper[4789]: I1122 08:21:26.419136 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bcf20d51-f782-4c2a-b093-489e9cbd300e-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-xqcfq\" (UID: \"bcf20d51-f782-4c2a-b093-489e9cbd300e\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xqcfq" Nov 22 08:21:26 crc kubenswrapper[4789]: I1122 08:21:26.419501 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bcf20d51-f782-4c2a-b093-489e9cbd300e-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-xqcfq\" (UID: \"bcf20d51-f782-4c2a-b093-489e9cbd300e\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xqcfq" Nov 22 08:21:26 crc kubenswrapper[4789]: I1122 08:21:26.431912 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5xbbw\" (UniqueName: \"kubernetes.io/projected/bcf20d51-f782-4c2a-b093-489e9cbd300e-kube-api-access-5xbbw\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-xqcfq\" (UID: \"bcf20d51-f782-4c2a-b093-489e9cbd300e\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xqcfq" Nov 22 08:21:26 crc kubenswrapper[4789]: I1122 08:21:26.503005 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xqcfq" Nov 22 08:21:26 crc kubenswrapper[4789]: I1122 08:21:26.999772 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xqcfq"] Nov 22 08:21:27 crc kubenswrapper[4789]: W1122 08:21:27.001979 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbcf20d51_f782_4c2a_b093_489e9cbd300e.slice/crio-6936f2d6cf973d800c001d0f08aa22feb48d48ee23bef6a25c2d53a5c7abf0c2 WatchSource:0}: Error finding container 6936f2d6cf973d800c001d0f08aa22feb48d48ee23bef6a25c2d53a5c7abf0c2: Status 404 returned error can't find the container with id 6936f2d6cf973d800c001d0f08aa22feb48d48ee23bef6a25c2d53a5c7abf0c2 Nov 22 08:21:27 crc kubenswrapper[4789]: I1122 08:21:27.125303 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xqcfq" event={"ID":"bcf20d51-f782-4c2a-b093-489e9cbd300e","Type":"ContainerStarted","Data":"6936f2d6cf973d800c001d0f08aa22feb48d48ee23bef6a25c2d53a5c7abf0c2"} Nov 22 08:21:27 crc kubenswrapper[4789]: I1122 08:21:27.807407 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 22 08:21:28 crc kubenswrapper[4789]: I1122 08:21:28.134107 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xqcfq" event={"ID":"bcf20d51-f782-4c2a-b093-489e9cbd300e","Type":"ContainerStarted","Data":"1357b63657791ed648b6839cf628c9f025201f14b21c296b033237e330e7ac34"} Nov 22 08:21:28 crc kubenswrapper[4789]: I1122 08:21:28.155684 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xqcfq" podStartSLOduration=1.35540214 podStartE2EDuration="2.155667681s" podCreationTimestamp="2025-11-22 08:21:26 +0000 UTC" firstStartedPulling="2025-11-22 08:21:27.004158169 +0000 UTC m=+1501.238558442" lastFinishedPulling="2025-11-22 08:21:27.80442371 +0000 UTC m=+1502.038823983" observedRunningTime="2025-11-22 08:21:28.148618529 +0000 UTC m=+1502.383018802" watchObservedRunningTime="2025-11-22 08:21:28.155667681 +0000 UTC m=+1502.390067954" Nov 22 08:21:35 crc kubenswrapper[4789]: I1122 08:21:35.372510 4789 patch_prober.go:28] interesting pod/machine-config-daemon-pxvrm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 08:21:35 crc kubenswrapper[4789]: I1122 08:21:35.373133 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 08:22:05 crc kubenswrapper[4789]: I1122 08:22:05.371917 4789 patch_prober.go:28] interesting pod/machine-config-daemon-pxvrm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 08:22:05 crc kubenswrapper[4789]: I1122 08:22:05.372491 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 08:22:24 crc kubenswrapper[4789]: I1122 08:22:24.224640 4789 scope.go:117] "RemoveContainer" containerID="83b2ef9fe1addd84e58855d1a0866dfb7dd425eeeeff2289d91aa552b31f47eb" Nov 22 08:22:24 crc kubenswrapper[4789]: I1122 08:22:24.277433 4789 scope.go:117] "RemoveContainer" containerID="404f9e8585b31f57145fa7f61ae33da7eacd4bcd8fd95b769ea20c94366662cb" Nov 22 08:22:35 crc kubenswrapper[4789]: I1122 08:22:35.371945 4789 patch_prober.go:28] interesting pod/machine-config-daemon-pxvrm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 08:22:35 crc kubenswrapper[4789]: I1122 08:22:35.372684 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 08:22:35 crc kubenswrapper[4789]: I1122 08:22:35.372787 4789 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" Nov 22 08:22:35 crc kubenswrapper[4789]: I1122 08:22:35.373706 4789 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"cd9752d9c3408de592a39b57fc6c7443fb52d47aecefc0deabddc20b20428c47"} pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 22 08:22:35 crc kubenswrapper[4789]: I1122 08:22:35.373832 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" containerID="cri-o://cd9752d9c3408de592a39b57fc6c7443fb52d47aecefc0deabddc20b20428c47" gracePeriod=600 Nov 22 08:22:35 crc kubenswrapper[4789]: E1122 08:22:35.493881 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 08:22:35 crc kubenswrapper[4789]: I1122 08:22:35.739148 4789 generic.go:334] "Generic (PLEG): container finished" podID="58aa8071-7ecc-4692-8789-2db04ec70647" containerID="cd9752d9c3408de592a39b57fc6c7443fb52d47aecefc0deabddc20b20428c47" exitCode=0 Nov 22 08:22:35 crc kubenswrapper[4789]: I1122 08:22:35.739195 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" event={"ID":"58aa8071-7ecc-4692-8789-2db04ec70647","Type":"ContainerDied","Data":"cd9752d9c3408de592a39b57fc6c7443fb52d47aecefc0deabddc20b20428c47"} Nov 22 08:22:35 crc kubenswrapper[4789]: I1122 08:22:35.739232 4789 scope.go:117] "RemoveContainer" containerID="3d1ad86d18575ad837bca3f627df48477a6e890e1d7c69736133f6389315bc07" Nov 22 08:22:35 crc kubenswrapper[4789]: I1122 08:22:35.740092 4789 scope.go:117] "RemoveContainer" containerID="cd9752d9c3408de592a39b57fc6c7443fb52d47aecefc0deabddc20b20428c47" Nov 22 08:22:35 crc kubenswrapper[4789]: E1122 08:22:35.740656 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 08:22:48 crc kubenswrapper[4789]: I1122 08:22:48.965134 4789 scope.go:117] "RemoveContainer" containerID="cd9752d9c3408de592a39b57fc6c7443fb52d47aecefc0deabddc20b20428c47" Nov 22 08:22:48 crc kubenswrapper[4789]: E1122 08:22:48.965897 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 08:23:02 crc kubenswrapper[4789]: I1122 08:23:02.965804 4789 scope.go:117] "RemoveContainer" containerID="cd9752d9c3408de592a39b57fc6c7443fb52d47aecefc0deabddc20b20428c47" Nov 22 08:23:02 crc kubenswrapper[4789]: E1122 08:23:02.966621 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 08:23:13 crc kubenswrapper[4789]: I1122 08:23:13.965969 4789 scope.go:117] "RemoveContainer" containerID="cd9752d9c3408de592a39b57fc6c7443fb52d47aecefc0deabddc20b20428c47" Nov 22 08:23:13 crc kubenswrapper[4789]: E1122 08:23:13.967573 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 08:23:25 crc kubenswrapper[4789]: I1122 08:23:25.965329 4789 scope.go:117] "RemoveContainer" containerID="cd9752d9c3408de592a39b57fc6c7443fb52d47aecefc0deabddc20b20428c47" Nov 22 08:23:25 crc kubenswrapper[4789]: E1122 08:23:25.966131 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 08:23:40 crc kubenswrapper[4789]: I1122 08:23:40.965078 4789 scope.go:117] "RemoveContainer" containerID="cd9752d9c3408de592a39b57fc6c7443fb52d47aecefc0deabddc20b20428c47" Nov 22 08:23:40 crc kubenswrapper[4789]: E1122 08:23:40.965891 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 08:23:41 crc kubenswrapper[4789]: I1122 08:23:41.604235 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-z5h7b"] Nov 22 08:23:41 crc kubenswrapper[4789]: I1122 08:23:41.606537 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-z5h7b" Nov 22 08:23:41 crc kubenswrapper[4789]: I1122 08:23:41.615059 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-z5h7b"] Nov 22 08:23:41 crc kubenswrapper[4789]: I1122 08:23:41.689837 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dd7c6845-e2de-4b77-a7d9-e880c22a6f31-utilities\") pod \"certified-operators-z5h7b\" (UID: \"dd7c6845-e2de-4b77-a7d9-e880c22a6f31\") " pod="openshift-marketplace/certified-operators-z5h7b" Nov 22 08:23:41 crc kubenswrapper[4789]: I1122 08:23:41.689927 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n4hqv\" (UniqueName: \"kubernetes.io/projected/dd7c6845-e2de-4b77-a7d9-e880c22a6f31-kube-api-access-n4hqv\") pod \"certified-operators-z5h7b\" (UID: \"dd7c6845-e2de-4b77-a7d9-e880c22a6f31\") " pod="openshift-marketplace/certified-operators-z5h7b" Nov 22 08:23:41 crc kubenswrapper[4789]: I1122 08:23:41.690019 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dd7c6845-e2de-4b77-a7d9-e880c22a6f31-catalog-content\") pod \"certified-operators-z5h7b\" (UID: \"dd7c6845-e2de-4b77-a7d9-e880c22a6f31\") " pod="openshift-marketplace/certified-operators-z5h7b" Nov 22 08:23:41 crc kubenswrapper[4789]: I1122 08:23:41.791832 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dd7c6845-e2de-4b77-a7d9-e880c22a6f31-utilities\") pod \"certified-operators-z5h7b\" (UID: \"dd7c6845-e2de-4b77-a7d9-e880c22a6f31\") " pod="openshift-marketplace/certified-operators-z5h7b" Nov 22 08:23:41 crc kubenswrapper[4789]: I1122 08:23:41.791915 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n4hqv\" (UniqueName: \"kubernetes.io/projected/dd7c6845-e2de-4b77-a7d9-e880c22a6f31-kube-api-access-n4hqv\") pod \"certified-operators-z5h7b\" (UID: \"dd7c6845-e2de-4b77-a7d9-e880c22a6f31\") " pod="openshift-marketplace/certified-operators-z5h7b" Nov 22 08:23:41 crc kubenswrapper[4789]: I1122 08:23:41.791965 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dd7c6845-e2de-4b77-a7d9-e880c22a6f31-catalog-content\") pod \"certified-operators-z5h7b\" (UID: \"dd7c6845-e2de-4b77-a7d9-e880c22a6f31\") " pod="openshift-marketplace/certified-operators-z5h7b" Nov 22 08:23:41 crc kubenswrapper[4789]: I1122 08:23:41.792416 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dd7c6845-e2de-4b77-a7d9-e880c22a6f31-utilities\") pod \"certified-operators-z5h7b\" (UID: \"dd7c6845-e2de-4b77-a7d9-e880c22a6f31\") " pod="openshift-marketplace/certified-operators-z5h7b" Nov 22 08:23:41 crc kubenswrapper[4789]: I1122 08:23:41.792529 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dd7c6845-e2de-4b77-a7d9-e880c22a6f31-catalog-content\") pod \"certified-operators-z5h7b\" (UID: \"dd7c6845-e2de-4b77-a7d9-e880c22a6f31\") " pod="openshift-marketplace/certified-operators-z5h7b" Nov 22 08:23:41 crc kubenswrapper[4789]: I1122 08:23:41.821160 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n4hqv\" (UniqueName: \"kubernetes.io/projected/dd7c6845-e2de-4b77-a7d9-e880c22a6f31-kube-api-access-n4hqv\") pod \"certified-operators-z5h7b\" (UID: \"dd7c6845-e2de-4b77-a7d9-e880c22a6f31\") " pod="openshift-marketplace/certified-operators-z5h7b" Nov 22 08:23:41 crc kubenswrapper[4789]: I1122 08:23:41.935041 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-z5h7b" Nov 22 08:23:42 crc kubenswrapper[4789]: I1122 08:23:42.427863 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-z5h7b"] Nov 22 08:23:43 crc kubenswrapper[4789]: I1122 08:23:43.325509 4789 generic.go:334] "Generic (PLEG): container finished" podID="dd7c6845-e2de-4b77-a7d9-e880c22a6f31" containerID="86c15041c75b12c17d6dead9c36ba6ac08e0e55ab9fec19575c9e5ff0bccb627" exitCode=0 Nov 22 08:23:43 crc kubenswrapper[4789]: I1122 08:23:43.325594 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z5h7b" event={"ID":"dd7c6845-e2de-4b77-a7d9-e880c22a6f31","Type":"ContainerDied","Data":"86c15041c75b12c17d6dead9c36ba6ac08e0e55ab9fec19575c9e5ff0bccb627"} Nov 22 08:23:43 crc kubenswrapper[4789]: I1122 08:23:43.325845 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z5h7b" event={"ID":"dd7c6845-e2de-4b77-a7d9-e880c22a6f31","Type":"ContainerStarted","Data":"c6bc2dab5d4484621e4ecc4aabe2ec665a4d723b25ed6e32b805c6b61cdb465f"} Nov 22 08:23:44 crc kubenswrapper[4789]: I1122 08:23:44.336904 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z5h7b" event={"ID":"dd7c6845-e2de-4b77-a7d9-e880c22a6f31","Type":"ContainerStarted","Data":"c0a90326df46ea44b7a007246ae09355d64ba80c5ba84af4ebb427de9eb43566"} Nov 22 08:23:45 crc kubenswrapper[4789]: I1122 08:23:45.346286 4789 generic.go:334] "Generic (PLEG): container finished" podID="dd7c6845-e2de-4b77-a7d9-e880c22a6f31" containerID="c0a90326df46ea44b7a007246ae09355d64ba80c5ba84af4ebb427de9eb43566" exitCode=0 Nov 22 08:23:45 crc kubenswrapper[4789]: I1122 08:23:45.346340 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z5h7b" event={"ID":"dd7c6845-e2de-4b77-a7d9-e880c22a6f31","Type":"ContainerDied","Data":"c0a90326df46ea44b7a007246ae09355d64ba80c5ba84af4ebb427de9eb43566"} Nov 22 08:23:46 crc kubenswrapper[4789]: I1122 08:23:46.359570 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z5h7b" event={"ID":"dd7c6845-e2de-4b77-a7d9-e880c22a6f31","Type":"ContainerStarted","Data":"660b1c7fa010e40545cdd99e4a5c3c3257942800f68d68c947716b2bda4a653e"} Nov 22 08:23:46 crc kubenswrapper[4789]: I1122 08:23:46.379893 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-z5h7b" podStartSLOduration=2.9512399350000003 podStartE2EDuration="5.379873014s" podCreationTimestamp="2025-11-22 08:23:41 +0000 UTC" firstStartedPulling="2025-11-22 08:23:43.32781063 +0000 UTC m=+1637.562210903" lastFinishedPulling="2025-11-22 08:23:45.756443709 +0000 UTC m=+1639.990843982" observedRunningTime="2025-11-22 08:23:46.376171963 +0000 UTC m=+1640.610572236" watchObservedRunningTime="2025-11-22 08:23:46.379873014 +0000 UTC m=+1640.614273287" Nov 22 08:23:51 crc kubenswrapper[4789]: I1122 08:23:51.935607 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-z5h7b" Nov 22 08:23:51 crc kubenswrapper[4789]: I1122 08:23:51.936199 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-z5h7b" Nov 22 08:23:51 crc kubenswrapper[4789]: I1122 08:23:51.981889 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-z5h7b" Nov 22 08:23:52 crc kubenswrapper[4789]: I1122 08:23:52.457783 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-z5h7b" Nov 22 08:23:52 crc kubenswrapper[4789]: I1122 08:23:52.507529 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-z5h7b"] Nov 22 08:23:54 crc kubenswrapper[4789]: I1122 08:23:54.427721 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-z5h7b" podUID="dd7c6845-e2de-4b77-a7d9-e880c22a6f31" containerName="registry-server" containerID="cri-o://660b1c7fa010e40545cdd99e4a5c3c3257942800f68d68c947716b2bda4a653e" gracePeriod=2 Nov 22 08:23:56 crc kubenswrapper[4789]: I1122 08:23:54.851910 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-z5h7b" Nov 22 08:23:56 crc kubenswrapper[4789]: I1122 08:23:55.013574 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dd7c6845-e2de-4b77-a7d9-e880c22a6f31-catalog-content\") pod \"dd7c6845-e2de-4b77-a7d9-e880c22a6f31\" (UID: \"dd7c6845-e2de-4b77-a7d9-e880c22a6f31\") " Nov 22 08:23:56 crc kubenswrapper[4789]: I1122 08:23:55.013707 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n4hqv\" (UniqueName: \"kubernetes.io/projected/dd7c6845-e2de-4b77-a7d9-e880c22a6f31-kube-api-access-n4hqv\") pod \"dd7c6845-e2de-4b77-a7d9-e880c22a6f31\" (UID: \"dd7c6845-e2de-4b77-a7d9-e880c22a6f31\") " Nov 22 08:23:56 crc kubenswrapper[4789]: I1122 08:23:55.013797 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dd7c6845-e2de-4b77-a7d9-e880c22a6f31-utilities\") pod \"dd7c6845-e2de-4b77-a7d9-e880c22a6f31\" (UID: \"dd7c6845-e2de-4b77-a7d9-e880c22a6f31\") " Nov 22 08:23:56 crc kubenswrapper[4789]: I1122 08:23:55.014622 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dd7c6845-e2de-4b77-a7d9-e880c22a6f31-utilities" (OuterVolumeSpecName: "utilities") pod "dd7c6845-e2de-4b77-a7d9-e880c22a6f31" (UID: "dd7c6845-e2de-4b77-a7d9-e880c22a6f31"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:23:56 crc kubenswrapper[4789]: I1122 08:23:55.020087 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd7c6845-e2de-4b77-a7d9-e880c22a6f31-kube-api-access-n4hqv" (OuterVolumeSpecName: "kube-api-access-n4hqv") pod "dd7c6845-e2de-4b77-a7d9-e880c22a6f31" (UID: "dd7c6845-e2de-4b77-a7d9-e880c22a6f31"). InnerVolumeSpecName "kube-api-access-n4hqv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:23:56 crc kubenswrapper[4789]: I1122 08:23:55.116482 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n4hqv\" (UniqueName: \"kubernetes.io/projected/dd7c6845-e2de-4b77-a7d9-e880c22a6f31-kube-api-access-n4hqv\") on node \"crc\" DevicePath \"\"" Nov 22 08:23:56 crc kubenswrapper[4789]: I1122 08:23:55.116604 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dd7c6845-e2de-4b77-a7d9-e880c22a6f31-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 08:23:56 crc kubenswrapper[4789]: I1122 08:23:55.438934 4789 generic.go:334] "Generic (PLEG): container finished" podID="dd7c6845-e2de-4b77-a7d9-e880c22a6f31" containerID="660b1c7fa010e40545cdd99e4a5c3c3257942800f68d68c947716b2bda4a653e" exitCode=0 Nov 22 08:23:56 crc kubenswrapper[4789]: I1122 08:23:55.438981 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z5h7b" event={"ID":"dd7c6845-e2de-4b77-a7d9-e880c22a6f31","Type":"ContainerDied","Data":"660b1c7fa010e40545cdd99e4a5c3c3257942800f68d68c947716b2bda4a653e"} Nov 22 08:23:56 crc kubenswrapper[4789]: I1122 08:23:55.439000 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-z5h7b" Nov 22 08:23:56 crc kubenswrapper[4789]: I1122 08:23:55.439015 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z5h7b" event={"ID":"dd7c6845-e2de-4b77-a7d9-e880c22a6f31","Type":"ContainerDied","Data":"c6bc2dab5d4484621e4ecc4aabe2ec665a4d723b25ed6e32b805c6b61cdb465f"} Nov 22 08:23:56 crc kubenswrapper[4789]: I1122 08:23:55.439035 4789 scope.go:117] "RemoveContainer" containerID="660b1c7fa010e40545cdd99e4a5c3c3257942800f68d68c947716b2bda4a653e" Nov 22 08:23:56 crc kubenswrapper[4789]: I1122 08:23:55.457273 4789 scope.go:117] "RemoveContainer" containerID="c0a90326df46ea44b7a007246ae09355d64ba80c5ba84af4ebb427de9eb43566" Nov 22 08:23:56 crc kubenswrapper[4789]: I1122 08:23:55.479674 4789 scope.go:117] "RemoveContainer" containerID="86c15041c75b12c17d6dead9c36ba6ac08e0e55ab9fec19575c9e5ff0bccb627" Nov 22 08:23:56 crc kubenswrapper[4789]: I1122 08:23:55.526720 4789 scope.go:117] "RemoveContainer" containerID="660b1c7fa010e40545cdd99e4a5c3c3257942800f68d68c947716b2bda4a653e" Nov 22 08:23:56 crc kubenswrapper[4789]: E1122 08:23:55.527229 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"660b1c7fa010e40545cdd99e4a5c3c3257942800f68d68c947716b2bda4a653e\": container with ID starting with 660b1c7fa010e40545cdd99e4a5c3c3257942800f68d68c947716b2bda4a653e not found: ID does not exist" containerID="660b1c7fa010e40545cdd99e4a5c3c3257942800f68d68c947716b2bda4a653e" Nov 22 08:23:56 crc kubenswrapper[4789]: I1122 08:23:55.527271 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"660b1c7fa010e40545cdd99e4a5c3c3257942800f68d68c947716b2bda4a653e"} err="failed to get container status \"660b1c7fa010e40545cdd99e4a5c3c3257942800f68d68c947716b2bda4a653e\": rpc error: code = NotFound desc = could not find container \"660b1c7fa010e40545cdd99e4a5c3c3257942800f68d68c947716b2bda4a653e\": container with ID starting with 660b1c7fa010e40545cdd99e4a5c3c3257942800f68d68c947716b2bda4a653e not found: ID does not exist" Nov 22 08:23:56 crc kubenswrapper[4789]: I1122 08:23:55.527349 4789 scope.go:117] "RemoveContainer" containerID="c0a90326df46ea44b7a007246ae09355d64ba80c5ba84af4ebb427de9eb43566" Nov 22 08:23:56 crc kubenswrapper[4789]: E1122 08:23:55.527637 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c0a90326df46ea44b7a007246ae09355d64ba80c5ba84af4ebb427de9eb43566\": container with ID starting with c0a90326df46ea44b7a007246ae09355d64ba80c5ba84af4ebb427de9eb43566 not found: ID does not exist" containerID="c0a90326df46ea44b7a007246ae09355d64ba80c5ba84af4ebb427de9eb43566" Nov 22 08:23:56 crc kubenswrapper[4789]: I1122 08:23:55.527657 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c0a90326df46ea44b7a007246ae09355d64ba80c5ba84af4ebb427de9eb43566"} err="failed to get container status \"c0a90326df46ea44b7a007246ae09355d64ba80c5ba84af4ebb427de9eb43566\": rpc error: code = NotFound desc = could not find container \"c0a90326df46ea44b7a007246ae09355d64ba80c5ba84af4ebb427de9eb43566\": container with ID starting with c0a90326df46ea44b7a007246ae09355d64ba80c5ba84af4ebb427de9eb43566 not found: ID does not exist" Nov 22 08:23:56 crc kubenswrapper[4789]: I1122 08:23:55.527692 4789 scope.go:117] "RemoveContainer" containerID="86c15041c75b12c17d6dead9c36ba6ac08e0e55ab9fec19575c9e5ff0bccb627" Nov 22 08:23:56 crc kubenswrapper[4789]: E1122 08:23:55.528179 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"86c15041c75b12c17d6dead9c36ba6ac08e0e55ab9fec19575c9e5ff0bccb627\": container with ID starting with 86c15041c75b12c17d6dead9c36ba6ac08e0e55ab9fec19575c9e5ff0bccb627 not found: ID does not exist" containerID="86c15041c75b12c17d6dead9c36ba6ac08e0e55ab9fec19575c9e5ff0bccb627" Nov 22 08:23:56 crc kubenswrapper[4789]: I1122 08:23:55.528248 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"86c15041c75b12c17d6dead9c36ba6ac08e0e55ab9fec19575c9e5ff0bccb627"} err="failed to get container status \"86c15041c75b12c17d6dead9c36ba6ac08e0e55ab9fec19575c9e5ff0bccb627\": rpc error: code = NotFound desc = could not find container \"86c15041c75b12c17d6dead9c36ba6ac08e0e55ab9fec19575c9e5ff0bccb627\": container with ID starting with 86c15041c75b12c17d6dead9c36ba6ac08e0e55ab9fec19575c9e5ff0bccb627 not found: ID does not exist" Nov 22 08:23:56 crc kubenswrapper[4789]: I1122 08:23:55.571710 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dd7c6845-e2de-4b77-a7d9-e880c22a6f31-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "dd7c6845-e2de-4b77-a7d9-e880c22a6f31" (UID: "dd7c6845-e2de-4b77-a7d9-e880c22a6f31"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:23:56 crc kubenswrapper[4789]: I1122 08:23:55.624628 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dd7c6845-e2de-4b77-a7d9-e880c22a6f31-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 08:23:56 crc kubenswrapper[4789]: I1122 08:23:55.773576 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-z5h7b"] Nov 22 08:23:56 crc kubenswrapper[4789]: I1122 08:23:55.784346 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-z5h7b"] Nov 22 08:23:56 crc kubenswrapper[4789]: I1122 08:23:55.964969 4789 scope.go:117] "RemoveContainer" containerID="cd9752d9c3408de592a39b57fc6c7443fb52d47aecefc0deabddc20b20428c47" Nov 22 08:23:56 crc kubenswrapper[4789]: E1122 08:23:55.965389 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 08:23:56 crc kubenswrapper[4789]: I1122 08:23:55.977240 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dd7c6845-e2de-4b77-a7d9-e880c22a6f31" path="/var/lib/kubelet/pods/dd7c6845-e2de-4b77-a7d9-e880c22a6f31/volumes" Nov 22 08:24:07 crc kubenswrapper[4789]: I1122 08:24:07.973014 4789 scope.go:117] "RemoveContainer" containerID="cd9752d9c3408de592a39b57fc6c7443fb52d47aecefc0deabddc20b20428c47" Nov 22 08:24:07 crc kubenswrapper[4789]: E1122 08:24:07.973851 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 08:24:21 crc kubenswrapper[4789]: I1122 08:24:21.966123 4789 scope.go:117] "RemoveContainer" containerID="cd9752d9c3408de592a39b57fc6c7443fb52d47aecefc0deabddc20b20428c47" Nov 22 08:24:21 crc kubenswrapper[4789]: E1122 08:24:21.966957 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 08:24:36 crc kubenswrapper[4789]: I1122 08:24:36.965255 4789 scope.go:117] "RemoveContainer" containerID="cd9752d9c3408de592a39b57fc6c7443fb52d47aecefc0deabddc20b20428c47" Nov 22 08:24:36 crc kubenswrapper[4789]: E1122 08:24:36.965946 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 08:24:47 crc kubenswrapper[4789]: I1122 08:24:47.970350 4789 scope.go:117] "RemoveContainer" containerID="cd9752d9c3408de592a39b57fc6c7443fb52d47aecefc0deabddc20b20428c47" Nov 22 08:24:47 crc kubenswrapper[4789]: E1122 08:24:47.971050 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 08:24:59 crc kubenswrapper[4789]: I1122 08:24:59.965290 4789 scope.go:117] "RemoveContainer" containerID="cd9752d9c3408de592a39b57fc6c7443fb52d47aecefc0deabddc20b20428c47" Nov 22 08:24:59 crc kubenswrapper[4789]: E1122 08:24:59.966032 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 08:25:01 crc kubenswrapper[4789]: I1122 08:25:01.039320 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-2sgdb"] Nov 22 08:25:01 crc kubenswrapper[4789]: I1122 08:25:01.049932 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-fdc5-account-create-lqqqs"] Nov 22 08:25:01 crc kubenswrapper[4789]: I1122 08:25:01.059526 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-2sgdb"] Nov 22 08:25:01 crc kubenswrapper[4789]: I1122 08:25:01.067878 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-fdc5-account-create-lqqqs"] Nov 22 08:25:01 crc kubenswrapper[4789]: I1122 08:25:01.990268 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="033d4c4c-73bc-4aa4-a124-aff465a011d3" path="/var/lib/kubelet/pods/033d4c4c-73bc-4aa4-a124-aff465a011d3/volumes" Nov 22 08:25:02 crc kubenswrapper[4789]: I1122 08:25:02.015517 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a5b8f813-91e7-4b56-803d-44167089fdf3" path="/var/lib/kubelet/pods/a5b8f813-91e7-4b56-803d-44167089fdf3/volumes" Nov 22 08:25:02 crc kubenswrapper[4789]: I1122 08:25:02.039035 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-d5fc-account-create-tbdhr"] Nov 22 08:25:02 crc kubenswrapper[4789]: I1122 08:25:02.055241 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-e4c3-account-create-679fs"] Nov 22 08:25:02 crc kubenswrapper[4789]: I1122 08:25:02.065505 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-9bv8q"] Nov 22 08:25:02 crc kubenswrapper[4789]: I1122 08:25:02.074562 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-b867h"] Nov 22 08:25:02 crc kubenswrapper[4789]: I1122 08:25:02.083494 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-d5fc-account-create-tbdhr"] Nov 22 08:25:02 crc kubenswrapper[4789]: I1122 08:25:02.092869 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-9bv8q"] Nov 22 08:25:02 crc kubenswrapper[4789]: I1122 08:25:02.101513 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-b867h"] Nov 22 08:25:02 crc kubenswrapper[4789]: I1122 08:25:02.109796 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-e4c3-account-create-679fs"] Nov 22 08:25:04 crc kubenswrapper[4789]: I1122 08:25:04.018894 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="28b7b6e7-2620-4b4b-b3ab-2da093176b81" path="/var/lib/kubelet/pods/28b7b6e7-2620-4b4b-b3ab-2da093176b81/volumes" Nov 22 08:25:04 crc kubenswrapper[4789]: I1122 08:25:04.026218 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="56293959-d97b-4df2-948e-d8422250945f" path="/var/lib/kubelet/pods/56293959-d97b-4df2-948e-d8422250945f/volumes" Nov 22 08:25:04 crc kubenswrapper[4789]: I1122 08:25:04.033861 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="95dc84f1-08e8-4e12-8181-4a52830326c1" path="/var/lib/kubelet/pods/95dc84f1-08e8-4e12-8181-4a52830326c1/volumes" Nov 22 08:25:04 crc kubenswrapper[4789]: I1122 08:25:04.042502 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dab42b3a-5fd5-43cf-801d-eeb4eef0ac98" path="/var/lib/kubelet/pods/dab42b3a-5fd5-43cf-801d-eeb4eef0ac98/volumes" Nov 22 08:25:14 crc kubenswrapper[4789]: I1122 08:25:14.965097 4789 scope.go:117] "RemoveContainer" containerID="cd9752d9c3408de592a39b57fc6c7443fb52d47aecefc0deabddc20b20428c47" Nov 22 08:25:14 crc kubenswrapper[4789]: E1122 08:25:14.965872 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 08:25:23 crc kubenswrapper[4789]: I1122 08:25:23.152773 4789 generic.go:334] "Generic (PLEG): container finished" podID="bcf20d51-f782-4c2a-b093-489e9cbd300e" containerID="1357b63657791ed648b6839cf628c9f025201f14b21c296b033237e330e7ac34" exitCode=0 Nov 22 08:25:23 crc kubenswrapper[4789]: I1122 08:25:23.152848 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xqcfq" event={"ID":"bcf20d51-f782-4c2a-b093-489e9cbd300e","Type":"ContainerDied","Data":"1357b63657791ed648b6839cf628c9f025201f14b21c296b033237e330e7ac34"} Nov 22 08:25:24 crc kubenswrapper[4789]: I1122 08:25:24.396691 4789 scope.go:117] "RemoveContainer" containerID="a7f9032ef73f529862f33e1e701f05ec568e012f2117e4463fe6532cb25f24b7" Nov 22 08:25:24 crc kubenswrapper[4789]: I1122 08:25:24.446717 4789 scope.go:117] "RemoveContainer" containerID="b84bea1d6e8d82a040a02e9f4861206802c1408e3e270e86181e6fa19997b2a4" Nov 22 08:25:24 crc kubenswrapper[4789]: I1122 08:25:24.477188 4789 scope.go:117] "RemoveContainer" containerID="c2f913b8d0601111137c4e3f14da86191b8ed52f5df71712374e6677e4851fbf" Nov 22 08:25:24 crc kubenswrapper[4789]: I1122 08:25:24.517991 4789 scope.go:117] "RemoveContainer" containerID="c9d9a47ab6213afb223079802b4594dbfeae14df6a8064de3fdfbc3d3211e5be" Nov 22 08:25:24 crc kubenswrapper[4789]: I1122 08:25:24.586820 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xqcfq" Nov 22 08:25:24 crc kubenswrapper[4789]: I1122 08:25:24.630636 4789 scope.go:117] "RemoveContainer" containerID="182aec76e43aa8c76574cf4f35eb8b948a8c3f29ee5bab1c406a601bb189642a" Nov 22 08:25:24 crc kubenswrapper[4789]: I1122 08:25:24.667077 4789 scope.go:117] "RemoveContainer" containerID="f3a3af787ea165ab682819efab7cdd97d21971e55c7d4175c91c2858a3196481" Nov 22 08:25:24 crc kubenswrapper[4789]: I1122 08:25:24.692253 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5xbbw\" (UniqueName: \"kubernetes.io/projected/bcf20d51-f782-4c2a-b093-489e9cbd300e-kube-api-access-5xbbw\") pod \"bcf20d51-f782-4c2a-b093-489e9cbd300e\" (UID: \"bcf20d51-f782-4c2a-b093-489e9cbd300e\") " Nov 22 08:25:24 crc kubenswrapper[4789]: I1122 08:25:24.692396 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bcf20d51-f782-4c2a-b093-489e9cbd300e-inventory\") pod \"bcf20d51-f782-4c2a-b093-489e9cbd300e\" (UID: \"bcf20d51-f782-4c2a-b093-489e9cbd300e\") " Nov 22 08:25:24 crc kubenswrapper[4789]: I1122 08:25:24.692513 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bcf20d51-f782-4c2a-b093-489e9cbd300e-ssh-key\") pod \"bcf20d51-f782-4c2a-b093-489e9cbd300e\" (UID: \"bcf20d51-f782-4c2a-b093-489e9cbd300e\") " Nov 22 08:25:24 crc kubenswrapper[4789]: I1122 08:25:24.692581 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bcf20d51-f782-4c2a-b093-489e9cbd300e-bootstrap-combined-ca-bundle\") pod \"bcf20d51-f782-4c2a-b093-489e9cbd300e\" (UID: \"bcf20d51-f782-4c2a-b093-489e9cbd300e\") " Nov 22 08:25:24 crc kubenswrapper[4789]: I1122 08:25:24.699455 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bcf20d51-f782-4c2a-b093-489e9cbd300e-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "bcf20d51-f782-4c2a-b093-489e9cbd300e" (UID: "bcf20d51-f782-4c2a-b093-489e9cbd300e"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:25:24 crc kubenswrapper[4789]: I1122 08:25:24.699490 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bcf20d51-f782-4c2a-b093-489e9cbd300e-kube-api-access-5xbbw" (OuterVolumeSpecName: "kube-api-access-5xbbw") pod "bcf20d51-f782-4c2a-b093-489e9cbd300e" (UID: "bcf20d51-f782-4c2a-b093-489e9cbd300e"). InnerVolumeSpecName "kube-api-access-5xbbw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:25:24 crc kubenswrapper[4789]: I1122 08:25:24.723845 4789 scope.go:117] "RemoveContainer" containerID="b30099df66f7f5c0ada8e51fea6fd51be78fed2a973b15f47861296095d47eb8" Nov 22 08:25:24 crc kubenswrapper[4789]: I1122 08:25:24.738659 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bcf20d51-f782-4c2a-b093-489e9cbd300e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "bcf20d51-f782-4c2a-b093-489e9cbd300e" (UID: "bcf20d51-f782-4c2a-b093-489e9cbd300e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:25:24 crc kubenswrapper[4789]: I1122 08:25:24.740921 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bcf20d51-f782-4c2a-b093-489e9cbd300e-inventory" (OuterVolumeSpecName: "inventory") pod "bcf20d51-f782-4c2a-b093-489e9cbd300e" (UID: "bcf20d51-f782-4c2a-b093-489e9cbd300e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:25:24 crc kubenswrapper[4789]: I1122 08:25:24.796348 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5xbbw\" (UniqueName: \"kubernetes.io/projected/bcf20d51-f782-4c2a-b093-489e9cbd300e-kube-api-access-5xbbw\") on node \"crc\" DevicePath \"\"" Nov 22 08:25:24 crc kubenswrapper[4789]: I1122 08:25:24.796408 4789 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bcf20d51-f782-4c2a-b093-489e9cbd300e-inventory\") on node \"crc\" DevicePath \"\"" Nov 22 08:25:24 crc kubenswrapper[4789]: I1122 08:25:24.796422 4789 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bcf20d51-f782-4c2a-b093-489e9cbd300e-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 22 08:25:24 crc kubenswrapper[4789]: I1122 08:25:24.796433 4789 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bcf20d51-f782-4c2a-b093-489e9cbd300e-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:25:25 crc kubenswrapper[4789]: I1122 08:25:25.197815 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xqcfq" event={"ID":"bcf20d51-f782-4c2a-b093-489e9cbd300e","Type":"ContainerDied","Data":"6936f2d6cf973d800c001d0f08aa22feb48d48ee23bef6a25c2d53a5c7abf0c2"} Nov 22 08:25:25 crc kubenswrapper[4789]: I1122 08:25:25.198398 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6936f2d6cf973d800c001d0f08aa22feb48d48ee23bef6a25c2d53a5c7abf0c2" Nov 22 08:25:25 crc kubenswrapper[4789]: I1122 08:25:25.198035 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xqcfq" Nov 22 08:25:25 crc kubenswrapper[4789]: I1122 08:25:25.253311 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qp94q"] Nov 22 08:25:25 crc kubenswrapper[4789]: E1122 08:25:25.253725 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd7c6845-e2de-4b77-a7d9-e880c22a6f31" containerName="extract-utilities" Nov 22 08:25:25 crc kubenswrapper[4789]: I1122 08:25:25.253747 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd7c6845-e2de-4b77-a7d9-e880c22a6f31" containerName="extract-utilities" Nov 22 08:25:25 crc kubenswrapper[4789]: E1122 08:25:25.253787 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd7c6845-e2de-4b77-a7d9-e880c22a6f31" containerName="extract-content" Nov 22 08:25:25 crc kubenswrapper[4789]: I1122 08:25:25.253794 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd7c6845-e2de-4b77-a7d9-e880c22a6f31" containerName="extract-content" Nov 22 08:25:25 crc kubenswrapper[4789]: E1122 08:25:25.253813 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd7c6845-e2de-4b77-a7d9-e880c22a6f31" containerName="registry-server" Nov 22 08:25:25 crc kubenswrapper[4789]: I1122 08:25:25.253821 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd7c6845-e2de-4b77-a7d9-e880c22a6f31" containerName="registry-server" Nov 22 08:25:25 crc kubenswrapper[4789]: E1122 08:25:25.253851 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bcf20d51-f782-4c2a-b093-489e9cbd300e" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 22 08:25:25 crc kubenswrapper[4789]: I1122 08:25:25.253860 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="bcf20d51-f782-4c2a-b093-489e9cbd300e" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 22 08:25:25 crc kubenswrapper[4789]: I1122 08:25:25.254040 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="bcf20d51-f782-4c2a-b093-489e9cbd300e" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 22 08:25:25 crc kubenswrapper[4789]: I1122 08:25:25.254072 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd7c6845-e2de-4b77-a7d9-e880c22a6f31" containerName="registry-server" Nov 22 08:25:25 crc kubenswrapper[4789]: I1122 08:25:25.254795 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qp94q" Nov 22 08:25:25 crc kubenswrapper[4789]: I1122 08:25:25.257005 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-glmr7" Nov 22 08:25:25 crc kubenswrapper[4789]: I1122 08:25:25.257315 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 22 08:25:25 crc kubenswrapper[4789]: I1122 08:25:25.263297 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 22 08:25:25 crc kubenswrapper[4789]: I1122 08:25:25.263702 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 22 08:25:25 crc kubenswrapper[4789]: I1122 08:25:25.268147 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qp94q"] Nov 22 08:25:25 crc kubenswrapper[4789]: I1122 08:25:25.407531 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/424c42d2-3148-4308-91a1-9935d0ab2441-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-qp94q\" (UID: \"424c42d2-3148-4308-91a1-9935d0ab2441\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qp94q" Nov 22 08:25:25 crc kubenswrapper[4789]: I1122 08:25:25.407591 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/424c42d2-3148-4308-91a1-9935d0ab2441-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-qp94q\" (UID: \"424c42d2-3148-4308-91a1-9935d0ab2441\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qp94q" Nov 22 08:25:25 crc kubenswrapper[4789]: I1122 08:25:25.407667 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8mcfp\" (UniqueName: \"kubernetes.io/projected/424c42d2-3148-4308-91a1-9935d0ab2441-kube-api-access-8mcfp\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-qp94q\" (UID: \"424c42d2-3148-4308-91a1-9935d0ab2441\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qp94q" Nov 22 08:25:25 crc kubenswrapper[4789]: I1122 08:25:25.508735 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8mcfp\" (UniqueName: \"kubernetes.io/projected/424c42d2-3148-4308-91a1-9935d0ab2441-kube-api-access-8mcfp\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-qp94q\" (UID: \"424c42d2-3148-4308-91a1-9935d0ab2441\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qp94q" Nov 22 08:25:25 crc kubenswrapper[4789]: I1122 08:25:25.508889 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/424c42d2-3148-4308-91a1-9935d0ab2441-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-qp94q\" (UID: \"424c42d2-3148-4308-91a1-9935d0ab2441\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qp94q" Nov 22 08:25:25 crc kubenswrapper[4789]: I1122 08:25:25.508921 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/424c42d2-3148-4308-91a1-9935d0ab2441-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-qp94q\" (UID: \"424c42d2-3148-4308-91a1-9935d0ab2441\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qp94q" Nov 22 08:25:25 crc kubenswrapper[4789]: I1122 08:25:25.513466 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/424c42d2-3148-4308-91a1-9935d0ab2441-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-qp94q\" (UID: \"424c42d2-3148-4308-91a1-9935d0ab2441\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qp94q" Nov 22 08:25:25 crc kubenswrapper[4789]: I1122 08:25:25.513789 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/424c42d2-3148-4308-91a1-9935d0ab2441-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-qp94q\" (UID: \"424c42d2-3148-4308-91a1-9935d0ab2441\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qp94q" Nov 22 08:25:25 crc kubenswrapper[4789]: I1122 08:25:25.531484 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8mcfp\" (UniqueName: \"kubernetes.io/projected/424c42d2-3148-4308-91a1-9935d0ab2441-kube-api-access-8mcfp\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-qp94q\" (UID: \"424c42d2-3148-4308-91a1-9935d0ab2441\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qp94q" Nov 22 08:25:25 crc kubenswrapper[4789]: I1122 08:25:25.579403 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qp94q" Nov 22 08:25:26 crc kubenswrapper[4789]: I1122 08:25:26.093187 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qp94q"] Nov 22 08:25:26 crc kubenswrapper[4789]: I1122 08:25:26.102057 4789 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 22 08:25:26 crc kubenswrapper[4789]: I1122 08:25:26.206765 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qp94q" event={"ID":"424c42d2-3148-4308-91a1-9935d0ab2441","Type":"ContainerStarted","Data":"f4e1ada55e6830382a42135896b5eba183c5e5ce2b5c1f66eb40c350eb782eca"} Nov 22 08:25:26 crc kubenswrapper[4789]: I1122 08:25:26.965419 4789 scope.go:117] "RemoveContainer" containerID="cd9752d9c3408de592a39b57fc6c7443fb52d47aecefc0deabddc20b20428c47" Nov 22 08:25:26 crc kubenswrapper[4789]: E1122 08:25:26.967008 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 08:25:27 crc kubenswrapper[4789]: I1122 08:25:27.219035 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qp94q" event={"ID":"424c42d2-3148-4308-91a1-9935d0ab2441","Type":"ContainerStarted","Data":"a60cf380354af9c3e52384cb95aab1c1637904d531e3dc1cbe439dd62de69622"} Nov 22 08:25:27 crc kubenswrapper[4789]: I1122 08:25:27.247518 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qp94q" podStartSLOduration=1.765515389 podStartE2EDuration="2.247487293s" podCreationTimestamp="2025-11-22 08:25:25 +0000 UTC" firstStartedPulling="2025-11-22 08:25:26.101691364 +0000 UTC m=+1740.336091637" lastFinishedPulling="2025-11-22 08:25:26.583663268 +0000 UTC m=+1740.818063541" observedRunningTime="2025-11-22 08:25:27.23566131 +0000 UTC m=+1741.470061603" watchObservedRunningTime="2025-11-22 08:25:27.247487293 +0000 UTC m=+1741.481887586" Nov 22 08:25:38 crc kubenswrapper[4789]: I1122 08:25:38.964693 4789 scope.go:117] "RemoveContainer" containerID="cd9752d9c3408de592a39b57fc6c7443fb52d47aecefc0deabddc20b20428c47" Nov 22 08:25:38 crc kubenswrapper[4789]: E1122 08:25:38.965409 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 08:25:43 crc kubenswrapper[4789]: I1122 08:25:43.039397 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-64ckc"] Nov 22 08:25:43 crc kubenswrapper[4789]: I1122 08:25:43.048139 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-nsptz"] Nov 22 08:25:43 crc kubenswrapper[4789]: I1122 08:25:43.058857 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-rnxhf"] Nov 22 08:25:43 crc kubenswrapper[4789]: I1122 08:25:43.071141 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-5a05-account-create-dcglb"] Nov 22 08:25:43 crc kubenswrapper[4789]: I1122 08:25:43.079100 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-411a-account-create-dh7ln"] Nov 22 08:25:43 crc kubenswrapper[4789]: I1122 08:25:43.087237 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-64ckc"] Nov 22 08:25:43 crc kubenswrapper[4789]: I1122 08:25:43.094604 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-xhfrw"] Nov 22 08:25:43 crc kubenswrapper[4789]: I1122 08:25:43.101539 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-2375-account-create-chvgz"] Nov 22 08:25:43 crc kubenswrapper[4789]: I1122 08:25:43.107966 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-nsptz"] Nov 22 08:25:43 crc kubenswrapper[4789]: I1122 08:25:43.114801 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-2375-account-create-chvgz"] Nov 22 08:25:43 crc kubenswrapper[4789]: I1122 08:25:43.122956 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-xhfrw"] Nov 22 08:25:43 crc kubenswrapper[4789]: I1122 08:25:43.130470 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-rnxhf"] Nov 22 08:25:43 crc kubenswrapper[4789]: I1122 08:25:43.136905 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-411a-account-create-dh7ln"] Nov 22 08:25:43 crc kubenswrapper[4789]: I1122 08:25:43.144301 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-5a05-account-create-dcglb"] Nov 22 08:25:43 crc kubenswrapper[4789]: I1122 08:25:43.976775 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c9e2b87-ef93-4176-9362-046e849033a5" path="/var/lib/kubelet/pods/0c9e2b87-ef93-4176-9362-046e849033a5/volumes" Nov 22 08:25:43 crc kubenswrapper[4789]: I1122 08:25:43.978294 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0d6e8982-c10f-441a-be60-34888bc0acf2" path="/var/lib/kubelet/pods/0d6e8982-c10f-441a-be60-34888bc0acf2/volumes" Nov 22 08:25:43 crc kubenswrapper[4789]: I1122 08:25:43.978851 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="201449e6-debb-4b67-80c6-524b309dac1f" path="/var/lib/kubelet/pods/201449e6-debb-4b67-80c6-524b309dac1f/volumes" Nov 22 08:25:43 crc kubenswrapper[4789]: I1122 08:25:43.979364 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="41a9281e-b846-4bd9-9c3a-085fe7d8e6f2" path="/var/lib/kubelet/pods/41a9281e-b846-4bd9-9c3a-085fe7d8e6f2/volumes" Nov 22 08:25:43 crc kubenswrapper[4789]: I1122 08:25:43.981381 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="56bd722c-eb14-4063-9ddb-9c010f461238" path="/var/lib/kubelet/pods/56bd722c-eb14-4063-9ddb-9c010f461238/volumes" Nov 22 08:25:43 crc kubenswrapper[4789]: I1122 08:25:43.981940 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="afee6663-d80e-41bc-af23-7d64a23d4b03" path="/var/lib/kubelet/pods/afee6663-d80e-41bc-af23-7d64a23d4b03/volumes" Nov 22 08:25:43 crc kubenswrapper[4789]: I1122 08:25:43.983224 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b56702e5-c634-4162-9aab-75c723993be4" path="/var/lib/kubelet/pods/b56702e5-c634-4162-9aab-75c723993be4/volumes" Nov 22 08:25:48 crc kubenswrapper[4789]: I1122 08:25:48.042683 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-tx8g5"] Nov 22 08:25:48 crc kubenswrapper[4789]: I1122 08:25:48.055157 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-tx8g5"] Nov 22 08:25:49 crc kubenswrapper[4789]: I1122 08:25:49.977667 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c80081f-1494-4fd4-ba90-da18af310075" path="/var/lib/kubelet/pods/4c80081f-1494-4fd4-ba90-da18af310075/volumes" Nov 22 08:25:50 crc kubenswrapper[4789]: I1122 08:25:50.966000 4789 scope.go:117] "RemoveContainer" containerID="cd9752d9c3408de592a39b57fc6c7443fb52d47aecefc0deabddc20b20428c47" Nov 22 08:25:50 crc kubenswrapper[4789]: E1122 08:25:50.966229 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 08:26:04 crc kubenswrapper[4789]: I1122 08:26:04.965371 4789 scope.go:117] "RemoveContainer" containerID="cd9752d9c3408de592a39b57fc6c7443fb52d47aecefc0deabddc20b20428c47" Nov 22 08:26:04 crc kubenswrapper[4789]: E1122 08:26:04.966190 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 08:26:17 crc kubenswrapper[4789]: I1122 08:26:17.971851 4789 scope.go:117] "RemoveContainer" containerID="cd9752d9c3408de592a39b57fc6c7443fb52d47aecefc0deabddc20b20428c47" Nov 22 08:26:17 crc kubenswrapper[4789]: E1122 08:26:17.973314 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 08:26:24 crc kubenswrapper[4789]: I1122 08:26:24.928084 4789 scope.go:117] "RemoveContainer" containerID="5c5f0e394886a566012dcde583f9244b73eb7e8f80bc29c6ec5ea672207f15bc" Nov 22 08:26:24 crc kubenswrapper[4789]: I1122 08:26:24.956597 4789 scope.go:117] "RemoveContainer" containerID="c7ec533ad1968148cca44a814c0fb8cecc5ba6f1805ce8d5609fd0de736207a6" Nov 22 08:26:24 crc kubenswrapper[4789]: I1122 08:26:24.990589 4789 scope.go:117] "RemoveContainer" containerID="1cb148dd74abf9675a681dd2ec67770bd47fe31a37cdf5da5b1fc713bb4ffebc" Nov 22 08:26:25 crc kubenswrapper[4789]: I1122 08:26:25.046237 4789 scope.go:117] "RemoveContainer" containerID="165aaecc3c7daba27ab8232bc51f5d633e1f9d811f9fe7aab330b0a5a7ed65a2" Nov 22 08:26:25 crc kubenswrapper[4789]: I1122 08:26:25.078468 4789 scope.go:117] "RemoveContainer" containerID="578155d67afcd1f4a2024c85e97167323cd513d97c07c4883d838855bb132d55" Nov 22 08:26:25 crc kubenswrapper[4789]: I1122 08:26:25.126228 4789 scope.go:117] "RemoveContainer" containerID="aac635889d3cbe5eeec4bde45f41b0bd395e0ca68c5bb3cff673bbf27978bee3" Nov 22 08:26:25 crc kubenswrapper[4789]: I1122 08:26:25.166982 4789 scope.go:117] "RemoveContainer" containerID="93ffeb8fcbfbb327e457f3753e9f8e777b7a951794760ed8c26507beb61b1a70" Nov 22 08:26:25 crc kubenswrapper[4789]: I1122 08:26:25.210212 4789 scope.go:117] "RemoveContainer" containerID="a5976b507eba98c7547d2fd79706d4b04159ef43f9be684cc22656f65564058a" Nov 22 08:26:25 crc kubenswrapper[4789]: I1122 08:26:25.229242 4789 scope.go:117] "RemoveContainer" containerID="b1483ede4e565a0a93afedd8132895003b4f2a08d1541b3d3df1683ab24a7407" Nov 22 08:26:25 crc kubenswrapper[4789]: I1122 08:26:25.252117 4789 scope.go:117] "RemoveContainer" containerID="1dfe678ead9e68e4225b9febfd4b0a95e66b0d0f04f4d2e710036df1d295f976" Nov 22 08:26:30 crc kubenswrapper[4789]: I1122 08:26:30.965000 4789 scope.go:117] "RemoveContainer" containerID="cd9752d9c3408de592a39b57fc6c7443fb52d47aecefc0deabddc20b20428c47" Nov 22 08:26:30 crc kubenswrapper[4789]: E1122 08:26:30.965678 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 08:26:37 crc kubenswrapper[4789]: I1122 08:26:37.036147 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-2ss46"] Nov 22 08:26:37 crc kubenswrapper[4789]: I1122 08:26:37.046208 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-rnzhp"] Nov 22 08:26:37 crc kubenswrapper[4789]: I1122 08:26:37.053272 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-rnzhp"] Nov 22 08:26:37 crc kubenswrapper[4789]: I1122 08:26:37.060155 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-2ss46"] Nov 22 08:26:37 crc kubenswrapper[4789]: I1122 08:26:37.976010 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1f1fc498-45f3-4bec-a4a5-b9722d2445ea" path="/var/lib/kubelet/pods/1f1fc498-45f3-4bec-a4a5-b9722d2445ea/volumes" Nov 22 08:26:37 crc kubenswrapper[4789]: I1122 08:26:37.977873 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2eb187f8-1892-4d72-b442-43592e75d489" path="/var/lib/kubelet/pods/2eb187f8-1892-4d72-b442-43592e75d489/volumes" Nov 22 08:26:43 crc kubenswrapper[4789]: I1122 08:26:43.965998 4789 scope.go:117] "RemoveContainer" containerID="cd9752d9c3408de592a39b57fc6c7443fb52d47aecefc0deabddc20b20428c47" Nov 22 08:26:43 crc kubenswrapper[4789]: E1122 08:26:43.966854 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 08:26:46 crc kubenswrapper[4789]: I1122 08:26:46.025879 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-r6nst"] Nov 22 08:26:46 crc kubenswrapper[4789]: I1122 08:26:46.034007 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-r6nst"] Nov 22 08:26:47 crc kubenswrapper[4789]: I1122 08:26:47.037723 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-n4s7g"] Nov 22 08:26:47 crc kubenswrapper[4789]: I1122 08:26:47.045195 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-n4s7g"] Nov 22 08:26:47 crc kubenswrapper[4789]: I1122 08:26:47.975485 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="733e94b9-65c2-406e-9ebe-244585e4e15d" path="/var/lib/kubelet/pods/733e94b9-65c2-406e-9ebe-244585e4e15d/volumes" Nov 22 08:26:47 crc kubenswrapper[4789]: I1122 08:26:47.977880 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f3ee5cf5-1275-448f-8713-e2b5a0d24439" path="/var/lib/kubelet/pods/f3ee5cf5-1275-448f-8713-e2b5a0d24439/volumes" Nov 22 08:26:49 crc kubenswrapper[4789]: I1122 08:26:49.026828 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-lg59m"] Nov 22 08:26:49 crc kubenswrapper[4789]: I1122 08:26:49.033350 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-lg59m"] Nov 22 08:26:49 crc kubenswrapper[4789]: I1122 08:26:49.975386 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6bf8aa3e-a7e7-4f0c-a542-d668a5623707" path="/var/lib/kubelet/pods/6bf8aa3e-a7e7-4f0c-a542-d668a5623707/volumes" Nov 22 08:26:56 crc kubenswrapper[4789]: I1122 08:26:56.965040 4789 scope.go:117] "RemoveContainer" containerID="cd9752d9c3408de592a39b57fc6c7443fb52d47aecefc0deabddc20b20428c47" Nov 22 08:26:56 crc kubenswrapper[4789]: E1122 08:26:56.966091 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 08:26:58 crc kubenswrapper[4789]: I1122 08:26:58.976595 4789 generic.go:334] "Generic (PLEG): container finished" podID="424c42d2-3148-4308-91a1-9935d0ab2441" containerID="a60cf380354af9c3e52384cb95aab1c1637904d531e3dc1cbe439dd62de69622" exitCode=0 Nov 22 08:26:58 crc kubenswrapper[4789]: I1122 08:26:58.976644 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qp94q" event={"ID":"424c42d2-3148-4308-91a1-9935d0ab2441","Type":"ContainerDied","Data":"a60cf380354af9c3e52384cb95aab1c1637904d531e3dc1cbe439dd62de69622"} Nov 22 08:27:00 crc kubenswrapper[4789]: I1122 08:27:00.505298 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qp94q" Nov 22 08:27:00 crc kubenswrapper[4789]: I1122 08:27:00.602297 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/424c42d2-3148-4308-91a1-9935d0ab2441-ssh-key\") pod \"424c42d2-3148-4308-91a1-9935d0ab2441\" (UID: \"424c42d2-3148-4308-91a1-9935d0ab2441\") " Nov 22 08:27:00 crc kubenswrapper[4789]: I1122 08:27:00.602424 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/424c42d2-3148-4308-91a1-9935d0ab2441-inventory\") pod \"424c42d2-3148-4308-91a1-9935d0ab2441\" (UID: \"424c42d2-3148-4308-91a1-9935d0ab2441\") " Nov 22 08:27:00 crc kubenswrapper[4789]: I1122 08:27:00.602534 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8mcfp\" (UniqueName: \"kubernetes.io/projected/424c42d2-3148-4308-91a1-9935d0ab2441-kube-api-access-8mcfp\") pod \"424c42d2-3148-4308-91a1-9935d0ab2441\" (UID: \"424c42d2-3148-4308-91a1-9935d0ab2441\") " Nov 22 08:27:00 crc kubenswrapper[4789]: I1122 08:27:00.614638 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/424c42d2-3148-4308-91a1-9935d0ab2441-kube-api-access-8mcfp" (OuterVolumeSpecName: "kube-api-access-8mcfp") pod "424c42d2-3148-4308-91a1-9935d0ab2441" (UID: "424c42d2-3148-4308-91a1-9935d0ab2441"). InnerVolumeSpecName "kube-api-access-8mcfp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:27:00 crc kubenswrapper[4789]: I1122 08:27:00.631821 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/424c42d2-3148-4308-91a1-9935d0ab2441-inventory" (OuterVolumeSpecName: "inventory") pod "424c42d2-3148-4308-91a1-9935d0ab2441" (UID: "424c42d2-3148-4308-91a1-9935d0ab2441"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:27:00 crc kubenswrapper[4789]: I1122 08:27:00.632481 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/424c42d2-3148-4308-91a1-9935d0ab2441-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "424c42d2-3148-4308-91a1-9935d0ab2441" (UID: "424c42d2-3148-4308-91a1-9935d0ab2441"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:27:00 crc kubenswrapper[4789]: I1122 08:27:00.705037 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8mcfp\" (UniqueName: \"kubernetes.io/projected/424c42d2-3148-4308-91a1-9935d0ab2441-kube-api-access-8mcfp\") on node \"crc\" DevicePath \"\"" Nov 22 08:27:00 crc kubenswrapper[4789]: I1122 08:27:00.705284 4789 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/424c42d2-3148-4308-91a1-9935d0ab2441-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 22 08:27:00 crc kubenswrapper[4789]: I1122 08:27:00.705341 4789 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/424c42d2-3148-4308-91a1-9935d0ab2441-inventory\") on node \"crc\" DevicePath \"\"" Nov 22 08:27:00 crc kubenswrapper[4789]: I1122 08:27:00.992629 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qp94q" event={"ID":"424c42d2-3148-4308-91a1-9935d0ab2441","Type":"ContainerDied","Data":"f4e1ada55e6830382a42135896b5eba183c5e5ce2b5c1f66eb40c350eb782eca"} Nov 22 08:27:00 crc kubenswrapper[4789]: I1122 08:27:00.992998 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f4e1ada55e6830382a42135896b5eba183c5e5ce2b5c1f66eb40c350eb782eca" Nov 22 08:27:00 crc kubenswrapper[4789]: I1122 08:27:00.992691 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qp94q" Nov 22 08:27:01 crc kubenswrapper[4789]: I1122 08:27:01.065574 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-75ls5"] Nov 22 08:27:01 crc kubenswrapper[4789]: E1122 08:27:01.066023 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="424c42d2-3148-4308-91a1-9935d0ab2441" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 22 08:27:01 crc kubenswrapper[4789]: I1122 08:27:01.066045 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="424c42d2-3148-4308-91a1-9935d0ab2441" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 22 08:27:01 crc kubenswrapper[4789]: I1122 08:27:01.066209 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="424c42d2-3148-4308-91a1-9935d0ab2441" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 22 08:27:01 crc kubenswrapper[4789]: I1122 08:27:01.066859 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-75ls5" Nov 22 08:27:01 crc kubenswrapper[4789]: I1122 08:27:01.068921 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 22 08:27:01 crc kubenswrapper[4789]: I1122 08:27:01.068948 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 22 08:27:01 crc kubenswrapper[4789]: I1122 08:27:01.069437 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-glmr7" Nov 22 08:27:01 crc kubenswrapper[4789]: I1122 08:27:01.069655 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 22 08:27:01 crc kubenswrapper[4789]: I1122 08:27:01.079821 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-75ls5"] Nov 22 08:27:01 crc kubenswrapper[4789]: I1122 08:27:01.214415 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/276e9f73-f469-44c1-b346-3cca0768cc87-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-75ls5\" (UID: \"276e9f73-f469-44c1-b346-3cca0768cc87\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-75ls5" Nov 22 08:27:01 crc kubenswrapper[4789]: I1122 08:27:01.214493 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/276e9f73-f469-44c1-b346-3cca0768cc87-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-75ls5\" (UID: \"276e9f73-f469-44c1-b346-3cca0768cc87\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-75ls5" Nov 22 08:27:01 crc kubenswrapper[4789]: I1122 08:27:01.214622 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wp8gm\" (UniqueName: \"kubernetes.io/projected/276e9f73-f469-44c1-b346-3cca0768cc87-kube-api-access-wp8gm\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-75ls5\" (UID: \"276e9f73-f469-44c1-b346-3cca0768cc87\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-75ls5" Nov 22 08:27:01 crc kubenswrapper[4789]: I1122 08:27:01.315979 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wp8gm\" (UniqueName: \"kubernetes.io/projected/276e9f73-f469-44c1-b346-3cca0768cc87-kube-api-access-wp8gm\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-75ls5\" (UID: \"276e9f73-f469-44c1-b346-3cca0768cc87\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-75ls5" Nov 22 08:27:01 crc kubenswrapper[4789]: I1122 08:27:01.316059 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/276e9f73-f469-44c1-b346-3cca0768cc87-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-75ls5\" (UID: \"276e9f73-f469-44c1-b346-3cca0768cc87\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-75ls5" Nov 22 08:27:01 crc kubenswrapper[4789]: I1122 08:27:01.316082 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/276e9f73-f469-44c1-b346-3cca0768cc87-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-75ls5\" (UID: \"276e9f73-f469-44c1-b346-3cca0768cc87\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-75ls5" Nov 22 08:27:01 crc kubenswrapper[4789]: I1122 08:27:01.323194 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/276e9f73-f469-44c1-b346-3cca0768cc87-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-75ls5\" (UID: \"276e9f73-f469-44c1-b346-3cca0768cc87\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-75ls5" Nov 22 08:27:01 crc kubenswrapper[4789]: I1122 08:27:01.323280 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/276e9f73-f469-44c1-b346-3cca0768cc87-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-75ls5\" (UID: \"276e9f73-f469-44c1-b346-3cca0768cc87\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-75ls5" Nov 22 08:27:01 crc kubenswrapper[4789]: I1122 08:27:01.336630 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wp8gm\" (UniqueName: \"kubernetes.io/projected/276e9f73-f469-44c1-b346-3cca0768cc87-kube-api-access-wp8gm\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-75ls5\" (UID: \"276e9f73-f469-44c1-b346-3cca0768cc87\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-75ls5" Nov 22 08:27:01 crc kubenswrapper[4789]: I1122 08:27:01.390044 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-75ls5" Nov 22 08:27:01 crc kubenswrapper[4789]: I1122 08:27:01.910763 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-75ls5"] Nov 22 08:27:02 crc kubenswrapper[4789]: I1122 08:27:02.002617 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-75ls5" event={"ID":"276e9f73-f469-44c1-b346-3cca0768cc87","Type":"ContainerStarted","Data":"e849ce91b27a09334b280c535ea122be1607ed71083cb33b0087720a2223cc59"} Nov 22 08:27:03 crc kubenswrapper[4789]: I1122 08:27:03.015179 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-75ls5" event={"ID":"276e9f73-f469-44c1-b346-3cca0768cc87","Type":"ContainerStarted","Data":"e1de92c1a996b7edeb1227765e69a94a410b28c11e3c8a73369fab521842f428"} Nov 22 08:27:03 crc kubenswrapper[4789]: I1122 08:27:03.034925 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-75ls5" podStartSLOduration=1.343649034 podStartE2EDuration="2.034902675s" podCreationTimestamp="2025-11-22 08:27:01 +0000 UTC" firstStartedPulling="2025-11-22 08:27:01.916714089 +0000 UTC m=+1836.151114362" lastFinishedPulling="2025-11-22 08:27:02.60796773 +0000 UTC m=+1836.842368003" observedRunningTime="2025-11-22 08:27:03.029526328 +0000 UTC m=+1837.263926601" watchObservedRunningTime="2025-11-22 08:27:03.034902675 +0000 UTC m=+1837.269302948" Nov 22 08:27:08 crc kubenswrapper[4789]: I1122 08:27:08.406933 4789 generic.go:334] "Generic (PLEG): container finished" podID="276e9f73-f469-44c1-b346-3cca0768cc87" containerID="e1de92c1a996b7edeb1227765e69a94a410b28c11e3c8a73369fab521842f428" exitCode=0 Nov 22 08:27:08 crc kubenswrapper[4789]: I1122 08:27:08.407034 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-75ls5" event={"ID":"276e9f73-f469-44c1-b346-3cca0768cc87","Type":"ContainerDied","Data":"e1de92c1a996b7edeb1227765e69a94a410b28c11e3c8a73369fab521842f428"} Nov 22 08:27:09 crc kubenswrapper[4789]: I1122 08:27:09.798223 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-75ls5" Nov 22 08:27:09 crc kubenswrapper[4789]: I1122 08:27:09.857011 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/276e9f73-f469-44c1-b346-3cca0768cc87-ssh-key\") pod \"276e9f73-f469-44c1-b346-3cca0768cc87\" (UID: \"276e9f73-f469-44c1-b346-3cca0768cc87\") " Nov 22 08:27:09 crc kubenswrapper[4789]: I1122 08:27:09.857452 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wp8gm\" (UniqueName: \"kubernetes.io/projected/276e9f73-f469-44c1-b346-3cca0768cc87-kube-api-access-wp8gm\") pod \"276e9f73-f469-44c1-b346-3cca0768cc87\" (UID: \"276e9f73-f469-44c1-b346-3cca0768cc87\") " Nov 22 08:27:09 crc kubenswrapper[4789]: I1122 08:27:09.857601 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/276e9f73-f469-44c1-b346-3cca0768cc87-inventory\") pod \"276e9f73-f469-44c1-b346-3cca0768cc87\" (UID: \"276e9f73-f469-44c1-b346-3cca0768cc87\") " Nov 22 08:27:09 crc kubenswrapper[4789]: I1122 08:27:09.863143 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/276e9f73-f469-44c1-b346-3cca0768cc87-kube-api-access-wp8gm" (OuterVolumeSpecName: "kube-api-access-wp8gm") pod "276e9f73-f469-44c1-b346-3cca0768cc87" (UID: "276e9f73-f469-44c1-b346-3cca0768cc87"). InnerVolumeSpecName "kube-api-access-wp8gm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:27:09 crc kubenswrapper[4789]: I1122 08:27:09.886066 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/276e9f73-f469-44c1-b346-3cca0768cc87-inventory" (OuterVolumeSpecName: "inventory") pod "276e9f73-f469-44c1-b346-3cca0768cc87" (UID: "276e9f73-f469-44c1-b346-3cca0768cc87"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:27:09 crc kubenswrapper[4789]: I1122 08:27:09.888151 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/276e9f73-f469-44c1-b346-3cca0768cc87-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "276e9f73-f469-44c1-b346-3cca0768cc87" (UID: "276e9f73-f469-44c1-b346-3cca0768cc87"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:27:09 crc kubenswrapper[4789]: I1122 08:27:09.961137 4789 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/276e9f73-f469-44c1-b346-3cca0768cc87-inventory\") on node \"crc\" DevicePath \"\"" Nov 22 08:27:09 crc kubenswrapper[4789]: I1122 08:27:09.961197 4789 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/276e9f73-f469-44c1-b346-3cca0768cc87-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 22 08:27:09 crc kubenswrapper[4789]: I1122 08:27:09.961211 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wp8gm\" (UniqueName: \"kubernetes.io/projected/276e9f73-f469-44c1-b346-3cca0768cc87-kube-api-access-wp8gm\") on node \"crc\" DevicePath \"\"" Nov 22 08:27:10 crc kubenswrapper[4789]: I1122 08:27:10.430460 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-75ls5" event={"ID":"276e9f73-f469-44c1-b346-3cca0768cc87","Type":"ContainerDied","Data":"e849ce91b27a09334b280c535ea122be1607ed71083cb33b0087720a2223cc59"} Nov 22 08:27:10 crc kubenswrapper[4789]: I1122 08:27:10.430508 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e849ce91b27a09334b280c535ea122be1607ed71083cb33b0087720a2223cc59" Nov 22 08:27:10 crc kubenswrapper[4789]: I1122 08:27:10.430531 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-75ls5" Nov 22 08:27:10 crc kubenswrapper[4789]: I1122 08:27:10.484167 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-8rxgw"] Nov 22 08:27:10 crc kubenswrapper[4789]: E1122 08:27:10.484551 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="276e9f73-f469-44c1-b346-3cca0768cc87" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 22 08:27:10 crc kubenswrapper[4789]: I1122 08:27:10.484571 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="276e9f73-f469-44c1-b346-3cca0768cc87" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 22 08:27:10 crc kubenswrapper[4789]: I1122 08:27:10.484765 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="276e9f73-f469-44c1-b346-3cca0768cc87" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 22 08:27:10 crc kubenswrapper[4789]: I1122 08:27:10.485396 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-8rxgw" Nov 22 08:27:10 crc kubenswrapper[4789]: I1122 08:27:10.487353 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-glmr7" Nov 22 08:27:10 crc kubenswrapper[4789]: I1122 08:27:10.487432 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 22 08:27:10 crc kubenswrapper[4789]: I1122 08:27:10.487659 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 22 08:27:10 crc kubenswrapper[4789]: I1122 08:27:10.487791 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 22 08:27:10 crc kubenswrapper[4789]: I1122 08:27:10.502272 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-8rxgw"] Nov 22 08:27:10 crc kubenswrapper[4789]: I1122 08:27:10.572803 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fb964d35-f23a-4c63-a480-eeeea1e2f305-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-8rxgw\" (UID: \"fb964d35-f23a-4c63-a480-eeeea1e2f305\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-8rxgw" Nov 22 08:27:10 crc kubenswrapper[4789]: I1122 08:27:10.573300 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qqqkk\" (UniqueName: \"kubernetes.io/projected/fb964d35-f23a-4c63-a480-eeeea1e2f305-kube-api-access-qqqkk\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-8rxgw\" (UID: \"fb964d35-f23a-4c63-a480-eeeea1e2f305\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-8rxgw" Nov 22 08:27:10 crc kubenswrapper[4789]: I1122 08:27:10.573405 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fb964d35-f23a-4c63-a480-eeeea1e2f305-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-8rxgw\" (UID: \"fb964d35-f23a-4c63-a480-eeeea1e2f305\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-8rxgw" Nov 22 08:27:10 crc kubenswrapper[4789]: I1122 08:27:10.675102 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fb964d35-f23a-4c63-a480-eeeea1e2f305-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-8rxgw\" (UID: \"fb964d35-f23a-4c63-a480-eeeea1e2f305\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-8rxgw" Nov 22 08:27:10 crc kubenswrapper[4789]: I1122 08:27:10.675307 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fb964d35-f23a-4c63-a480-eeeea1e2f305-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-8rxgw\" (UID: \"fb964d35-f23a-4c63-a480-eeeea1e2f305\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-8rxgw" Nov 22 08:27:10 crc kubenswrapper[4789]: I1122 08:27:10.675454 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qqqkk\" (UniqueName: \"kubernetes.io/projected/fb964d35-f23a-4c63-a480-eeeea1e2f305-kube-api-access-qqqkk\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-8rxgw\" (UID: \"fb964d35-f23a-4c63-a480-eeeea1e2f305\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-8rxgw" Nov 22 08:27:10 crc kubenswrapper[4789]: I1122 08:27:10.681064 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fb964d35-f23a-4c63-a480-eeeea1e2f305-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-8rxgw\" (UID: \"fb964d35-f23a-4c63-a480-eeeea1e2f305\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-8rxgw" Nov 22 08:27:10 crc kubenswrapper[4789]: I1122 08:27:10.681297 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fb964d35-f23a-4c63-a480-eeeea1e2f305-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-8rxgw\" (UID: \"fb964d35-f23a-4c63-a480-eeeea1e2f305\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-8rxgw" Nov 22 08:27:10 crc kubenswrapper[4789]: I1122 08:27:10.693444 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qqqkk\" (UniqueName: \"kubernetes.io/projected/fb964d35-f23a-4c63-a480-eeeea1e2f305-kube-api-access-qqqkk\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-8rxgw\" (UID: \"fb964d35-f23a-4c63-a480-eeeea1e2f305\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-8rxgw" Nov 22 08:27:10 crc kubenswrapper[4789]: I1122 08:27:10.802487 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-8rxgw" Nov 22 08:27:11 crc kubenswrapper[4789]: I1122 08:27:11.282910 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-8rxgw"] Nov 22 08:27:11 crc kubenswrapper[4789]: I1122 08:27:11.439828 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-8rxgw" event={"ID":"fb964d35-f23a-4c63-a480-eeeea1e2f305","Type":"ContainerStarted","Data":"8c779c4cdb28b34caf51c237b3e6b552a3f0e040832e93ccfb092f965ce11ed1"} Nov 22 08:27:11 crc kubenswrapper[4789]: I1122 08:27:11.965393 4789 scope.go:117] "RemoveContainer" containerID="cd9752d9c3408de592a39b57fc6c7443fb52d47aecefc0deabddc20b20428c47" Nov 22 08:27:11 crc kubenswrapper[4789]: E1122 08:27:11.965877 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 08:27:12 crc kubenswrapper[4789]: I1122 08:27:12.449871 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-8rxgw" event={"ID":"fb964d35-f23a-4c63-a480-eeeea1e2f305","Type":"ContainerStarted","Data":"645d7595e012c68741c83ef8491e3fc0e745a63b28c053fe594370ee4c2d89b5"} Nov 22 08:27:12 crc kubenswrapper[4789]: I1122 08:27:12.475393 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-8rxgw" podStartSLOduration=2.063333838 podStartE2EDuration="2.475373157s" podCreationTimestamp="2025-11-22 08:27:10 +0000 UTC" firstStartedPulling="2025-11-22 08:27:11.289771644 +0000 UTC m=+1845.524171917" lastFinishedPulling="2025-11-22 08:27:11.701810973 +0000 UTC m=+1845.936211236" observedRunningTime="2025-11-22 08:27:12.469194809 +0000 UTC m=+1846.703595082" watchObservedRunningTime="2025-11-22 08:27:12.475373157 +0000 UTC m=+1846.709773430" Nov 22 08:27:22 crc kubenswrapper[4789]: I1122 08:27:22.964865 4789 scope.go:117] "RemoveContainer" containerID="cd9752d9c3408de592a39b57fc6c7443fb52d47aecefc0deabddc20b20428c47" Nov 22 08:27:22 crc kubenswrapper[4789]: E1122 08:27:22.965640 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 08:27:25 crc kubenswrapper[4789]: I1122 08:27:25.565905 4789 scope.go:117] "RemoveContainer" containerID="848a353afae49b131c8b3082febbe24a3efdd0d7749888b4a1aa1c94bf99fb13" Nov 22 08:27:25 crc kubenswrapper[4789]: I1122 08:27:25.596303 4789 scope.go:117] "RemoveContainer" containerID="50273be882e06c5bebf9373f13c41c1860a0308e06ec2d3632e22d25ca316617" Nov 22 08:27:25 crc kubenswrapper[4789]: I1122 08:27:25.642967 4789 scope.go:117] "RemoveContainer" containerID="aa73593836103e7d3c745c9dd3a425820bddd9b5607f5bdc6e6dfaa4e6e0506f" Nov 22 08:27:25 crc kubenswrapper[4789]: I1122 08:27:25.699424 4789 scope.go:117] "RemoveContainer" containerID="bf24bad4cf9f730c08c10af9dcc81cda4c4d4644510e1624d12b515e8bde48e6" Nov 22 08:27:25 crc kubenswrapper[4789]: I1122 08:27:25.716253 4789 scope.go:117] "RemoveContainer" containerID="e7c9611deeb5f070f55bebba5b8617f188d07f83303c25a7a62ae83a003facef" Nov 22 08:27:25 crc kubenswrapper[4789]: I1122 08:27:25.755336 4789 scope.go:117] "RemoveContainer" containerID="43622e76a3aa6a505c8df9823f7041e41a83be4429150db9572330e047631f10" Nov 22 08:27:28 crc kubenswrapper[4789]: I1122 08:27:28.037507 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-frb4x"] Nov 22 08:27:28 crc kubenswrapper[4789]: I1122 08:27:28.045665 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-frb4x"] Nov 22 08:27:29 crc kubenswrapper[4789]: I1122 08:27:29.031547 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-8cdf-account-create-hzl44"] Nov 22 08:27:29 crc kubenswrapper[4789]: I1122 08:27:29.040686 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-5mvvx"] Nov 22 08:27:29 crc kubenswrapper[4789]: I1122 08:27:29.048457 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-vcmgb"] Nov 22 08:27:29 crc kubenswrapper[4789]: I1122 08:27:29.056392 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-8cdf-account-create-hzl44"] Nov 22 08:27:29 crc kubenswrapper[4789]: I1122 08:27:29.063450 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-8958-account-create-626c2"] Nov 22 08:27:29 crc kubenswrapper[4789]: I1122 08:27:29.070764 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-9a6c-account-create-2cwpt"] Nov 22 08:27:29 crc kubenswrapper[4789]: I1122 08:27:29.076859 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-5mvvx"] Nov 22 08:27:29 crc kubenswrapper[4789]: I1122 08:27:29.083130 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-vcmgb"] Nov 22 08:27:29 crc kubenswrapper[4789]: I1122 08:27:29.089043 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-9a6c-account-create-2cwpt"] Nov 22 08:27:29 crc kubenswrapper[4789]: I1122 08:27:29.094875 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-8958-account-create-626c2"] Nov 22 08:27:30 crc kubenswrapper[4789]: I1122 08:27:30.084032 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="07941151-cfeb-4356-ad4e-439fdb875687" path="/var/lib/kubelet/pods/07941151-cfeb-4356-ad4e-439fdb875687/volumes" Nov 22 08:27:30 crc kubenswrapper[4789]: I1122 08:27:30.085368 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="83e51032-30aa-47cb-b856-36b4280c2775" path="/var/lib/kubelet/pods/83e51032-30aa-47cb-b856-36b4280c2775/volumes" Nov 22 08:27:30 crc kubenswrapper[4789]: I1122 08:27:30.085919 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9e886f6c-babc-4118-b532-50b72dacb370" path="/var/lib/kubelet/pods/9e886f6c-babc-4118-b532-50b72dacb370/volumes" Nov 22 08:27:30 crc kubenswrapper[4789]: I1122 08:27:30.086536 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aa7f415d-4528-41d4-996f-57c061b72885" path="/var/lib/kubelet/pods/aa7f415d-4528-41d4-996f-57c061b72885/volumes" Nov 22 08:27:30 crc kubenswrapper[4789]: I1122 08:27:30.100888 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c61136a6-9398-4fec-89a0-e4d7291b99fc" path="/var/lib/kubelet/pods/c61136a6-9398-4fec-89a0-e4d7291b99fc/volumes" Nov 22 08:27:30 crc kubenswrapper[4789]: I1122 08:27:30.102204 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d315bf00-e54d-40cb-a261-ff8bd068e535" path="/var/lib/kubelet/pods/d315bf00-e54d-40cb-a261-ff8bd068e535/volumes" Nov 22 08:27:37 crc kubenswrapper[4789]: I1122 08:27:37.971767 4789 scope.go:117] "RemoveContainer" containerID="cd9752d9c3408de592a39b57fc6c7443fb52d47aecefc0deabddc20b20428c47" Nov 22 08:27:38 crc kubenswrapper[4789]: I1122 08:27:38.671773 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" event={"ID":"58aa8071-7ecc-4692-8789-2db04ec70647","Type":"ContainerStarted","Data":"aaa33370060895e49b5a7fda013e1521882a1ec6a21f755348d821f46b72a57a"} Nov 22 08:27:52 crc kubenswrapper[4789]: I1122 08:27:52.794972 4789 generic.go:334] "Generic (PLEG): container finished" podID="fb964d35-f23a-4c63-a480-eeeea1e2f305" containerID="645d7595e012c68741c83ef8491e3fc0e745a63b28c053fe594370ee4c2d89b5" exitCode=0 Nov 22 08:27:52 crc kubenswrapper[4789]: I1122 08:27:52.795579 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-8rxgw" event={"ID":"fb964d35-f23a-4c63-a480-eeeea1e2f305","Type":"ContainerDied","Data":"645d7595e012c68741c83ef8491e3fc0e745a63b28c053fe594370ee4c2d89b5"} Nov 22 08:27:54 crc kubenswrapper[4789]: I1122 08:27:54.222104 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-8rxgw" Nov 22 08:27:54 crc kubenswrapper[4789]: I1122 08:27:54.250346 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fb964d35-f23a-4c63-a480-eeeea1e2f305-inventory\") pod \"fb964d35-f23a-4c63-a480-eeeea1e2f305\" (UID: \"fb964d35-f23a-4c63-a480-eeeea1e2f305\") " Nov 22 08:27:54 crc kubenswrapper[4789]: I1122 08:27:54.250438 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qqqkk\" (UniqueName: \"kubernetes.io/projected/fb964d35-f23a-4c63-a480-eeeea1e2f305-kube-api-access-qqqkk\") pod \"fb964d35-f23a-4c63-a480-eeeea1e2f305\" (UID: \"fb964d35-f23a-4c63-a480-eeeea1e2f305\") " Nov 22 08:27:54 crc kubenswrapper[4789]: I1122 08:27:54.250513 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fb964d35-f23a-4c63-a480-eeeea1e2f305-ssh-key\") pod \"fb964d35-f23a-4c63-a480-eeeea1e2f305\" (UID: \"fb964d35-f23a-4c63-a480-eeeea1e2f305\") " Nov 22 08:27:54 crc kubenswrapper[4789]: I1122 08:27:54.256233 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb964d35-f23a-4c63-a480-eeeea1e2f305-kube-api-access-qqqkk" (OuterVolumeSpecName: "kube-api-access-qqqkk") pod "fb964d35-f23a-4c63-a480-eeeea1e2f305" (UID: "fb964d35-f23a-4c63-a480-eeeea1e2f305"). InnerVolumeSpecName "kube-api-access-qqqkk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:27:54 crc kubenswrapper[4789]: I1122 08:27:54.275334 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb964d35-f23a-4c63-a480-eeeea1e2f305-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "fb964d35-f23a-4c63-a480-eeeea1e2f305" (UID: "fb964d35-f23a-4c63-a480-eeeea1e2f305"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:27:54 crc kubenswrapper[4789]: I1122 08:27:54.278829 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb964d35-f23a-4c63-a480-eeeea1e2f305-inventory" (OuterVolumeSpecName: "inventory") pod "fb964d35-f23a-4c63-a480-eeeea1e2f305" (UID: "fb964d35-f23a-4c63-a480-eeeea1e2f305"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:27:54 crc kubenswrapper[4789]: I1122 08:27:54.353114 4789 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fb964d35-f23a-4c63-a480-eeeea1e2f305-inventory\") on node \"crc\" DevicePath \"\"" Nov 22 08:27:54 crc kubenswrapper[4789]: I1122 08:27:54.353354 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qqqkk\" (UniqueName: \"kubernetes.io/projected/fb964d35-f23a-4c63-a480-eeeea1e2f305-kube-api-access-qqqkk\") on node \"crc\" DevicePath \"\"" Nov 22 08:27:54 crc kubenswrapper[4789]: I1122 08:27:54.353412 4789 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fb964d35-f23a-4c63-a480-eeeea1e2f305-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 22 08:27:54 crc kubenswrapper[4789]: I1122 08:27:54.813027 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-8rxgw" event={"ID":"fb964d35-f23a-4c63-a480-eeeea1e2f305","Type":"ContainerDied","Data":"8c779c4cdb28b34caf51c237b3e6b552a3f0e040832e93ccfb092f965ce11ed1"} Nov 22 08:27:54 crc kubenswrapper[4789]: I1122 08:27:54.813064 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8c779c4cdb28b34caf51c237b3e6b552a3f0e040832e93ccfb092f965ce11ed1" Nov 22 08:27:54 crc kubenswrapper[4789]: I1122 08:27:54.813077 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-8rxgw" Nov 22 08:27:54 crc kubenswrapper[4789]: I1122 08:27:54.887213 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-gt7gt"] Nov 22 08:27:54 crc kubenswrapper[4789]: E1122 08:27:54.887730 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb964d35-f23a-4c63-a480-eeeea1e2f305" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 22 08:27:54 crc kubenswrapper[4789]: I1122 08:27:54.887813 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb964d35-f23a-4c63-a480-eeeea1e2f305" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 22 08:27:54 crc kubenswrapper[4789]: I1122 08:27:54.888032 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb964d35-f23a-4c63-a480-eeeea1e2f305" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 22 08:27:54 crc kubenswrapper[4789]: I1122 08:27:54.888724 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-gt7gt" Nov 22 08:27:54 crc kubenswrapper[4789]: I1122 08:27:54.893175 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 22 08:27:54 crc kubenswrapper[4789]: I1122 08:27:54.893187 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 22 08:27:54 crc kubenswrapper[4789]: I1122 08:27:54.893792 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-glmr7" Nov 22 08:27:54 crc kubenswrapper[4789]: I1122 08:27:54.897243 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 22 08:27:54 crc kubenswrapper[4789]: I1122 08:27:54.900252 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-gt7gt"] Nov 22 08:27:54 crc kubenswrapper[4789]: I1122 08:27:54.963926 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8470bb71-39d5-4a92-b60e-856624688535-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-gt7gt\" (UID: \"8470bb71-39d5-4a92-b60e-856624688535\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-gt7gt" Nov 22 08:27:54 crc kubenswrapper[4789]: I1122 08:27:54.964000 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ltdlg\" (UniqueName: \"kubernetes.io/projected/8470bb71-39d5-4a92-b60e-856624688535-kube-api-access-ltdlg\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-gt7gt\" (UID: \"8470bb71-39d5-4a92-b60e-856624688535\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-gt7gt" Nov 22 08:27:54 crc kubenswrapper[4789]: I1122 08:27:54.964024 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8470bb71-39d5-4a92-b60e-856624688535-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-gt7gt\" (UID: \"8470bb71-39d5-4a92-b60e-856624688535\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-gt7gt" Nov 22 08:27:55 crc kubenswrapper[4789]: I1122 08:27:55.065436 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8470bb71-39d5-4a92-b60e-856624688535-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-gt7gt\" (UID: \"8470bb71-39d5-4a92-b60e-856624688535\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-gt7gt" Nov 22 08:27:55 crc kubenswrapper[4789]: I1122 08:27:55.065860 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ltdlg\" (UniqueName: \"kubernetes.io/projected/8470bb71-39d5-4a92-b60e-856624688535-kube-api-access-ltdlg\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-gt7gt\" (UID: \"8470bb71-39d5-4a92-b60e-856624688535\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-gt7gt" Nov 22 08:27:55 crc kubenswrapper[4789]: I1122 08:27:55.065889 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8470bb71-39d5-4a92-b60e-856624688535-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-gt7gt\" (UID: \"8470bb71-39d5-4a92-b60e-856624688535\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-gt7gt" Nov 22 08:27:55 crc kubenswrapper[4789]: I1122 08:27:55.071302 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8470bb71-39d5-4a92-b60e-856624688535-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-gt7gt\" (UID: \"8470bb71-39d5-4a92-b60e-856624688535\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-gt7gt" Nov 22 08:27:55 crc kubenswrapper[4789]: I1122 08:27:55.072359 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8470bb71-39d5-4a92-b60e-856624688535-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-gt7gt\" (UID: \"8470bb71-39d5-4a92-b60e-856624688535\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-gt7gt" Nov 22 08:27:55 crc kubenswrapper[4789]: I1122 08:27:55.093652 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ltdlg\" (UniqueName: \"kubernetes.io/projected/8470bb71-39d5-4a92-b60e-856624688535-kube-api-access-ltdlg\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-gt7gt\" (UID: \"8470bb71-39d5-4a92-b60e-856624688535\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-gt7gt" Nov 22 08:27:55 crc kubenswrapper[4789]: I1122 08:27:55.207191 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-gt7gt" Nov 22 08:27:55 crc kubenswrapper[4789]: I1122 08:27:55.693582 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-gt7gt"] Nov 22 08:27:55 crc kubenswrapper[4789]: I1122 08:27:55.822281 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-gt7gt" event={"ID":"8470bb71-39d5-4a92-b60e-856624688535","Type":"ContainerStarted","Data":"e054a94d50f333bb3dd4302b487ba617c199f29614683152cb5b337ab7e5a4f4"} Nov 22 08:27:56 crc kubenswrapper[4789]: I1122 08:27:56.834274 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-gt7gt" event={"ID":"8470bb71-39d5-4a92-b60e-856624688535","Type":"ContainerStarted","Data":"4d4f6ad56c7b19b2010140fc50c9e595f63e55d53556011d40962dbd1a9d123d"} Nov 22 08:28:00 crc kubenswrapper[4789]: I1122 08:28:00.873641 4789 generic.go:334] "Generic (PLEG): container finished" podID="8470bb71-39d5-4a92-b60e-856624688535" containerID="4d4f6ad56c7b19b2010140fc50c9e595f63e55d53556011d40962dbd1a9d123d" exitCode=0 Nov 22 08:28:00 crc kubenswrapper[4789]: I1122 08:28:00.873726 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-gt7gt" event={"ID":"8470bb71-39d5-4a92-b60e-856624688535","Type":"ContainerDied","Data":"4d4f6ad56c7b19b2010140fc50c9e595f63e55d53556011d40962dbd1a9d123d"} Nov 22 08:28:02 crc kubenswrapper[4789]: I1122 08:28:02.265554 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-gt7gt" Nov 22 08:28:02 crc kubenswrapper[4789]: I1122 08:28:02.396096 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ltdlg\" (UniqueName: \"kubernetes.io/projected/8470bb71-39d5-4a92-b60e-856624688535-kube-api-access-ltdlg\") pod \"8470bb71-39d5-4a92-b60e-856624688535\" (UID: \"8470bb71-39d5-4a92-b60e-856624688535\") " Nov 22 08:28:02 crc kubenswrapper[4789]: I1122 08:28:02.396155 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8470bb71-39d5-4a92-b60e-856624688535-ssh-key\") pod \"8470bb71-39d5-4a92-b60e-856624688535\" (UID: \"8470bb71-39d5-4a92-b60e-856624688535\") " Nov 22 08:28:02 crc kubenswrapper[4789]: I1122 08:28:02.396273 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8470bb71-39d5-4a92-b60e-856624688535-inventory\") pod \"8470bb71-39d5-4a92-b60e-856624688535\" (UID: \"8470bb71-39d5-4a92-b60e-856624688535\") " Nov 22 08:28:02 crc kubenswrapper[4789]: I1122 08:28:02.401937 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8470bb71-39d5-4a92-b60e-856624688535-kube-api-access-ltdlg" (OuterVolumeSpecName: "kube-api-access-ltdlg") pod "8470bb71-39d5-4a92-b60e-856624688535" (UID: "8470bb71-39d5-4a92-b60e-856624688535"). InnerVolumeSpecName "kube-api-access-ltdlg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:28:02 crc kubenswrapper[4789]: I1122 08:28:02.425446 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8470bb71-39d5-4a92-b60e-856624688535-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "8470bb71-39d5-4a92-b60e-856624688535" (UID: "8470bb71-39d5-4a92-b60e-856624688535"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:28:02 crc kubenswrapper[4789]: I1122 08:28:02.426139 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8470bb71-39d5-4a92-b60e-856624688535-inventory" (OuterVolumeSpecName: "inventory") pod "8470bb71-39d5-4a92-b60e-856624688535" (UID: "8470bb71-39d5-4a92-b60e-856624688535"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:28:02 crc kubenswrapper[4789]: I1122 08:28:02.498880 4789 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8470bb71-39d5-4a92-b60e-856624688535-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 22 08:28:02 crc kubenswrapper[4789]: I1122 08:28:02.498915 4789 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8470bb71-39d5-4a92-b60e-856624688535-inventory\") on node \"crc\" DevicePath \"\"" Nov 22 08:28:02 crc kubenswrapper[4789]: I1122 08:28:02.498926 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ltdlg\" (UniqueName: \"kubernetes.io/projected/8470bb71-39d5-4a92-b60e-856624688535-kube-api-access-ltdlg\") on node \"crc\" DevicePath \"\"" Nov 22 08:28:02 crc kubenswrapper[4789]: I1122 08:28:02.896465 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-gt7gt" event={"ID":"8470bb71-39d5-4a92-b60e-856624688535","Type":"ContainerDied","Data":"e054a94d50f333bb3dd4302b487ba617c199f29614683152cb5b337ab7e5a4f4"} Nov 22 08:28:02 crc kubenswrapper[4789]: I1122 08:28:02.896519 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e054a94d50f333bb3dd4302b487ba617c199f29614683152cb5b337ab7e5a4f4" Nov 22 08:28:02 crc kubenswrapper[4789]: I1122 08:28:02.896665 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-gt7gt" Nov 22 08:28:02 crc kubenswrapper[4789]: I1122 08:28:02.962661 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kdxvv"] Nov 22 08:28:02 crc kubenswrapper[4789]: E1122 08:28:02.963136 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8470bb71-39d5-4a92-b60e-856624688535" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Nov 22 08:28:02 crc kubenswrapper[4789]: I1122 08:28:02.963159 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="8470bb71-39d5-4a92-b60e-856624688535" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Nov 22 08:28:02 crc kubenswrapper[4789]: I1122 08:28:02.963342 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="8470bb71-39d5-4a92-b60e-856624688535" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Nov 22 08:28:02 crc kubenswrapper[4789]: I1122 08:28:02.963957 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kdxvv" Nov 22 08:28:02 crc kubenswrapper[4789]: I1122 08:28:02.965814 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 22 08:28:02 crc kubenswrapper[4789]: I1122 08:28:02.966255 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 22 08:28:02 crc kubenswrapper[4789]: I1122 08:28:02.967231 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 22 08:28:02 crc kubenswrapper[4789]: I1122 08:28:02.967928 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-glmr7" Nov 22 08:28:02 crc kubenswrapper[4789]: I1122 08:28:02.978284 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kdxvv"] Nov 22 08:28:03 crc kubenswrapper[4789]: I1122 08:28:03.007937 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0eba14cc-decf-4b92-adbb-3cf60adc26f3-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-kdxvv\" (UID: \"0eba14cc-decf-4b92-adbb-3cf60adc26f3\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kdxvv" Nov 22 08:28:03 crc kubenswrapper[4789]: I1122 08:28:03.008130 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rgj7z\" (UniqueName: \"kubernetes.io/projected/0eba14cc-decf-4b92-adbb-3cf60adc26f3-kube-api-access-rgj7z\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-kdxvv\" (UID: \"0eba14cc-decf-4b92-adbb-3cf60adc26f3\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kdxvv" Nov 22 08:28:03 crc kubenswrapper[4789]: I1122 08:28:03.008185 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0eba14cc-decf-4b92-adbb-3cf60adc26f3-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-kdxvv\" (UID: \"0eba14cc-decf-4b92-adbb-3cf60adc26f3\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kdxvv" Nov 22 08:28:03 crc kubenswrapper[4789]: I1122 08:28:03.110433 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rgj7z\" (UniqueName: \"kubernetes.io/projected/0eba14cc-decf-4b92-adbb-3cf60adc26f3-kube-api-access-rgj7z\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-kdxvv\" (UID: \"0eba14cc-decf-4b92-adbb-3cf60adc26f3\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kdxvv" Nov 22 08:28:03 crc kubenswrapper[4789]: I1122 08:28:03.110508 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0eba14cc-decf-4b92-adbb-3cf60adc26f3-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-kdxvv\" (UID: \"0eba14cc-decf-4b92-adbb-3cf60adc26f3\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kdxvv" Nov 22 08:28:03 crc kubenswrapper[4789]: I1122 08:28:03.110619 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0eba14cc-decf-4b92-adbb-3cf60adc26f3-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-kdxvv\" (UID: \"0eba14cc-decf-4b92-adbb-3cf60adc26f3\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kdxvv" Nov 22 08:28:03 crc kubenswrapper[4789]: I1122 08:28:03.114630 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0eba14cc-decf-4b92-adbb-3cf60adc26f3-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-kdxvv\" (UID: \"0eba14cc-decf-4b92-adbb-3cf60adc26f3\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kdxvv" Nov 22 08:28:03 crc kubenswrapper[4789]: I1122 08:28:03.115189 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0eba14cc-decf-4b92-adbb-3cf60adc26f3-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-kdxvv\" (UID: \"0eba14cc-decf-4b92-adbb-3cf60adc26f3\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kdxvv" Nov 22 08:28:03 crc kubenswrapper[4789]: I1122 08:28:03.129012 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rgj7z\" (UniqueName: \"kubernetes.io/projected/0eba14cc-decf-4b92-adbb-3cf60adc26f3-kube-api-access-rgj7z\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-kdxvv\" (UID: \"0eba14cc-decf-4b92-adbb-3cf60adc26f3\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kdxvv" Nov 22 08:28:03 crc kubenswrapper[4789]: I1122 08:28:03.285453 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kdxvv" Nov 22 08:28:03 crc kubenswrapper[4789]: I1122 08:28:03.792860 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kdxvv"] Nov 22 08:28:03 crc kubenswrapper[4789]: I1122 08:28:03.906324 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kdxvv" event={"ID":"0eba14cc-decf-4b92-adbb-3cf60adc26f3","Type":"ContainerStarted","Data":"4ab1eed33b8b849568389b29fb6bb6462ca4adaddecffa75c4620f3ccb4334cf"} Nov 22 08:28:04 crc kubenswrapper[4789]: I1122 08:28:04.916504 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kdxvv" event={"ID":"0eba14cc-decf-4b92-adbb-3cf60adc26f3","Type":"ContainerStarted","Data":"7c88ffd385b6dc5704d986c3f01a32015907ab893d1061705ab6115b81533966"} Nov 22 08:28:04 crc kubenswrapper[4789]: I1122 08:28:04.934204 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kdxvv" podStartSLOduration=2.568519267 podStartE2EDuration="2.934182822s" podCreationTimestamp="2025-11-22 08:28:02 +0000 UTC" firstStartedPulling="2025-11-22 08:28:03.804359073 +0000 UTC m=+1898.038759366" lastFinishedPulling="2025-11-22 08:28:04.170022648 +0000 UTC m=+1898.404422921" observedRunningTime="2025-11-22 08:28:04.932604049 +0000 UTC m=+1899.167004332" watchObservedRunningTime="2025-11-22 08:28:04.934182822 +0000 UTC m=+1899.168583095" Nov 22 08:28:25 crc kubenswrapper[4789]: I1122 08:28:25.876278 4789 scope.go:117] "RemoveContainer" containerID="d2be3b96f1f09315958202746fe2b6ff79bc62d91dcf7feef03096e72661d2f5" Nov 22 08:28:25 crc kubenswrapper[4789]: I1122 08:28:25.909050 4789 scope.go:117] "RemoveContainer" containerID="9b84e92c2e789ca0e69d366aec32c9b0a04e8eb06bc59b5b87e9d586cef870a1" Nov 22 08:28:25 crc kubenswrapper[4789]: I1122 08:28:25.967009 4789 scope.go:117] "RemoveContainer" containerID="4ccac87ec671fab5f21281811992e6a8edb929190b35525ba420234d05dff837" Nov 22 08:28:25 crc kubenswrapper[4789]: I1122 08:28:25.998811 4789 scope.go:117] "RemoveContainer" containerID="d969b22b590e1304aa90136e93d1646335c192a7ee87707af43ab394c65b5768" Nov 22 08:28:26 crc kubenswrapper[4789]: I1122 08:28:26.040287 4789 scope.go:117] "RemoveContainer" containerID="8c25a7cf637bf7ef49c67f9e27a1ebd73539326339bf61342245cf3d89013cce" Nov 22 08:28:26 crc kubenswrapper[4789]: I1122 08:28:26.080022 4789 scope.go:117] "RemoveContainer" containerID="ccc2c4d2207917c6e4fa2291159a5b1d29db8d298af4a1d1ef3d948cf7e498b5" Nov 22 08:28:28 crc kubenswrapper[4789]: I1122 08:28:28.038942 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-mg9lj"] Nov 22 08:28:28 crc kubenswrapper[4789]: I1122 08:28:28.052014 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-mg9lj"] Nov 22 08:28:29 crc kubenswrapper[4789]: I1122 08:28:29.975439 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8ea41118-064f-4760-a1cc-0ebe3fc81a02" path="/var/lib/kubelet/pods/8ea41118-064f-4760-a1cc-0ebe3fc81a02/volumes" Nov 22 08:28:50 crc kubenswrapper[4789]: I1122 08:28:50.036185 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-fzs5c"] Nov 22 08:28:50 crc kubenswrapper[4789]: I1122 08:28:50.042857 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-fzs5c"] Nov 22 08:28:51 crc kubenswrapper[4789]: I1122 08:28:51.023419 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-skqfc"] Nov 22 08:28:51 crc kubenswrapper[4789]: I1122 08:28:51.029687 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-skqfc"] Nov 22 08:28:51 crc kubenswrapper[4789]: I1122 08:28:51.975587 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="79d1ad09-f02f-421e-b1e5-10860722cc77" path="/var/lib/kubelet/pods/79d1ad09-f02f-421e-b1e5-10860722cc77/volumes" Nov 22 08:28:51 crc kubenswrapper[4789]: I1122 08:28:51.976862 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d4270eeb-3da5-464a-806a-358dbc4bd6ba" path="/var/lib/kubelet/pods/d4270eeb-3da5-464a-806a-358dbc4bd6ba/volumes" Nov 22 08:28:58 crc kubenswrapper[4789]: I1122 08:28:58.378557 4789 generic.go:334] "Generic (PLEG): container finished" podID="0eba14cc-decf-4b92-adbb-3cf60adc26f3" containerID="7c88ffd385b6dc5704d986c3f01a32015907ab893d1061705ab6115b81533966" exitCode=0 Nov 22 08:28:58 crc kubenswrapper[4789]: I1122 08:28:58.378630 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kdxvv" event={"ID":"0eba14cc-decf-4b92-adbb-3cf60adc26f3","Type":"ContainerDied","Data":"7c88ffd385b6dc5704d986c3f01a32015907ab893d1061705ab6115b81533966"} Nov 22 08:28:59 crc kubenswrapper[4789]: I1122 08:28:59.762938 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kdxvv" Nov 22 08:28:59 crc kubenswrapper[4789]: I1122 08:28:59.807810 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rgj7z\" (UniqueName: \"kubernetes.io/projected/0eba14cc-decf-4b92-adbb-3cf60adc26f3-kube-api-access-rgj7z\") pod \"0eba14cc-decf-4b92-adbb-3cf60adc26f3\" (UID: \"0eba14cc-decf-4b92-adbb-3cf60adc26f3\") " Nov 22 08:28:59 crc kubenswrapper[4789]: I1122 08:28:59.807909 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0eba14cc-decf-4b92-adbb-3cf60adc26f3-inventory\") pod \"0eba14cc-decf-4b92-adbb-3cf60adc26f3\" (UID: \"0eba14cc-decf-4b92-adbb-3cf60adc26f3\") " Nov 22 08:28:59 crc kubenswrapper[4789]: I1122 08:28:59.807937 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0eba14cc-decf-4b92-adbb-3cf60adc26f3-ssh-key\") pod \"0eba14cc-decf-4b92-adbb-3cf60adc26f3\" (UID: \"0eba14cc-decf-4b92-adbb-3cf60adc26f3\") " Nov 22 08:28:59 crc kubenswrapper[4789]: I1122 08:28:59.814555 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0eba14cc-decf-4b92-adbb-3cf60adc26f3-kube-api-access-rgj7z" (OuterVolumeSpecName: "kube-api-access-rgj7z") pod "0eba14cc-decf-4b92-adbb-3cf60adc26f3" (UID: "0eba14cc-decf-4b92-adbb-3cf60adc26f3"). InnerVolumeSpecName "kube-api-access-rgj7z". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:28:59 crc kubenswrapper[4789]: I1122 08:28:59.836088 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0eba14cc-decf-4b92-adbb-3cf60adc26f3-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "0eba14cc-decf-4b92-adbb-3cf60adc26f3" (UID: "0eba14cc-decf-4b92-adbb-3cf60adc26f3"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:28:59 crc kubenswrapper[4789]: I1122 08:28:59.838112 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0eba14cc-decf-4b92-adbb-3cf60adc26f3-inventory" (OuterVolumeSpecName: "inventory") pod "0eba14cc-decf-4b92-adbb-3cf60adc26f3" (UID: "0eba14cc-decf-4b92-adbb-3cf60adc26f3"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:28:59 crc kubenswrapper[4789]: I1122 08:28:59.910970 4789 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0eba14cc-decf-4b92-adbb-3cf60adc26f3-inventory\") on node \"crc\" DevicePath \"\"" Nov 22 08:28:59 crc kubenswrapper[4789]: I1122 08:28:59.911001 4789 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0eba14cc-decf-4b92-adbb-3cf60adc26f3-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 22 08:28:59 crc kubenswrapper[4789]: I1122 08:28:59.911012 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rgj7z\" (UniqueName: \"kubernetes.io/projected/0eba14cc-decf-4b92-adbb-3cf60adc26f3-kube-api-access-rgj7z\") on node \"crc\" DevicePath \"\"" Nov 22 08:29:00 crc kubenswrapper[4789]: I1122 08:29:00.394624 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kdxvv" event={"ID":"0eba14cc-decf-4b92-adbb-3cf60adc26f3","Type":"ContainerDied","Data":"4ab1eed33b8b849568389b29fb6bb6462ca4adaddecffa75c4620f3ccb4334cf"} Nov 22 08:29:00 crc kubenswrapper[4789]: I1122 08:29:00.394665 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4ab1eed33b8b849568389b29fb6bb6462ca4adaddecffa75c4620f3ccb4334cf" Nov 22 08:29:00 crc kubenswrapper[4789]: I1122 08:29:00.394707 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kdxvv" Nov 22 08:29:00 crc kubenswrapper[4789]: I1122 08:29:00.470188 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-k2c8r"] Nov 22 08:29:00 crc kubenswrapper[4789]: E1122 08:29:00.470637 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0eba14cc-decf-4b92-adbb-3cf60adc26f3" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 22 08:29:00 crc kubenswrapper[4789]: I1122 08:29:00.470659 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="0eba14cc-decf-4b92-adbb-3cf60adc26f3" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 22 08:29:00 crc kubenswrapper[4789]: I1122 08:29:00.470946 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="0eba14cc-decf-4b92-adbb-3cf60adc26f3" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 22 08:29:00 crc kubenswrapper[4789]: I1122 08:29:00.471560 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-k2c8r" Nov 22 08:29:00 crc kubenswrapper[4789]: I1122 08:29:00.473895 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 22 08:29:00 crc kubenswrapper[4789]: I1122 08:29:00.474263 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 22 08:29:00 crc kubenswrapper[4789]: I1122 08:29:00.474474 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-glmr7" Nov 22 08:29:00 crc kubenswrapper[4789]: I1122 08:29:00.480339 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 22 08:29:00 crc kubenswrapper[4789]: I1122 08:29:00.491495 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-k2c8r"] Nov 22 08:29:00 crc kubenswrapper[4789]: I1122 08:29:00.527387 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/4cac03ab-292f-4255-aec2-76120f6f8c24-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-k2c8r\" (UID: \"4cac03ab-292f-4255-aec2-76120f6f8c24\") " pod="openstack/ssh-known-hosts-edpm-deployment-k2c8r" Nov 22 08:29:00 crc kubenswrapper[4789]: I1122 08:29:00.527463 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-db29g\" (UniqueName: \"kubernetes.io/projected/4cac03ab-292f-4255-aec2-76120f6f8c24-kube-api-access-db29g\") pod \"ssh-known-hosts-edpm-deployment-k2c8r\" (UID: \"4cac03ab-292f-4255-aec2-76120f6f8c24\") " pod="openstack/ssh-known-hosts-edpm-deployment-k2c8r" Nov 22 08:29:00 crc kubenswrapper[4789]: I1122 08:29:00.527508 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/4cac03ab-292f-4255-aec2-76120f6f8c24-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-k2c8r\" (UID: \"4cac03ab-292f-4255-aec2-76120f6f8c24\") " pod="openstack/ssh-known-hosts-edpm-deployment-k2c8r" Nov 22 08:29:00 crc kubenswrapper[4789]: I1122 08:29:00.629650 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/4cac03ab-292f-4255-aec2-76120f6f8c24-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-k2c8r\" (UID: \"4cac03ab-292f-4255-aec2-76120f6f8c24\") " pod="openstack/ssh-known-hosts-edpm-deployment-k2c8r" Nov 22 08:29:00 crc kubenswrapper[4789]: I1122 08:29:00.629735 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-db29g\" (UniqueName: \"kubernetes.io/projected/4cac03ab-292f-4255-aec2-76120f6f8c24-kube-api-access-db29g\") pod \"ssh-known-hosts-edpm-deployment-k2c8r\" (UID: \"4cac03ab-292f-4255-aec2-76120f6f8c24\") " pod="openstack/ssh-known-hosts-edpm-deployment-k2c8r" Nov 22 08:29:00 crc kubenswrapper[4789]: I1122 08:29:00.629824 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/4cac03ab-292f-4255-aec2-76120f6f8c24-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-k2c8r\" (UID: \"4cac03ab-292f-4255-aec2-76120f6f8c24\") " pod="openstack/ssh-known-hosts-edpm-deployment-k2c8r" Nov 22 08:29:00 crc kubenswrapper[4789]: I1122 08:29:00.633116 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/4cac03ab-292f-4255-aec2-76120f6f8c24-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-k2c8r\" (UID: \"4cac03ab-292f-4255-aec2-76120f6f8c24\") " pod="openstack/ssh-known-hosts-edpm-deployment-k2c8r" Nov 22 08:29:00 crc kubenswrapper[4789]: I1122 08:29:00.633256 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/4cac03ab-292f-4255-aec2-76120f6f8c24-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-k2c8r\" (UID: \"4cac03ab-292f-4255-aec2-76120f6f8c24\") " pod="openstack/ssh-known-hosts-edpm-deployment-k2c8r" Nov 22 08:29:00 crc kubenswrapper[4789]: I1122 08:29:00.644556 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-db29g\" (UniqueName: \"kubernetes.io/projected/4cac03ab-292f-4255-aec2-76120f6f8c24-kube-api-access-db29g\") pod \"ssh-known-hosts-edpm-deployment-k2c8r\" (UID: \"4cac03ab-292f-4255-aec2-76120f6f8c24\") " pod="openstack/ssh-known-hosts-edpm-deployment-k2c8r" Nov 22 08:29:00 crc kubenswrapper[4789]: I1122 08:29:00.793155 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-k2c8r" Nov 22 08:29:01 crc kubenswrapper[4789]: I1122 08:29:01.310814 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-k2c8r"] Nov 22 08:29:01 crc kubenswrapper[4789]: I1122 08:29:01.402623 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-k2c8r" event={"ID":"4cac03ab-292f-4255-aec2-76120f6f8c24","Type":"ContainerStarted","Data":"9bed59e36f789f3603550f2f948d0c3391363c76469f598a85c317482ab40369"} Nov 22 08:29:02 crc kubenswrapper[4789]: I1122 08:29:02.410971 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-k2c8r" event={"ID":"4cac03ab-292f-4255-aec2-76120f6f8c24","Type":"ContainerStarted","Data":"5741378a3aff31bf12bdcc33ca2eb5dda6d02fa7bd0ecc769b25e098f493849c"} Nov 22 08:29:02 crc kubenswrapper[4789]: I1122 08:29:02.434508 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-k2c8r" podStartSLOduration=2.02876965 podStartE2EDuration="2.434490935s" podCreationTimestamp="2025-11-22 08:29:00 +0000 UTC" firstStartedPulling="2025-11-22 08:29:01.316008124 +0000 UTC m=+1955.550408397" lastFinishedPulling="2025-11-22 08:29:01.721729409 +0000 UTC m=+1955.956129682" observedRunningTime="2025-11-22 08:29:02.42767459 +0000 UTC m=+1956.662074863" watchObservedRunningTime="2025-11-22 08:29:02.434490935 +0000 UTC m=+1956.668891208" Nov 22 08:29:09 crc kubenswrapper[4789]: I1122 08:29:09.461140 4789 generic.go:334] "Generic (PLEG): container finished" podID="4cac03ab-292f-4255-aec2-76120f6f8c24" containerID="5741378a3aff31bf12bdcc33ca2eb5dda6d02fa7bd0ecc769b25e098f493849c" exitCode=0 Nov 22 08:29:09 crc kubenswrapper[4789]: I1122 08:29:09.461207 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-k2c8r" event={"ID":"4cac03ab-292f-4255-aec2-76120f6f8c24","Type":"ContainerDied","Data":"5741378a3aff31bf12bdcc33ca2eb5dda6d02fa7bd0ecc769b25e098f493849c"} Nov 22 08:29:10 crc kubenswrapper[4789]: I1122 08:29:10.822710 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-k2c8r" Nov 22 08:29:10 crc kubenswrapper[4789]: I1122 08:29:10.911744 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/4cac03ab-292f-4255-aec2-76120f6f8c24-ssh-key-openstack-edpm-ipam\") pod \"4cac03ab-292f-4255-aec2-76120f6f8c24\" (UID: \"4cac03ab-292f-4255-aec2-76120f6f8c24\") " Nov 22 08:29:10 crc kubenswrapper[4789]: I1122 08:29:10.938313 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4cac03ab-292f-4255-aec2-76120f6f8c24-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "4cac03ab-292f-4255-aec2-76120f6f8c24" (UID: "4cac03ab-292f-4255-aec2-76120f6f8c24"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:29:11 crc kubenswrapper[4789]: I1122 08:29:11.012794 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-db29g\" (UniqueName: \"kubernetes.io/projected/4cac03ab-292f-4255-aec2-76120f6f8c24-kube-api-access-db29g\") pod \"4cac03ab-292f-4255-aec2-76120f6f8c24\" (UID: \"4cac03ab-292f-4255-aec2-76120f6f8c24\") " Nov 22 08:29:11 crc kubenswrapper[4789]: I1122 08:29:11.013697 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/4cac03ab-292f-4255-aec2-76120f6f8c24-inventory-0\") pod \"4cac03ab-292f-4255-aec2-76120f6f8c24\" (UID: \"4cac03ab-292f-4255-aec2-76120f6f8c24\") " Nov 22 08:29:11 crc kubenswrapper[4789]: I1122 08:29:11.015619 4789 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/4cac03ab-292f-4255-aec2-76120f6f8c24-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Nov 22 08:29:11 crc kubenswrapper[4789]: I1122 08:29:11.017011 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4cac03ab-292f-4255-aec2-76120f6f8c24-kube-api-access-db29g" (OuterVolumeSpecName: "kube-api-access-db29g") pod "4cac03ab-292f-4255-aec2-76120f6f8c24" (UID: "4cac03ab-292f-4255-aec2-76120f6f8c24"). InnerVolumeSpecName "kube-api-access-db29g". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:29:11 crc kubenswrapper[4789]: I1122 08:29:11.046743 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4cac03ab-292f-4255-aec2-76120f6f8c24-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "4cac03ab-292f-4255-aec2-76120f6f8c24" (UID: "4cac03ab-292f-4255-aec2-76120f6f8c24"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:29:11 crc kubenswrapper[4789]: I1122 08:29:11.117363 4789 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/4cac03ab-292f-4255-aec2-76120f6f8c24-inventory-0\") on node \"crc\" DevicePath \"\"" Nov 22 08:29:11 crc kubenswrapper[4789]: I1122 08:29:11.117401 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-db29g\" (UniqueName: \"kubernetes.io/projected/4cac03ab-292f-4255-aec2-76120f6f8c24-kube-api-access-db29g\") on node \"crc\" DevicePath \"\"" Nov 22 08:29:11 crc kubenswrapper[4789]: I1122 08:29:11.483592 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-k2c8r" event={"ID":"4cac03ab-292f-4255-aec2-76120f6f8c24","Type":"ContainerDied","Data":"9bed59e36f789f3603550f2f948d0c3391363c76469f598a85c317482ab40369"} Nov 22 08:29:11 crc kubenswrapper[4789]: I1122 08:29:11.483633 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9bed59e36f789f3603550f2f948d0c3391363c76469f598a85c317482ab40369" Nov 22 08:29:11 crc kubenswrapper[4789]: I1122 08:29:11.483707 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-k2c8r" Nov 22 08:29:11 crc kubenswrapper[4789]: I1122 08:29:11.540737 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-xdxkl"] Nov 22 08:29:11 crc kubenswrapper[4789]: E1122 08:29:11.541704 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4cac03ab-292f-4255-aec2-76120f6f8c24" containerName="ssh-known-hosts-edpm-deployment" Nov 22 08:29:11 crc kubenswrapper[4789]: I1122 08:29:11.541730 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="4cac03ab-292f-4255-aec2-76120f6f8c24" containerName="ssh-known-hosts-edpm-deployment" Nov 22 08:29:11 crc kubenswrapper[4789]: I1122 08:29:11.541989 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="4cac03ab-292f-4255-aec2-76120f6f8c24" containerName="ssh-known-hosts-edpm-deployment" Nov 22 08:29:11 crc kubenswrapper[4789]: I1122 08:29:11.542732 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-xdxkl" Nov 22 08:29:11 crc kubenswrapper[4789]: I1122 08:29:11.546211 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-glmr7" Nov 22 08:29:11 crc kubenswrapper[4789]: I1122 08:29:11.546233 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 22 08:29:11 crc kubenswrapper[4789]: I1122 08:29:11.546244 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 22 08:29:11 crc kubenswrapper[4789]: I1122 08:29:11.546555 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 22 08:29:11 crc kubenswrapper[4789]: I1122 08:29:11.548238 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-xdxkl"] Nov 22 08:29:11 crc kubenswrapper[4789]: I1122 08:29:11.624200 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0f452ff9-9dbd-4504-a089-86c16fcd690d-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-xdxkl\" (UID: \"0f452ff9-9dbd-4504-a089-86c16fcd690d\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-xdxkl" Nov 22 08:29:11 crc kubenswrapper[4789]: I1122 08:29:11.624273 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8mtxd\" (UniqueName: \"kubernetes.io/projected/0f452ff9-9dbd-4504-a089-86c16fcd690d-kube-api-access-8mtxd\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-xdxkl\" (UID: \"0f452ff9-9dbd-4504-a089-86c16fcd690d\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-xdxkl" Nov 22 08:29:11 crc kubenswrapper[4789]: I1122 08:29:11.624341 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0f452ff9-9dbd-4504-a089-86c16fcd690d-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-xdxkl\" (UID: \"0f452ff9-9dbd-4504-a089-86c16fcd690d\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-xdxkl" Nov 22 08:29:11 crc kubenswrapper[4789]: I1122 08:29:11.725649 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0f452ff9-9dbd-4504-a089-86c16fcd690d-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-xdxkl\" (UID: \"0f452ff9-9dbd-4504-a089-86c16fcd690d\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-xdxkl" Nov 22 08:29:11 crc kubenswrapper[4789]: I1122 08:29:11.725692 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8mtxd\" (UniqueName: \"kubernetes.io/projected/0f452ff9-9dbd-4504-a089-86c16fcd690d-kube-api-access-8mtxd\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-xdxkl\" (UID: \"0f452ff9-9dbd-4504-a089-86c16fcd690d\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-xdxkl" Nov 22 08:29:11 crc kubenswrapper[4789]: I1122 08:29:11.725760 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0f452ff9-9dbd-4504-a089-86c16fcd690d-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-xdxkl\" (UID: \"0f452ff9-9dbd-4504-a089-86c16fcd690d\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-xdxkl" Nov 22 08:29:11 crc kubenswrapper[4789]: I1122 08:29:11.729362 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0f452ff9-9dbd-4504-a089-86c16fcd690d-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-xdxkl\" (UID: \"0f452ff9-9dbd-4504-a089-86c16fcd690d\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-xdxkl" Nov 22 08:29:11 crc kubenswrapper[4789]: I1122 08:29:11.729379 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0f452ff9-9dbd-4504-a089-86c16fcd690d-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-xdxkl\" (UID: \"0f452ff9-9dbd-4504-a089-86c16fcd690d\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-xdxkl" Nov 22 08:29:11 crc kubenswrapper[4789]: I1122 08:29:11.755588 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8mtxd\" (UniqueName: \"kubernetes.io/projected/0f452ff9-9dbd-4504-a089-86c16fcd690d-kube-api-access-8mtxd\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-xdxkl\" (UID: \"0f452ff9-9dbd-4504-a089-86c16fcd690d\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-xdxkl" Nov 22 08:29:11 crc kubenswrapper[4789]: I1122 08:29:11.866509 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-xdxkl" Nov 22 08:29:12 crc kubenswrapper[4789]: I1122 08:29:12.376199 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-xdxkl"] Nov 22 08:29:12 crc kubenswrapper[4789]: I1122 08:29:12.492009 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-xdxkl" event={"ID":"0f452ff9-9dbd-4504-a089-86c16fcd690d","Type":"ContainerStarted","Data":"dbaffb4406f4413916756a3d8b2e370622b59dda852b4449faa3102224b34ce8"} Nov 22 08:29:13 crc kubenswrapper[4789]: I1122 08:29:13.499600 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-xdxkl" event={"ID":"0f452ff9-9dbd-4504-a089-86c16fcd690d","Type":"ContainerStarted","Data":"099f7b3f545e289261b5f69258034452bfecf664007fa741897156878c82d262"} Nov 22 08:29:13 crc kubenswrapper[4789]: I1122 08:29:13.517102 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-xdxkl" podStartSLOduration=2.042655532 podStartE2EDuration="2.517089585s" podCreationTimestamp="2025-11-22 08:29:11 +0000 UTC" firstStartedPulling="2025-11-22 08:29:12.383433312 +0000 UTC m=+1966.617833585" lastFinishedPulling="2025-11-22 08:29:12.857867365 +0000 UTC m=+1967.092267638" observedRunningTime="2025-11-22 08:29:13.511984997 +0000 UTC m=+1967.746385270" watchObservedRunningTime="2025-11-22 08:29:13.517089585 +0000 UTC m=+1967.751489848" Nov 22 08:29:21 crc kubenswrapper[4789]: I1122 08:29:21.563185 4789 generic.go:334] "Generic (PLEG): container finished" podID="0f452ff9-9dbd-4504-a089-86c16fcd690d" containerID="099f7b3f545e289261b5f69258034452bfecf664007fa741897156878c82d262" exitCode=0 Nov 22 08:29:21 crc kubenswrapper[4789]: I1122 08:29:21.563342 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-xdxkl" event={"ID":"0f452ff9-9dbd-4504-a089-86c16fcd690d","Type":"ContainerDied","Data":"099f7b3f545e289261b5f69258034452bfecf664007fa741897156878c82d262"} Nov 22 08:29:22 crc kubenswrapper[4789]: I1122 08:29:22.943869 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-xdxkl" Nov 22 08:29:23 crc kubenswrapper[4789]: I1122 08:29:23.123416 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8mtxd\" (UniqueName: \"kubernetes.io/projected/0f452ff9-9dbd-4504-a089-86c16fcd690d-kube-api-access-8mtxd\") pod \"0f452ff9-9dbd-4504-a089-86c16fcd690d\" (UID: \"0f452ff9-9dbd-4504-a089-86c16fcd690d\") " Nov 22 08:29:23 crc kubenswrapper[4789]: I1122 08:29:23.123663 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0f452ff9-9dbd-4504-a089-86c16fcd690d-inventory\") pod \"0f452ff9-9dbd-4504-a089-86c16fcd690d\" (UID: \"0f452ff9-9dbd-4504-a089-86c16fcd690d\") " Nov 22 08:29:23 crc kubenswrapper[4789]: I1122 08:29:23.123741 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0f452ff9-9dbd-4504-a089-86c16fcd690d-ssh-key\") pod \"0f452ff9-9dbd-4504-a089-86c16fcd690d\" (UID: \"0f452ff9-9dbd-4504-a089-86c16fcd690d\") " Nov 22 08:29:23 crc kubenswrapper[4789]: I1122 08:29:23.130180 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f452ff9-9dbd-4504-a089-86c16fcd690d-kube-api-access-8mtxd" (OuterVolumeSpecName: "kube-api-access-8mtxd") pod "0f452ff9-9dbd-4504-a089-86c16fcd690d" (UID: "0f452ff9-9dbd-4504-a089-86c16fcd690d"). InnerVolumeSpecName "kube-api-access-8mtxd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:29:23 crc kubenswrapper[4789]: I1122 08:29:23.156274 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f452ff9-9dbd-4504-a089-86c16fcd690d-inventory" (OuterVolumeSpecName: "inventory") pod "0f452ff9-9dbd-4504-a089-86c16fcd690d" (UID: "0f452ff9-9dbd-4504-a089-86c16fcd690d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:29:23 crc kubenswrapper[4789]: I1122 08:29:23.156826 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f452ff9-9dbd-4504-a089-86c16fcd690d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "0f452ff9-9dbd-4504-a089-86c16fcd690d" (UID: "0f452ff9-9dbd-4504-a089-86c16fcd690d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:29:23 crc kubenswrapper[4789]: I1122 08:29:23.225442 4789 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0f452ff9-9dbd-4504-a089-86c16fcd690d-inventory\") on node \"crc\" DevicePath \"\"" Nov 22 08:29:23 crc kubenswrapper[4789]: I1122 08:29:23.225476 4789 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0f452ff9-9dbd-4504-a089-86c16fcd690d-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 22 08:29:23 crc kubenswrapper[4789]: I1122 08:29:23.225487 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8mtxd\" (UniqueName: \"kubernetes.io/projected/0f452ff9-9dbd-4504-a089-86c16fcd690d-kube-api-access-8mtxd\") on node \"crc\" DevicePath \"\"" Nov 22 08:29:23 crc kubenswrapper[4789]: I1122 08:29:23.580729 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-xdxkl" event={"ID":"0f452ff9-9dbd-4504-a089-86c16fcd690d","Type":"ContainerDied","Data":"dbaffb4406f4413916756a3d8b2e370622b59dda852b4449faa3102224b34ce8"} Nov 22 08:29:23 crc kubenswrapper[4789]: I1122 08:29:23.580984 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dbaffb4406f4413916756a3d8b2e370622b59dda852b4449faa3102224b34ce8" Nov 22 08:29:23 crc kubenswrapper[4789]: I1122 08:29:23.580793 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-xdxkl" Nov 22 08:29:23 crc kubenswrapper[4789]: I1122 08:29:23.661743 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7qscl"] Nov 22 08:29:23 crc kubenswrapper[4789]: E1122 08:29:23.662085 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f452ff9-9dbd-4504-a089-86c16fcd690d" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 22 08:29:23 crc kubenswrapper[4789]: I1122 08:29:23.662102 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f452ff9-9dbd-4504-a089-86c16fcd690d" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 22 08:29:23 crc kubenswrapper[4789]: I1122 08:29:23.662290 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f452ff9-9dbd-4504-a089-86c16fcd690d" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 22 08:29:23 crc kubenswrapper[4789]: I1122 08:29:23.662998 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7qscl" Nov 22 08:29:23 crc kubenswrapper[4789]: I1122 08:29:23.664940 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 22 08:29:23 crc kubenswrapper[4789]: I1122 08:29:23.665146 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-glmr7" Nov 22 08:29:23 crc kubenswrapper[4789]: I1122 08:29:23.665417 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 22 08:29:23 crc kubenswrapper[4789]: I1122 08:29:23.668926 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 22 08:29:23 crc kubenswrapper[4789]: I1122 08:29:23.674808 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7qscl"] Nov 22 08:29:23 crc kubenswrapper[4789]: I1122 08:29:23.734556 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fs4fv\" (UniqueName: \"kubernetes.io/projected/393d8a04-e1ff-4707-bf88-964105ed2008-kube-api-access-fs4fv\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-7qscl\" (UID: \"393d8a04-e1ff-4707-bf88-964105ed2008\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7qscl" Nov 22 08:29:23 crc kubenswrapper[4789]: I1122 08:29:23.734948 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/393d8a04-e1ff-4707-bf88-964105ed2008-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-7qscl\" (UID: \"393d8a04-e1ff-4707-bf88-964105ed2008\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7qscl" Nov 22 08:29:23 crc kubenswrapper[4789]: I1122 08:29:23.735109 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/393d8a04-e1ff-4707-bf88-964105ed2008-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-7qscl\" (UID: \"393d8a04-e1ff-4707-bf88-964105ed2008\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7qscl" Nov 22 08:29:23 crc kubenswrapper[4789]: I1122 08:29:23.837788 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fs4fv\" (UniqueName: \"kubernetes.io/projected/393d8a04-e1ff-4707-bf88-964105ed2008-kube-api-access-fs4fv\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-7qscl\" (UID: \"393d8a04-e1ff-4707-bf88-964105ed2008\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7qscl" Nov 22 08:29:23 crc kubenswrapper[4789]: I1122 08:29:23.838492 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/393d8a04-e1ff-4707-bf88-964105ed2008-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-7qscl\" (UID: \"393d8a04-e1ff-4707-bf88-964105ed2008\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7qscl" Nov 22 08:29:23 crc kubenswrapper[4789]: I1122 08:29:23.838589 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/393d8a04-e1ff-4707-bf88-964105ed2008-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-7qscl\" (UID: \"393d8a04-e1ff-4707-bf88-964105ed2008\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7qscl" Nov 22 08:29:23 crc kubenswrapper[4789]: I1122 08:29:23.845210 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/393d8a04-e1ff-4707-bf88-964105ed2008-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-7qscl\" (UID: \"393d8a04-e1ff-4707-bf88-964105ed2008\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7qscl" Nov 22 08:29:23 crc kubenswrapper[4789]: I1122 08:29:23.845472 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/393d8a04-e1ff-4707-bf88-964105ed2008-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-7qscl\" (UID: \"393d8a04-e1ff-4707-bf88-964105ed2008\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7qscl" Nov 22 08:29:23 crc kubenswrapper[4789]: I1122 08:29:23.855407 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fs4fv\" (UniqueName: \"kubernetes.io/projected/393d8a04-e1ff-4707-bf88-964105ed2008-kube-api-access-fs4fv\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-7qscl\" (UID: \"393d8a04-e1ff-4707-bf88-964105ed2008\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7qscl" Nov 22 08:29:23 crc kubenswrapper[4789]: I1122 08:29:23.981248 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7qscl" Nov 22 08:29:24 crc kubenswrapper[4789]: I1122 08:29:24.499930 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7qscl"] Nov 22 08:29:24 crc kubenswrapper[4789]: I1122 08:29:24.589779 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7qscl" event={"ID":"393d8a04-e1ff-4707-bf88-964105ed2008","Type":"ContainerStarted","Data":"0dfc0dfbb331da4d18c83835b4eba556445c9f869a01d6c3c4f25f3339cfc77c"} Nov 22 08:29:25 crc kubenswrapper[4789]: I1122 08:29:25.599059 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7qscl" event={"ID":"393d8a04-e1ff-4707-bf88-964105ed2008","Type":"ContainerStarted","Data":"f3cc3f695f3c7ebe4b186d126800d634b396ab3b012164684186dce8bcc05b10"} Nov 22 08:29:25 crc kubenswrapper[4789]: I1122 08:29:25.620684 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7qscl" podStartSLOduration=2.199452525 podStartE2EDuration="2.620655472s" podCreationTimestamp="2025-11-22 08:29:23 +0000 UTC" firstStartedPulling="2025-11-22 08:29:24.502708635 +0000 UTC m=+1978.737108908" lastFinishedPulling="2025-11-22 08:29:24.923911592 +0000 UTC m=+1979.158311855" observedRunningTime="2025-11-22 08:29:25.612843229 +0000 UTC m=+1979.847243512" watchObservedRunningTime="2025-11-22 08:29:25.620655472 +0000 UTC m=+1979.855055745" Nov 22 08:29:26 crc kubenswrapper[4789]: I1122 08:29:26.205245 4789 scope.go:117] "RemoveContainer" containerID="8f13b0868aa5253a8993b673578603567f018491fca6b9f2caa8e7ea94d33934" Nov 22 08:29:26 crc kubenswrapper[4789]: I1122 08:29:26.249350 4789 scope.go:117] "RemoveContainer" containerID="7128ef0b5659bd07e32c2171483c5a7e4125643355ad974d47dde4aa001b1fec" Nov 22 08:29:26 crc kubenswrapper[4789]: I1122 08:29:26.303504 4789 scope.go:117] "RemoveContainer" containerID="1d99062b301747b7ae1bd0307943d6d1d286baa1210d5067e5b0ada008972cf3" Nov 22 08:29:35 crc kubenswrapper[4789]: I1122 08:29:35.709537 4789 generic.go:334] "Generic (PLEG): container finished" podID="393d8a04-e1ff-4707-bf88-964105ed2008" containerID="f3cc3f695f3c7ebe4b186d126800d634b396ab3b012164684186dce8bcc05b10" exitCode=0 Nov 22 08:29:35 crc kubenswrapper[4789]: I1122 08:29:35.709639 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7qscl" event={"ID":"393d8a04-e1ff-4707-bf88-964105ed2008","Type":"ContainerDied","Data":"f3cc3f695f3c7ebe4b186d126800d634b396ab3b012164684186dce8bcc05b10"} Nov 22 08:29:37 crc kubenswrapper[4789]: I1122 08:29:37.044665 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-c94ng"] Nov 22 08:29:37 crc kubenswrapper[4789]: I1122 08:29:37.054174 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-c94ng"] Nov 22 08:29:37 crc kubenswrapper[4789]: I1122 08:29:37.116707 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7qscl" Nov 22 08:29:37 crc kubenswrapper[4789]: I1122 08:29:37.288589 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fs4fv\" (UniqueName: \"kubernetes.io/projected/393d8a04-e1ff-4707-bf88-964105ed2008-kube-api-access-fs4fv\") pod \"393d8a04-e1ff-4707-bf88-964105ed2008\" (UID: \"393d8a04-e1ff-4707-bf88-964105ed2008\") " Nov 22 08:29:37 crc kubenswrapper[4789]: I1122 08:29:37.288691 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/393d8a04-e1ff-4707-bf88-964105ed2008-ssh-key\") pod \"393d8a04-e1ff-4707-bf88-964105ed2008\" (UID: \"393d8a04-e1ff-4707-bf88-964105ed2008\") " Nov 22 08:29:37 crc kubenswrapper[4789]: I1122 08:29:37.288735 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/393d8a04-e1ff-4707-bf88-964105ed2008-inventory\") pod \"393d8a04-e1ff-4707-bf88-964105ed2008\" (UID: \"393d8a04-e1ff-4707-bf88-964105ed2008\") " Nov 22 08:29:37 crc kubenswrapper[4789]: I1122 08:29:37.309179 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/393d8a04-e1ff-4707-bf88-964105ed2008-kube-api-access-fs4fv" (OuterVolumeSpecName: "kube-api-access-fs4fv") pod "393d8a04-e1ff-4707-bf88-964105ed2008" (UID: "393d8a04-e1ff-4707-bf88-964105ed2008"). InnerVolumeSpecName "kube-api-access-fs4fv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:29:37 crc kubenswrapper[4789]: I1122 08:29:37.315685 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/393d8a04-e1ff-4707-bf88-964105ed2008-inventory" (OuterVolumeSpecName: "inventory") pod "393d8a04-e1ff-4707-bf88-964105ed2008" (UID: "393d8a04-e1ff-4707-bf88-964105ed2008"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:29:37 crc kubenswrapper[4789]: I1122 08:29:37.318043 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/393d8a04-e1ff-4707-bf88-964105ed2008-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "393d8a04-e1ff-4707-bf88-964105ed2008" (UID: "393d8a04-e1ff-4707-bf88-964105ed2008"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:29:37 crc kubenswrapper[4789]: I1122 08:29:37.390829 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fs4fv\" (UniqueName: \"kubernetes.io/projected/393d8a04-e1ff-4707-bf88-964105ed2008-kube-api-access-fs4fv\") on node \"crc\" DevicePath \"\"" Nov 22 08:29:37 crc kubenswrapper[4789]: I1122 08:29:37.390871 4789 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/393d8a04-e1ff-4707-bf88-964105ed2008-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 22 08:29:37 crc kubenswrapper[4789]: I1122 08:29:37.390881 4789 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/393d8a04-e1ff-4707-bf88-964105ed2008-inventory\") on node \"crc\" DevicePath \"\"" Nov 22 08:29:37 crc kubenswrapper[4789]: I1122 08:29:37.728416 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7qscl" event={"ID":"393d8a04-e1ff-4707-bf88-964105ed2008","Type":"ContainerDied","Data":"0dfc0dfbb331da4d18c83835b4eba556445c9f869a01d6c3c4f25f3339cfc77c"} Nov 22 08:29:37 crc kubenswrapper[4789]: I1122 08:29:37.728456 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0dfc0dfbb331da4d18c83835b4eba556445c9f869a01d6c3c4f25f3339cfc77c" Nov 22 08:29:37 crc kubenswrapper[4789]: I1122 08:29:37.728513 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7qscl" Nov 22 08:29:37 crc kubenswrapper[4789]: I1122 08:29:37.978470 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="206a90d7-1c1c-4a17-992a-43bd718fc9ea" path="/var/lib/kubelet/pods/206a90d7-1c1c-4a17-992a-43bd718fc9ea/volumes" Nov 22 08:30:00 crc kubenswrapper[4789]: I1122 08:30:00.164757 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396670-ltqvr"] Nov 22 08:30:00 crc kubenswrapper[4789]: E1122 08:30:00.165839 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="393d8a04-e1ff-4707-bf88-964105ed2008" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 22 08:30:00 crc kubenswrapper[4789]: I1122 08:30:00.165853 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="393d8a04-e1ff-4707-bf88-964105ed2008" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 22 08:30:00 crc kubenswrapper[4789]: I1122 08:30:00.166181 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="393d8a04-e1ff-4707-bf88-964105ed2008" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 22 08:30:00 crc kubenswrapper[4789]: I1122 08:30:00.167017 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396670-ltqvr" Nov 22 08:30:00 crc kubenswrapper[4789]: I1122 08:30:00.173522 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 22 08:30:00 crc kubenswrapper[4789]: I1122 08:30:00.175034 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 22 08:30:00 crc kubenswrapper[4789]: I1122 08:30:00.195252 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396670-ltqvr"] Nov 22 08:30:00 crc kubenswrapper[4789]: I1122 08:30:00.277941 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tgwkb\" (UniqueName: \"kubernetes.io/projected/37c2f280-272f-45f0-993e-7d840ebf3b92-kube-api-access-tgwkb\") pod \"collect-profiles-29396670-ltqvr\" (UID: \"37c2f280-272f-45f0-993e-7d840ebf3b92\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396670-ltqvr" Nov 22 08:30:00 crc kubenswrapper[4789]: I1122 08:30:00.278429 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/37c2f280-272f-45f0-993e-7d840ebf3b92-config-volume\") pod \"collect-profiles-29396670-ltqvr\" (UID: \"37c2f280-272f-45f0-993e-7d840ebf3b92\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396670-ltqvr" Nov 22 08:30:00 crc kubenswrapper[4789]: I1122 08:30:00.278485 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/37c2f280-272f-45f0-993e-7d840ebf3b92-secret-volume\") pod \"collect-profiles-29396670-ltqvr\" (UID: \"37c2f280-272f-45f0-993e-7d840ebf3b92\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396670-ltqvr" Nov 22 08:30:00 crc kubenswrapper[4789]: I1122 08:30:00.380082 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tgwkb\" (UniqueName: \"kubernetes.io/projected/37c2f280-272f-45f0-993e-7d840ebf3b92-kube-api-access-tgwkb\") pod \"collect-profiles-29396670-ltqvr\" (UID: \"37c2f280-272f-45f0-993e-7d840ebf3b92\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396670-ltqvr" Nov 22 08:30:00 crc kubenswrapper[4789]: I1122 08:30:00.380272 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/37c2f280-272f-45f0-993e-7d840ebf3b92-config-volume\") pod \"collect-profiles-29396670-ltqvr\" (UID: \"37c2f280-272f-45f0-993e-7d840ebf3b92\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396670-ltqvr" Nov 22 08:30:00 crc kubenswrapper[4789]: I1122 08:30:00.380303 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/37c2f280-272f-45f0-993e-7d840ebf3b92-secret-volume\") pod \"collect-profiles-29396670-ltqvr\" (UID: \"37c2f280-272f-45f0-993e-7d840ebf3b92\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396670-ltqvr" Nov 22 08:30:00 crc kubenswrapper[4789]: I1122 08:30:00.381237 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/37c2f280-272f-45f0-993e-7d840ebf3b92-config-volume\") pod \"collect-profiles-29396670-ltqvr\" (UID: \"37c2f280-272f-45f0-993e-7d840ebf3b92\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396670-ltqvr" Nov 22 08:30:00 crc kubenswrapper[4789]: I1122 08:30:00.387976 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/37c2f280-272f-45f0-993e-7d840ebf3b92-secret-volume\") pod \"collect-profiles-29396670-ltqvr\" (UID: \"37c2f280-272f-45f0-993e-7d840ebf3b92\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396670-ltqvr" Nov 22 08:30:00 crc kubenswrapper[4789]: I1122 08:30:00.397722 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tgwkb\" (UniqueName: \"kubernetes.io/projected/37c2f280-272f-45f0-993e-7d840ebf3b92-kube-api-access-tgwkb\") pod \"collect-profiles-29396670-ltqvr\" (UID: \"37c2f280-272f-45f0-993e-7d840ebf3b92\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396670-ltqvr" Nov 22 08:30:00 crc kubenswrapper[4789]: I1122 08:30:00.498587 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396670-ltqvr" Nov 22 08:30:00 crc kubenswrapper[4789]: I1122 08:30:00.924490 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396670-ltqvr"] Nov 22 08:30:00 crc kubenswrapper[4789]: I1122 08:30:00.949701 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29396670-ltqvr" event={"ID":"37c2f280-272f-45f0-993e-7d840ebf3b92","Type":"ContainerStarted","Data":"9acf811bf075b52c6dda9a644803751cdb8a91328580455f6d47d7f206085daa"} Nov 22 08:30:01 crc kubenswrapper[4789]: I1122 08:30:01.959117 4789 generic.go:334] "Generic (PLEG): container finished" podID="37c2f280-272f-45f0-993e-7d840ebf3b92" containerID="f9252464a45b6db0e11c046144f08da1b92ba249212199e6a571d3d7c1b15f13" exitCode=0 Nov 22 08:30:01 crc kubenswrapper[4789]: I1122 08:30:01.959212 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29396670-ltqvr" event={"ID":"37c2f280-272f-45f0-993e-7d840ebf3b92","Type":"ContainerDied","Data":"f9252464a45b6db0e11c046144f08da1b92ba249212199e6a571d3d7c1b15f13"} Nov 22 08:30:03 crc kubenswrapper[4789]: I1122 08:30:03.278621 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396670-ltqvr" Nov 22 08:30:03 crc kubenswrapper[4789]: I1122 08:30:03.331241 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/37c2f280-272f-45f0-993e-7d840ebf3b92-secret-volume\") pod \"37c2f280-272f-45f0-993e-7d840ebf3b92\" (UID: \"37c2f280-272f-45f0-993e-7d840ebf3b92\") " Nov 22 08:30:03 crc kubenswrapper[4789]: I1122 08:30:03.331439 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tgwkb\" (UniqueName: \"kubernetes.io/projected/37c2f280-272f-45f0-993e-7d840ebf3b92-kube-api-access-tgwkb\") pod \"37c2f280-272f-45f0-993e-7d840ebf3b92\" (UID: \"37c2f280-272f-45f0-993e-7d840ebf3b92\") " Nov 22 08:30:03 crc kubenswrapper[4789]: I1122 08:30:03.331472 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/37c2f280-272f-45f0-993e-7d840ebf3b92-config-volume\") pod \"37c2f280-272f-45f0-993e-7d840ebf3b92\" (UID: \"37c2f280-272f-45f0-993e-7d840ebf3b92\") " Nov 22 08:30:03 crc kubenswrapper[4789]: I1122 08:30:03.332223 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/37c2f280-272f-45f0-993e-7d840ebf3b92-config-volume" (OuterVolumeSpecName: "config-volume") pod "37c2f280-272f-45f0-993e-7d840ebf3b92" (UID: "37c2f280-272f-45f0-993e-7d840ebf3b92"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:30:03 crc kubenswrapper[4789]: I1122 08:30:03.336983 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37c2f280-272f-45f0-993e-7d840ebf3b92-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "37c2f280-272f-45f0-993e-7d840ebf3b92" (UID: "37c2f280-272f-45f0-993e-7d840ebf3b92"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:30:03 crc kubenswrapper[4789]: I1122 08:30:03.337140 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37c2f280-272f-45f0-993e-7d840ebf3b92-kube-api-access-tgwkb" (OuterVolumeSpecName: "kube-api-access-tgwkb") pod "37c2f280-272f-45f0-993e-7d840ebf3b92" (UID: "37c2f280-272f-45f0-993e-7d840ebf3b92"). InnerVolumeSpecName "kube-api-access-tgwkb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:30:03 crc kubenswrapper[4789]: I1122 08:30:03.433681 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tgwkb\" (UniqueName: \"kubernetes.io/projected/37c2f280-272f-45f0-993e-7d840ebf3b92-kube-api-access-tgwkb\") on node \"crc\" DevicePath \"\"" Nov 22 08:30:03 crc kubenswrapper[4789]: I1122 08:30:03.433726 4789 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/37c2f280-272f-45f0-993e-7d840ebf3b92-config-volume\") on node \"crc\" DevicePath \"\"" Nov 22 08:30:03 crc kubenswrapper[4789]: I1122 08:30:03.433738 4789 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/37c2f280-272f-45f0-993e-7d840ebf3b92-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 22 08:30:03 crc kubenswrapper[4789]: I1122 08:30:03.986489 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29396670-ltqvr" event={"ID":"37c2f280-272f-45f0-993e-7d840ebf3b92","Type":"ContainerDied","Data":"9acf811bf075b52c6dda9a644803751cdb8a91328580455f6d47d7f206085daa"} Nov 22 08:30:03 crc kubenswrapper[4789]: I1122 08:30:03.986856 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9acf811bf075b52c6dda9a644803751cdb8a91328580455f6d47d7f206085daa" Nov 22 08:30:03 crc kubenswrapper[4789]: I1122 08:30:03.986933 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396670-ltqvr" Nov 22 08:30:04 crc kubenswrapper[4789]: I1122 08:30:04.341280 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396625-nrhcv"] Nov 22 08:30:04 crc kubenswrapper[4789]: I1122 08:30:04.347481 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396625-nrhcv"] Nov 22 08:30:05 crc kubenswrapper[4789]: I1122 08:30:05.372611 4789 patch_prober.go:28] interesting pod/machine-config-daemon-pxvrm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 08:30:05 crc kubenswrapper[4789]: I1122 08:30:05.372690 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 08:30:05 crc kubenswrapper[4789]: I1122 08:30:05.978834 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ac3d1daa-cb6d-41fc-9798-e687d9b2a21d" path="/var/lib/kubelet/pods/ac3d1daa-cb6d-41fc-9798-e687d9b2a21d/volumes" Nov 22 08:30:26 crc kubenswrapper[4789]: I1122 08:30:26.409454 4789 scope.go:117] "RemoveContainer" containerID="5cc6b25550f1ec45cf8e7b0340383254d666cc9cb3aa291b0a9b7d2543f04a4e" Nov 22 08:30:26 crc kubenswrapper[4789]: I1122 08:30:26.443128 4789 scope.go:117] "RemoveContainer" containerID="380143d9ea6b3ab19e351a49cb7d0db0752bada11ce7974131318d6c6d340ff3" Nov 22 08:30:35 crc kubenswrapper[4789]: I1122 08:30:35.371931 4789 patch_prober.go:28] interesting pod/machine-config-daemon-pxvrm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 08:30:35 crc kubenswrapper[4789]: I1122 08:30:35.373074 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 08:30:51 crc kubenswrapper[4789]: I1122 08:30:51.369785 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-dvlgg"] Nov 22 08:30:51 crc kubenswrapper[4789]: E1122 08:30:51.370933 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37c2f280-272f-45f0-993e-7d840ebf3b92" containerName="collect-profiles" Nov 22 08:30:51 crc kubenswrapper[4789]: I1122 08:30:51.370953 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="37c2f280-272f-45f0-993e-7d840ebf3b92" containerName="collect-profiles" Nov 22 08:30:51 crc kubenswrapper[4789]: I1122 08:30:51.371159 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="37c2f280-272f-45f0-993e-7d840ebf3b92" containerName="collect-profiles" Nov 22 08:30:51 crc kubenswrapper[4789]: I1122 08:30:51.372737 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dvlgg" Nov 22 08:30:51 crc kubenswrapper[4789]: I1122 08:30:51.385819 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dvlgg"] Nov 22 08:30:51 crc kubenswrapper[4789]: I1122 08:30:51.555699 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e70319d4-4101-4ceb-80a3-b615df927a18-utilities\") pod \"redhat-operators-dvlgg\" (UID: \"e70319d4-4101-4ceb-80a3-b615df927a18\") " pod="openshift-marketplace/redhat-operators-dvlgg" Nov 22 08:30:51 crc kubenswrapper[4789]: I1122 08:30:51.556224 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4h585\" (UniqueName: \"kubernetes.io/projected/e70319d4-4101-4ceb-80a3-b615df927a18-kube-api-access-4h585\") pod \"redhat-operators-dvlgg\" (UID: \"e70319d4-4101-4ceb-80a3-b615df927a18\") " pod="openshift-marketplace/redhat-operators-dvlgg" Nov 22 08:30:51 crc kubenswrapper[4789]: I1122 08:30:51.556412 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e70319d4-4101-4ceb-80a3-b615df927a18-catalog-content\") pod \"redhat-operators-dvlgg\" (UID: \"e70319d4-4101-4ceb-80a3-b615df927a18\") " pod="openshift-marketplace/redhat-operators-dvlgg" Nov 22 08:30:51 crc kubenswrapper[4789]: I1122 08:30:51.658839 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4h585\" (UniqueName: \"kubernetes.io/projected/e70319d4-4101-4ceb-80a3-b615df927a18-kube-api-access-4h585\") pod \"redhat-operators-dvlgg\" (UID: \"e70319d4-4101-4ceb-80a3-b615df927a18\") " pod="openshift-marketplace/redhat-operators-dvlgg" Nov 22 08:30:51 crc kubenswrapper[4789]: I1122 08:30:51.658895 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e70319d4-4101-4ceb-80a3-b615df927a18-catalog-content\") pod \"redhat-operators-dvlgg\" (UID: \"e70319d4-4101-4ceb-80a3-b615df927a18\") " pod="openshift-marketplace/redhat-operators-dvlgg" Nov 22 08:30:51 crc kubenswrapper[4789]: I1122 08:30:51.659012 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e70319d4-4101-4ceb-80a3-b615df927a18-utilities\") pod \"redhat-operators-dvlgg\" (UID: \"e70319d4-4101-4ceb-80a3-b615df927a18\") " pod="openshift-marketplace/redhat-operators-dvlgg" Nov 22 08:30:51 crc kubenswrapper[4789]: I1122 08:30:51.659448 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e70319d4-4101-4ceb-80a3-b615df927a18-catalog-content\") pod \"redhat-operators-dvlgg\" (UID: \"e70319d4-4101-4ceb-80a3-b615df927a18\") " pod="openshift-marketplace/redhat-operators-dvlgg" Nov 22 08:30:51 crc kubenswrapper[4789]: I1122 08:30:51.659527 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e70319d4-4101-4ceb-80a3-b615df927a18-utilities\") pod \"redhat-operators-dvlgg\" (UID: \"e70319d4-4101-4ceb-80a3-b615df927a18\") " pod="openshift-marketplace/redhat-operators-dvlgg" Nov 22 08:30:51 crc kubenswrapper[4789]: I1122 08:30:51.681279 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4h585\" (UniqueName: \"kubernetes.io/projected/e70319d4-4101-4ceb-80a3-b615df927a18-kube-api-access-4h585\") pod \"redhat-operators-dvlgg\" (UID: \"e70319d4-4101-4ceb-80a3-b615df927a18\") " pod="openshift-marketplace/redhat-operators-dvlgg" Nov 22 08:30:51 crc kubenswrapper[4789]: I1122 08:30:51.708108 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dvlgg" Nov 22 08:30:52 crc kubenswrapper[4789]: I1122 08:30:52.162545 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dvlgg"] Nov 22 08:30:52 crc kubenswrapper[4789]: I1122 08:30:52.414241 4789 generic.go:334] "Generic (PLEG): container finished" podID="e70319d4-4101-4ceb-80a3-b615df927a18" containerID="8dcfe552980a30fdcb621b3034b6e2afb955d77d9be9763829f46845ddfd0a57" exitCode=0 Nov 22 08:30:52 crc kubenswrapper[4789]: I1122 08:30:52.414310 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dvlgg" event={"ID":"e70319d4-4101-4ceb-80a3-b615df927a18","Type":"ContainerDied","Data":"8dcfe552980a30fdcb621b3034b6e2afb955d77d9be9763829f46845ddfd0a57"} Nov 22 08:30:52 crc kubenswrapper[4789]: I1122 08:30:52.414586 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dvlgg" event={"ID":"e70319d4-4101-4ceb-80a3-b615df927a18","Type":"ContainerStarted","Data":"5cf99c1c403112a32d78783bfe09e3b006b68f2103103db594a73e2711cfc507"} Nov 22 08:30:52 crc kubenswrapper[4789]: I1122 08:30:52.416151 4789 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 22 08:30:53 crc kubenswrapper[4789]: I1122 08:30:53.425049 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dvlgg" event={"ID":"e70319d4-4101-4ceb-80a3-b615df927a18","Type":"ContainerStarted","Data":"dfba41882e69b545a7add9897c427677d5141b54e729458b4ec32cade78d1d90"} Nov 22 08:30:54 crc kubenswrapper[4789]: I1122 08:30:54.438159 4789 generic.go:334] "Generic (PLEG): container finished" podID="e70319d4-4101-4ceb-80a3-b615df927a18" containerID="dfba41882e69b545a7add9897c427677d5141b54e729458b4ec32cade78d1d90" exitCode=0 Nov 22 08:30:54 crc kubenswrapper[4789]: I1122 08:30:54.438240 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dvlgg" event={"ID":"e70319d4-4101-4ceb-80a3-b615df927a18","Type":"ContainerDied","Data":"dfba41882e69b545a7add9897c427677d5141b54e729458b4ec32cade78d1d90"} Nov 22 08:30:56 crc kubenswrapper[4789]: I1122 08:30:56.461239 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dvlgg" event={"ID":"e70319d4-4101-4ceb-80a3-b615df927a18","Type":"ContainerStarted","Data":"660d54756912e4543f81365c1b5d691d57d8c8ac368d61de10db6920932ad699"} Nov 22 08:30:56 crc kubenswrapper[4789]: I1122 08:30:56.489420 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-dvlgg" podStartSLOduration=2.796852151 podStartE2EDuration="5.489393704s" podCreationTimestamp="2025-11-22 08:30:51 +0000 UTC" firstStartedPulling="2025-11-22 08:30:52.415931223 +0000 UTC m=+2066.650331496" lastFinishedPulling="2025-11-22 08:30:55.108472756 +0000 UTC m=+2069.342873049" observedRunningTime="2025-11-22 08:30:56.481175381 +0000 UTC m=+2070.715575654" watchObservedRunningTime="2025-11-22 08:30:56.489393704 +0000 UTC m=+2070.723793977" Nov 22 08:31:00 crc kubenswrapper[4789]: I1122 08:31:00.061294 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-vfzsg"] Nov 22 08:31:00 crc kubenswrapper[4789]: I1122 08:31:00.063577 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vfzsg" Nov 22 08:31:00 crc kubenswrapper[4789]: I1122 08:31:00.077700 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vfzsg"] Nov 22 08:31:00 crc kubenswrapper[4789]: I1122 08:31:00.215690 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/261de000-bcb1-4c6a-858c-ea3254254d0c-utilities\") pod \"community-operators-vfzsg\" (UID: \"261de000-bcb1-4c6a-858c-ea3254254d0c\") " pod="openshift-marketplace/community-operators-vfzsg" Nov 22 08:31:00 crc kubenswrapper[4789]: I1122 08:31:00.215930 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/261de000-bcb1-4c6a-858c-ea3254254d0c-catalog-content\") pod \"community-operators-vfzsg\" (UID: \"261de000-bcb1-4c6a-858c-ea3254254d0c\") " pod="openshift-marketplace/community-operators-vfzsg" Nov 22 08:31:00 crc kubenswrapper[4789]: I1122 08:31:00.216019 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qqvm2\" (UniqueName: \"kubernetes.io/projected/261de000-bcb1-4c6a-858c-ea3254254d0c-kube-api-access-qqvm2\") pod \"community-operators-vfzsg\" (UID: \"261de000-bcb1-4c6a-858c-ea3254254d0c\") " pod="openshift-marketplace/community-operators-vfzsg" Nov 22 08:31:00 crc kubenswrapper[4789]: I1122 08:31:00.317537 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/261de000-bcb1-4c6a-858c-ea3254254d0c-utilities\") pod \"community-operators-vfzsg\" (UID: \"261de000-bcb1-4c6a-858c-ea3254254d0c\") " pod="openshift-marketplace/community-operators-vfzsg" Nov 22 08:31:00 crc kubenswrapper[4789]: I1122 08:31:00.317615 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/261de000-bcb1-4c6a-858c-ea3254254d0c-catalog-content\") pod \"community-operators-vfzsg\" (UID: \"261de000-bcb1-4c6a-858c-ea3254254d0c\") " pod="openshift-marketplace/community-operators-vfzsg" Nov 22 08:31:00 crc kubenswrapper[4789]: I1122 08:31:00.317667 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qqvm2\" (UniqueName: \"kubernetes.io/projected/261de000-bcb1-4c6a-858c-ea3254254d0c-kube-api-access-qqvm2\") pod \"community-operators-vfzsg\" (UID: \"261de000-bcb1-4c6a-858c-ea3254254d0c\") " pod="openshift-marketplace/community-operators-vfzsg" Nov 22 08:31:00 crc kubenswrapper[4789]: I1122 08:31:00.318106 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/261de000-bcb1-4c6a-858c-ea3254254d0c-utilities\") pod \"community-operators-vfzsg\" (UID: \"261de000-bcb1-4c6a-858c-ea3254254d0c\") " pod="openshift-marketplace/community-operators-vfzsg" Nov 22 08:31:00 crc kubenswrapper[4789]: I1122 08:31:00.318222 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/261de000-bcb1-4c6a-858c-ea3254254d0c-catalog-content\") pod \"community-operators-vfzsg\" (UID: \"261de000-bcb1-4c6a-858c-ea3254254d0c\") " pod="openshift-marketplace/community-operators-vfzsg" Nov 22 08:31:00 crc kubenswrapper[4789]: I1122 08:31:00.337425 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qqvm2\" (UniqueName: \"kubernetes.io/projected/261de000-bcb1-4c6a-858c-ea3254254d0c-kube-api-access-qqvm2\") pod \"community-operators-vfzsg\" (UID: \"261de000-bcb1-4c6a-858c-ea3254254d0c\") " pod="openshift-marketplace/community-operators-vfzsg" Nov 22 08:31:00 crc kubenswrapper[4789]: I1122 08:31:00.385530 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vfzsg" Nov 22 08:31:00 crc kubenswrapper[4789]: I1122 08:31:00.927174 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vfzsg"] Nov 22 08:31:01 crc kubenswrapper[4789]: I1122 08:31:01.511463 4789 generic.go:334] "Generic (PLEG): container finished" podID="261de000-bcb1-4c6a-858c-ea3254254d0c" containerID="72eb86159b377c03739718819dcaa6ec61c19f9ec6cc06033dca2146af456bac" exitCode=0 Nov 22 08:31:01 crc kubenswrapper[4789]: I1122 08:31:01.511535 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vfzsg" event={"ID":"261de000-bcb1-4c6a-858c-ea3254254d0c","Type":"ContainerDied","Data":"72eb86159b377c03739718819dcaa6ec61c19f9ec6cc06033dca2146af456bac"} Nov 22 08:31:01 crc kubenswrapper[4789]: I1122 08:31:01.512182 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vfzsg" event={"ID":"261de000-bcb1-4c6a-858c-ea3254254d0c","Type":"ContainerStarted","Data":"cc829ca05a2a7ccd05d4ecbce8398b13472c8f45109d695ff1dccdaaa06723b9"} Nov 22 08:31:01 crc kubenswrapper[4789]: I1122 08:31:01.709042 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-dvlgg" Nov 22 08:31:01 crc kubenswrapper[4789]: I1122 08:31:01.709118 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-dvlgg" Nov 22 08:31:02 crc kubenswrapper[4789]: I1122 08:31:02.522551 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vfzsg" event={"ID":"261de000-bcb1-4c6a-858c-ea3254254d0c","Type":"ContainerStarted","Data":"f4961e0f4acd5e44db666766aea5ae467e1152770dca462e64af8ee202061e12"} Nov 22 08:31:02 crc kubenswrapper[4789]: I1122 08:31:02.752019 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-dvlgg" podUID="e70319d4-4101-4ceb-80a3-b615df927a18" containerName="registry-server" probeResult="failure" output=< Nov 22 08:31:02 crc kubenswrapper[4789]: timeout: failed to connect service ":50051" within 1s Nov 22 08:31:02 crc kubenswrapper[4789]: > Nov 22 08:31:03 crc kubenswrapper[4789]: I1122 08:31:03.252573 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-t5l9z"] Nov 22 08:31:03 crc kubenswrapper[4789]: I1122 08:31:03.254393 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-t5l9z" Nov 22 08:31:03 crc kubenswrapper[4789]: I1122 08:31:03.266313 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-t5l9z"] Nov 22 08:31:03 crc kubenswrapper[4789]: I1122 08:31:03.374135 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dsbcd\" (UniqueName: \"kubernetes.io/projected/01a8d978-74e0-41a2-8f45-81c6d611d570-kube-api-access-dsbcd\") pod \"redhat-marketplace-t5l9z\" (UID: \"01a8d978-74e0-41a2-8f45-81c6d611d570\") " pod="openshift-marketplace/redhat-marketplace-t5l9z" Nov 22 08:31:03 crc kubenswrapper[4789]: I1122 08:31:03.374311 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/01a8d978-74e0-41a2-8f45-81c6d611d570-catalog-content\") pod \"redhat-marketplace-t5l9z\" (UID: \"01a8d978-74e0-41a2-8f45-81c6d611d570\") " pod="openshift-marketplace/redhat-marketplace-t5l9z" Nov 22 08:31:03 crc kubenswrapper[4789]: I1122 08:31:03.374336 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/01a8d978-74e0-41a2-8f45-81c6d611d570-utilities\") pod \"redhat-marketplace-t5l9z\" (UID: \"01a8d978-74e0-41a2-8f45-81c6d611d570\") " pod="openshift-marketplace/redhat-marketplace-t5l9z" Nov 22 08:31:03 crc kubenswrapper[4789]: I1122 08:31:03.476041 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dsbcd\" (UniqueName: \"kubernetes.io/projected/01a8d978-74e0-41a2-8f45-81c6d611d570-kube-api-access-dsbcd\") pod \"redhat-marketplace-t5l9z\" (UID: \"01a8d978-74e0-41a2-8f45-81c6d611d570\") " pod="openshift-marketplace/redhat-marketplace-t5l9z" Nov 22 08:31:03 crc kubenswrapper[4789]: I1122 08:31:03.476389 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/01a8d978-74e0-41a2-8f45-81c6d611d570-catalog-content\") pod \"redhat-marketplace-t5l9z\" (UID: \"01a8d978-74e0-41a2-8f45-81c6d611d570\") " pod="openshift-marketplace/redhat-marketplace-t5l9z" Nov 22 08:31:03 crc kubenswrapper[4789]: I1122 08:31:03.476410 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/01a8d978-74e0-41a2-8f45-81c6d611d570-utilities\") pod \"redhat-marketplace-t5l9z\" (UID: \"01a8d978-74e0-41a2-8f45-81c6d611d570\") " pod="openshift-marketplace/redhat-marketplace-t5l9z" Nov 22 08:31:03 crc kubenswrapper[4789]: I1122 08:31:03.476867 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/01a8d978-74e0-41a2-8f45-81c6d611d570-catalog-content\") pod \"redhat-marketplace-t5l9z\" (UID: \"01a8d978-74e0-41a2-8f45-81c6d611d570\") " pod="openshift-marketplace/redhat-marketplace-t5l9z" Nov 22 08:31:03 crc kubenswrapper[4789]: I1122 08:31:03.476923 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/01a8d978-74e0-41a2-8f45-81c6d611d570-utilities\") pod \"redhat-marketplace-t5l9z\" (UID: \"01a8d978-74e0-41a2-8f45-81c6d611d570\") " pod="openshift-marketplace/redhat-marketplace-t5l9z" Nov 22 08:31:03 crc kubenswrapper[4789]: I1122 08:31:03.494566 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dsbcd\" (UniqueName: \"kubernetes.io/projected/01a8d978-74e0-41a2-8f45-81c6d611d570-kube-api-access-dsbcd\") pod \"redhat-marketplace-t5l9z\" (UID: \"01a8d978-74e0-41a2-8f45-81c6d611d570\") " pod="openshift-marketplace/redhat-marketplace-t5l9z" Nov 22 08:31:03 crc kubenswrapper[4789]: I1122 08:31:03.546319 4789 generic.go:334] "Generic (PLEG): container finished" podID="261de000-bcb1-4c6a-858c-ea3254254d0c" containerID="f4961e0f4acd5e44db666766aea5ae467e1152770dca462e64af8ee202061e12" exitCode=0 Nov 22 08:31:03 crc kubenswrapper[4789]: I1122 08:31:03.546372 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vfzsg" event={"ID":"261de000-bcb1-4c6a-858c-ea3254254d0c","Type":"ContainerDied","Data":"f4961e0f4acd5e44db666766aea5ae467e1152770dca462e64af8ee202061e12"} Nov 22 08:31:03 crc kubenswrapper[4789]: I1122 08:31:03.574072 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-t5l9z" Nov 22 08:31:04 crc kubenswrapper[4789]: I1122 08:31:04.020393 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-t5l9z"] Nov 22 08:31:04 crc kubenswrapper[4789]: W1122 08:31:04.029166 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod01a8d978_74e0_41a2_8f45_81c6d611d570.slice/crio-9e975a41e0857c0d0cd9f487486fcf9d0e50a9c9ac7d627588bd8130e3c7b0ca WatchSource:0}: Error finding container 9e975a41e0857c0d0cd9f487486fcf9d0e50a9c9ac7d627588bd8130e3c7b0ca: Status 404 returned error can't find the container with id 9e975a41e0857c0d0cd9f487486fcf9d0e50a9c9ac7d627588bd8130e3c7b0ca Nov 22 08:31:04 crc kubenswrapper[4789]: I1122 08:31:04.555405 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vfzsg" event={"ID":"261de000-bcb1-4c6a-858c-ea3254254d0c","Type":"ContainerStarted","Data":"f597edda162fa62bd56ac5fdfee78b7400d2084aeee2a3ac513b646ef23540e1"} Nov 22 08:31:04 crc kubenswrapper[4789]: I1122 08:31:04.558108 4789 generic.go:334] "Generic (PLEG): container finished" podID="01a8d978-74e0-41a2-8f45-81c6d611d570" containerID="207d6f9bbb91889ddc8585c03b2852eacc08d7623d46d65aec7db7e3be62e8c3" exitCode=0 Nov 22 08:31:04 crc kubenswrapper[4789]: I1122 08:31:04.558156 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t5l9z" event={"ID":"01a8d978-74e0-41a2-8f45-81c6d611d570","Type":"ContainerDied","Data":"207d6f9bbb91889ddc8585c03b2852eacc08d7623d46d65aec7db7e3be62e8c3"} Nov 22 08:31:04 crc kubenswrapper[4789]: I1122 08:31:04.558178 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t5l9z" event={"ID":"01a8d978-74e0-41a2-8f45-81c6d611d570","Type":"ContainerStarted","Data":"9e975a41e0857c0d0cd9f487486fcf9d0e50a9c9ac7d627588bd8130e3c7b0ca"} Nov 22 08:31:04 crc kubenswrapper[4789]: I1122 08:31:04.581246 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-vfzsg" podStartSLOduration=2.090915631 podStartE2EDuration="4.581219532s" podCreationTimestamp="2025-11-22 08:31:00 +0000 UTC" firstStartedPulling="2025-11-22 08:31:01.514850612 +0000 UTC m=+2075.749250875" lastFinishedPulling="2025-11-22 08:31:04.005154493 +0000 UTC m=+2078.239554776" observedRunningTime="2025-11-22 08:31:04.575278559 +0000 UTC m=+2078.809678852" watchObservedRunningTime="2025-11-22 08:31:04.581219532 +0000 UTC m=+2078.815619805" Nov 22 08:31:05 crc kubenswrapper[4789]: I1122 08:31:05.372418 4789 patch_prober.go:28] interesting pod/machine-config-daemon-pxvrm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 08:31:05 crc kubenswrapper[4789]: I1122 08:31:05.372802 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 08:31:05 crc kubenswrapper[4789]: I1122 08:31:05.372852 4789 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" Nov 22 08:31:05 crc kubenswrapper[4789]: I1122 08:31:05.373706 4789 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"aaa33370060895e49b5a7fda013e1521882a1ec6a21f755348d821f46b72a57a"} pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 22 08:31:05 crc kubenswrapper[4789]: I1122 08:31:05.373833 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" containerID="cri-o://aaa33370060895e49b5a7fda013e1521882a1ec6a21f755348d821f46b72a57a" gracePeriod=600 Nov 22 08:31:05 crc kubenswrapper[4789]: I1122 08:31:05.572941 4789 generic.go:334] "Generic (PLEG): container finished" podID="58aa8071-7ecc-4692-8789-2db04ec70647" containerID="aaa33370060895e49b5a7fda013e1521882a1ec6a21f755348d821f46b72a57a" exitCode=0 Nov 22 08:31:05 crc kubenswrapper[4789]: I1122 08:31:05.573011 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" event={"ID":"58aa8071-7ecc-4692-8789-2db04ec70647","Type":"ContainerDied","Data":"aaa33370060895e49b5a7fda013e1521882a1ec6a21f755348d821f46b72a57a"} Nov 22 08:31:05 crc kubenswrapper[4789]: I1122 08:31:05.573074 4789 scope.go:117] "RemoveContainer" containerID="cd9752d9c3408de592a39b57fc6c7443fb52d47aecefc0deabddc20b20428c47" Nov 22 08:31:05 crc kubenswrapper[4789]: I1122 08:31:05.576364 4789 generic.go:334] "Generic (PLEG): container finished" podID="01a8d978-74e0-41a2-8f45-81c6d611d570" containerID="3e17bcb1c3576a168118fbdebf1a03387615ebdb139dabad7f395d11f78069f3" exitCode=0 Nov 22 08:31:05 crc kubenswrapper[4789]: I1122 08:31:05.576456 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t5l9z" event={"ID":"01a8d978-74e0-41a2-8f45-81c6d611d570","Type":"ContainerDied","Data":"3e17bcb1c3576a168118fbdebf1a03387615ebdb139dabad7f395d11f78069f3"} Nov 22 08:31:06 crc kubenswrapper[4789]: I1122 08:31:06.588828 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" event={"ID":"58aa8071-7ecc-4692-8789-2db04ec70647","Type":"ContainerStarted","Data":"c7898b5c08854d5df6033d34564729f053b077e4fcadc5a1ea77706f1981be3b"} Nov 22 08:31:06 crc kubenswrapper[4789]: I1122 08:31:06.591477 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t5l9z" event={"ID":"01a8d978-74e0-41a2-8f45-81c6d611d570","Type":"ContainerStarted","Data":"9bf53ba8c3fe35cc355f92ed91df5e251b55325fe7e741f023fe2e0e347ba99d"} Nov 22 08:31:06 crc kubenswrapper[4789]: I1122 08:31:06.638161 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-t5l9z" podStartSLOduration=2.203321072 podStartE2EDuration="3.638134736s" podCreationTimestamp="2025-11-22 08:31:03 +0000 UTC" firstStartedPulling="2025-11-22 08:31:04.559733657 +0000 UTC m=+2078.794133930" lastFinishedPulling="2025-11-22 08:31:05.994547331 +0000 UTC m=+2080.228947594" observedRunningTime="2025-11-22 08:31:06.629676756 +0000 UTC m=+2080.864077039" watchObservedRunningTime="2025-11-22 08:31:06.638134736 +0000 UTC m=+2080.872535009" Nov 22 08:31:10 crc kubenswrapper[4789]: I1122 08:31:10.385673 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-vfzsg" Nov 22 08:31:10 crc kubenswrapper[4789]: I1122 08:31:10.386959 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-vfzsg" Nov 22 08:31:10 crc kubenswrapper[4789]: I1122 08:31:10.438852 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-vfzsg" Nov 22 08:31:10 crc kubenswrapper[4789]: I1122 08:31:10.704482 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-vfzsg" Nov 22 08:31:10 crc kubenswrapper[4789]: I1122 08:31:10.752643 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vfzsg"] Nov 22 08:31:11 crc kubenswrapper[4789]: I1122 08:31:11.752394 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-dvlgg" Nov 22 08:31:11 crc kubenswrapper[4789]: I1122 08:31:11.802798 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-dvlgg" Nov 22 08:31:12 crc kubenswrapper[4789]: I1122 08:31:12.645967 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-vfzsg" podUID="261de000-bcb1-4c6a-858c-ea3254254d0c" containerName="registry-server" containerID="cri-o://f597edda162fa62bd56ac5fdfee78b7400d2084aeee2a3ac513b646ef23540e1" gracePeriod=2 Nov 22 08:31:13 crc kubenswrapper[4789]: I1122 08:31:13.071742 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vfzsg" Nov 22 08:31:13 crc kubenswrapper[4789]: I1122 08:31:13.073221 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dvlgg"] Nov 22 08:31:13 crc kubenswrapper[4789]: I1122 08:31:13.259885 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qqvm2\" (UniqueName: \"kubernetes.io/projected/261de000-bcb1-4c6a-858c-ea3254254d0c-kube-api-access-qqvm2\") pod \"261de000-bcb1-4c6a-858c-ea3254254d0c\" (UID: \"261de000-bcb1-4c6a-858c-ea3254254d0c\") " Nov 22 08:31:13 crc kubenswrapper[4789]: I1122 08:31:13.260535 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/261de000-bcb1-4c6a-858c-ea3254254d0c-catalog-content\") pod \"261de000-bcb1-4c6a-858c-ea3254254d0c\" (UID: \"261de000-bcb1-4c6a-858c-ea3254254d0c\") " Nov 22 08:31:13 crc kubenswrapper[4789]: I1122 08:31:13.260784 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/261de000-bcb1-4c6a-858c-ea3254254d0c-utilities\") pod \"261de000-bcb1-4c6a-858c-ea3254254d0c\" (UID: \"261de000-bcb1-4c6a-858c-ea3254254d0c\") " Nov 22 08:31:13 crc kubenswrapper[4789]: I1122 08:31:13.262233 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/261de000-bcb1-4c6a-858c-ea3254254d0c-utilities" (OuterVolumeSpecName: "utilities") pod "261de000-bcb1-4c6a-858c-ea3254254d0c" (UID: "261de000-bcb1-4c6a-858c-ea3254254d0c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:31:13 crc kubenswrapper[4789]: I1122 08:31:13.265511 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/261de000-bcb1-4c6a-858c-ea3254254d0c-kube-api-access-qqvm2" (OuterVolumeSpecName: "kube-api-access-qqvm2") pod "261de000-bcb1-4c6a-858c-ea3254254d0c" (UID: "261de000-bcb1-4c6a-858c-ea3254254d0c"). InnerVolumeSpecName "kube-api-access-qqvm2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:31:13 crc kubenswrapper[4789]: I1122 08:31:13.316050 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/261de000-bcb1-4c6a-858c-ea3254254d0c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "261de000-bcb1-4c6a-858c-ea3254254d0c" (UID: "261de000-bcb1-4c6a-858c-ea3254254d0c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:31:13 crc kubenswrapper[4789]: I1122 08:31:13.363866 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qqvm2\" (UniqueName: \"kubernetes.io/projected/261de000-bcb1-4c6a-858c-ea3254254d0c-kube-api-access-qqvm2\") on node \"crc\" DevicePath \"\"" Nov 22 08:31:13 crc kubenswrapper[4789]: I1122 08:31:13.363922 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/261de000-bcb1-4c6a-858c-ea3254254d0c-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 08:31:13 crc kubenswrapper[4789]: I1122 08:31:13.363979 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/261de000-bcb1-4c6a-858c-ea3254254d0c-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 08:31:13 crc kubenswrapper[4789]: I1122 08:31:13.575591 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-t5l9z" Nov 22 08:31:13 crc kubenswrapper[4789]: I1122 08:31:13.575637 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-t5l9z" Nov 22 08:31:13 crc kubenswrapper[4789]: I1122 08:31:13.617999 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-t5l9z" Nov 22 08:31:13 crc kubenswrapper[4789]: I1122 08:31:13.656635 4789 generic.go:334] "Generic (PLEG): container finished" podID="261de000-bcb1-4c6a-858c-ea3254254d0c" containerID="f597edda162fa62bd56ac5fdfee78b7400d2084aeee2a3ac513b646ef23540e1" exitCode=0 Nov 22 08:31:13 crc kubenswrapper[4789]: I1122 08:31:13.656706 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vfzsg" Nov 22 08:31:13 crc kubenswrapper[4789]: I1122 08:31:13.656764 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vfzsg" event={"ID":"261de000-bcb1-4c6a-858c-ea3254254d0c","Type":"ContainerDied","Data":"f597edda162fa62bd56ac5fdfee78b7400d2084aeee2a3ac513b646ef23540e1"} Nov 22 08:31:13 crc kubenswrapper[4789]: I1122 08:31:13.656812 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vfzsg" event={"ID":"261de000-bcb1-4c6a-858c-ea3254254d0c","Type":"ContainerDied","Data":"cc829ca05a2a7ccd05d4ecbce8398b13472c8f45109d695ff1dccdaaa06723b9"} Nov 22 08:31:13 crc kubenswrapper[4789]: I1122 08:31:13.656831 4789 scope.go:117] "RemoveContainer" containerID="f597edda162fa62bd56ac5fdfee78b7400d2084aeee2a3ac513b646ef23540e1" Nov 22 08:31:13 crc kubenswrapper[4789]: I1122 08:31:13.656861 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-dvlgg" podUID="e70319d4-4101-4ceb-80a3-b615df927a18" containerName="registry-server" containerID="cri-o://660d54756912e4543f81365c1b5d691d57d8c8ac368d61de10db6920932ad699" gracePeriod=2 Nov 22 08:31:13 crc kubenswrapper[4789]: I1122 08:31:13.690807 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vfzsg"] Nov 22 08:31:13 crc kubenswrapper[4789]: I1122 08:31:13.690855 4789 scope.go:117] "RemoveContainer" containerID="f4961e0f4acd5e44db666766aea5ae467e1152770dca462e64af8ee202061e12" Nov 22 08:31:13 crc kubenswrapper[4789]: I1122 08:31:13.698220 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-vfzsg"] Nov 22 08:31:13 crc kubenswrapper[4789]: I1122 08:31:13.709773 4789 scope.go:117] "RemoveContainer" containerID="72eb86159b377c03739718819dcaa6ec61c19f9ec6cc06033dca2146af456bac" Nov 22 08:31:13 crc kubenswrapper[4789]: I1122 08:31:13.711701 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-t5l9z" Nov 22 08:31:13 crc kubenswrapper[4789]: I1122 08:31:13.839572 4789 scope.go:117] "RemoveContainer" containerID="f597edda162fa62bd56ac5fdfee78b7400d2084aeee2a3ac513b646ef23540e1" Nov 22 08:31:13 crc kubenswrapper[4789]: E1122 08:31:13.840223 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f597edda162fa62bd56ac5fdfee78b7400d2084aeee2a3ac513b646ef23540e1\": container with ID starting with f597edda162fa62bd56ac5fdfee78b7400d2084aeee2a3ac513b646ef23540e1 not found: ID does not exist" containerID="f597edda162fa62bd56ac5fdfee78b7400d2084aeee2a3ac513b646ef23540e1" Nov 22 08:31:13 crc kubenswrapper[4789]: I1122 08:31:13.840280 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f597edda162fa62bd56ac5fdfee78b7400d2084aeee2a3ac513b646ef23540e1"} err="failed to get container status \"f597edda162fa62bd56ac5fdfee78b7400d2084aeee2a3ac513b646ef23540e1\": rpc error: code = NotFound desc = could not find container \"f597edda162fa62bd56ac5fdfee78b7400d2084aeee2a3ac513b646ef23540e1\": container with ID starting with f597edda162fa62bd56ac5fdfee78b7400d2084aeee2a3ac513b646ef23540e1 not found: ID does not exist" Nov 22 08:31:13 crc kubenswrapper[4789]: I1122 08:31:13.840314 4789 scope.go:117] "RemoveContainer" containerID="f4961e0f4acd5e44db666766aea5ae467e1152770dca462e64af8ee202061e12" Nov 22 08:31:13 crc kubenswrapper[4789]: E1122 08:31:13.840961 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f4961e0f4acd5e44db666766aea5ae467e1152770dca462e64af8ee202061e12\": container with ID starting with f4961e0f4acd5e44db666766aea5ae467e1152770dca462e64af8ee202061e12 not found: ID does not exist" containerID="f4961e0f4acd5e44db666766aea5ae467e1152770dca462e64af8ee202061e12" Nov 22 08:31:13 crc kubenswrapper[4789]: I1122 08:31:13.841004 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f4961e0f4acd5e44db666766aea5ae467e1152770dca462e64af8ee202061e12"} err="failed to get container status \"f4961e0f4acd5e44db666766aea5ae467e1152770dca462e64af8ee202061e12\": rpc error: code = NotFound desc = could not find container \"f4961e0f4acd5e44db666766aea5ae467e1152770dca462e64af8ee202061e12\": container with ID starting with f4961e0f4acd5e44db666766aea5ae467e1152770dca462e64af8ee202061e12 not found: ID does not exist" Nov 22 08:31:13 crc kubenswrapper[4789]: I1122 08:31:13.841229 4789 scope.go:117] "RemoveContainer" containerID="72eb86159b377c03739718819dcaa6ec61c19f9ec6cc06033dca2146af456bac" Nov 22 08:31:13 crc kubenswrapper[4789]: E1122 08:31:13.841502 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"72eb86159b377c03739718819dcaa6ec61c19f9ec6cc06033dca2146af456bac\": container with ID starting with 72eb86159b377c03739718819dcaa6ec61c19f9ec6cc06033dca2146af456bac not found: ID does not exist" containerID="72eb86159b377c03739718819dcaa6ec61c19f9ec6cc06033dca2146af456bac" Nov 22 08:31:13 crc kubenswrapper[4789]: I1122 08:31:13.841553 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"72eb86159b377c03739718819dcaa6ec61c19f9ec6cc06033dca2146af456bac"} err="failed to get container status \"72eb86159b377c03739718819dcaa6ec61c19f9ec6cc06033dca2146af456bac\": rpc error: code = NotFound desc = could not find container \"72eb86159b377c03739718819dcaa6ec61c19f9ec6cc06033dca2146af456bac\": container with ID starting with 72eb86159b377c03739718819dcaa6ec61c19f9ec6cc06033dca2146af456bac not found: ID does not exist" Nov 22 08:31:13 crc kubenswrapper[4789]: I1122 08:31:13.974646 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="261de000-bcb1-4c6a-858c-ea3254254d0c" path="/var/lib/kubelet/pods/261de000-bcb1-4c6a-858c-ea3254254d0c/volumes" Nov 22 08:31:14 crc kubenswrapper[4789]: I1122 08:31:14.076379 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dvlgg" Nov 22 08:31:14 crc kubenswrapper[4789]: I1122 08:31:14.175382 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e70319d4-4101-4ceb-80a3-b615df927a18-utilities\") pod \"e70319d4-4101-4ceb-80a3-b615df927a18\" (UID: \"e70319d4-4101-4ceb-80a3-b615df927a18\") " Nov 22 08:31:14 crc kubenswrapper[4789]: I1122 08:31:14.175439 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e70319d4-4101-4ceb-80a3-b615df927a18-catalog-content\") pod \"e70319d4-4101-4ceb-80a3-b615df927a18\" (UID: \"e70319d4-4101-4ceb-80a3-b615df927a18\") " Nov 22 08:31:14 crc kubenswrapper[4789]: I1122 08:31:14.175463 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4h585\" (UniqueName: \"kubernetes.io/projected/e70319d4-4101-4ceb-80a3-b615df927a18-kube-api-access-4h585\") pod \"e70319d4-4101-4ceb-80a3-b615df927a18\" (UID: \"e70319d4-4101-4ceb-80a3-b615df927a18\") " Nov 22 08:31:14 crc kubenswrapper[4789]: I1122 08:31:14.176141 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e70319d4-4101-4ceb-80a3-b615df927a18-utilities" (OuterVolumeSpecName: "utilities") pod "e70319d4-4101-4ceb-80a3-b615df927a18" (UID: "e70319d4-4101-4ceb-80a3-b615df927a18"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:31:14 crc kubenswrapper[4789]: I1122 08:31:14.180149 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e70319d4-4101-4ceb-80a3-b615df927a18-kube-api-access-4h585" (OuterVolumeSpecName: "kube-api-access-4h585") pod "e70319d4-4101-4ceb-80a3-b615df927a18" (UID: "e70319d4-4101-4ceb-80a3-b615df927a18"). InnerVolumeSpecName "kube-api-access-4h585". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:31:14 crc kubenswrapper[4789]: I1122 08:31:14.258569 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e70319d4-4101-4ceb-80a3-b615df927a18-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e70319d4-4101-4ceb-80a3-b615df927a18" (UID: "e70319d4-4101-4ceb-80a3-b615df927a18"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:31:14 crc kubenswrapper[4789]: I1122 08:31:14.277800 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e70319d4-4101-4ceb-80a3-b615df927a18-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 08:31:14 crc kubenswrapper[4789]: I1122 08:31:14.277835 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e70319d4-4101-4ceb-80a3-b615df927a18-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 08:31:14 crc kubenswrapper[4789]: I1122 08:31:14.277845 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4h585\" (UniqueName: \"kubernetes.io/projected/e70319d4-4101-4ceb-80a3-b615df927a18-kube-api-access-4h585\") on node \"crc\" DevicePath \"\"" Nov 22 08:31:14 crc kubenswrapper[4789]: I1122 08:31:14.670722 4789 generic.go:334] "Generic (PLEG): container finished" podID="e70319d4-4101-4ceb-80a3-b615df927a18" containerID="660d54756912e4543f81365c1b5d691d57d8c8ac368d61de10db6920932ad699" exitCode=0 Nov 22 08:31:14 crc kubenswrapper[4789]: I1122 08:31:14.670851 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dvlgg" event={"ID":"e70319d4-4101-4ceb-80a3-b615df927a18","Type":"ContainerDied","Data":"660d54756912e4543f81365c1b5d691d57d8c8ac368d61de10db6920932ad699"} Nov 22 08:31:14 crc kubenswrapper[4789]: I1122 08:31:14.671417 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dvlgg" event={"ID":"e70319d4-4101-4ceb-80a3-b615df927a18","Type":"ContainerDied","Data":"5cf99c1c403112a32d78783bfe09e3b006b68f2103103db594a73e2711cfc507"} Nov 22 08:31:14 crc kubenswrapper[4789]: I1122 08:31:14.670858 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dvlgg" Nov 22 08:31:14 crc kubenswrapper[4789]: I1122 08:31:14.671451 4789 scope.go:117] "RemoveContainer" containerID="660d54756912e4543f81365c1b5d691d57d8c8ac368d61de10db6920932ad699" Nov 22 08:31:14 crc kubenswrapper[4789]: I1122 08:31:14.706369 4789 scope.go:117] "RemoveContainer" containerID="dfba41882e69b545a7add9897c427677d5141b54e729458b4ec32cade78d1d90" Nov 22 08:31:14 crc kubenswrapper[4789]: I1122 08:31:14.712714 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dvlgg"] Nov 22 08:31:14 crc kubenswrapper[4789]: I1122 08:31:14.720602 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-dvlgg"] Nov 22 08:31:14 crc kubenswrapper[4789]: I1122 08:31:14.729781 4789 scope.go:117] "RemoveContainer" containerID="8dcfe552980a30fdcb621b3034b6e2afb955d77d9be9763829f46845ddfd0a57" Nov 22 08:31:14 crc kubenswrapper[4789]: I1122 08:31:14.747355 4789 scope.go:117] "RemoveContainer" containerID="660d54756912e4543f81365c1b5d691d57d8c8ac368d61de10db6920932ad699" Nov 22 08:31:14 crc kubenswrapper[4789]: E1122 08:31:14.747854 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"660d54756912e4543f81365c1b5d691d57d8c8ac368d61de10db6920932ad699\": container with ID starting with 660d54756912e4543f81365c1b5d691d57d8c8ac368d61de10db6920932ad699 not found: ID does not exist" containerID="660d54756912e4543f81365c1b5d691d57d8c8ac368d61de10db6920932ad699" Nov 22 08:31:14 crc kubenswrapper[4789]: I1122 08:31:14.747927 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"660d54756912e4543f81365c1b5d691d57d8c8ac368d61de10db6920932ad699"} err="failed to get container status \"660d54756912e4543f81365c1b5d691d57d8c8ac368d61de10db6920932ad699\": rpc error: code = NotFound desc = could not find container \"660d54756912e4543f81365c1b5d691d57d8c8ac368d61de10db6920932ad699\": container with ID starting with 660d54756912e4543f81365c1b5d691d57d8c8ac368d61de10db6920932ad699 not found: ID does not exist" Nov 22 08:31:14 crc kubenswrapper[4789]: I1122 08:31:14.747961 4789 scope.go:117] "RemoveContainer" containerID="dfba41882e69b545a7add9897c427677d5141b54e729458b4ec32cade78d1d90" Nov 22 08:31:14 crc kubenswrapper[4789]: E1122 08:31:14.748242 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dfba41882e69b545a7add9897c427677d5141b54e729458b4ec32cade78d1d90\": container with ID starting with dfba41882e69b545a7add9897c427677d5141b54e729458b4ec32cade78d1d90 not found: ID does not exist" containerID="dfba41882e69b545a7add9897c427677d5141b54e729458b4ec32cade78d1d90" Nov 22 08:31:14 crc kubenswrapper[4789]: I1122 08:31:14.748274 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dfba41882e69b545a7add9897c427677d5141b54e729458b4ec32cade78d1d90"} err="failed to get container status \"dfba41882e69b545a7add9897c427677d5141b54e729458b4ec32cade78d1d90\": rpc error: code = NotFound desc = could not find container \"dfba41882e69b545a7add9897c427677d5141b54e729458b4ec32cade78d1d90\": container with ID starting with dfba41882e69b545a7add9897c427677d5141b54e729458b4ec32cade78d1d90 not found: ID does not exist" Nov 22 08:31:14 crc kubenswrapper[4789]: I1122 08:31:14.748298 4789 scope.go:117] "RemoveContainer" containerID="8dcfe552980a30fdcb621b3034b6e2afb955d77d9be9763829f46845ddfd0a57" Nov 22 08:31:14 crc kubenswrapper[4789]: E1122 08:31:14.748557 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8dcfe552980a30fdcb621b3034b6e2afb955d77d9be9763829f46845ddfd0a57\": container with ID starting with 8dcfe552980a30fdcb621b3034b6e2afb955d77d9be9763829f46845ddfd0a57 not found: ID does not exist" containerID="8dcfe552980a30fdcb621b3034b6e2afb955d77d9be9763829f46845ddfd0a57" Nov 22 08:31:14 crc kubenswrapper[4789]: I1122 08:31:14.748592 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8dcfe552980a30fdcb621b3034b6e2afb955d77d9be9763829f46845ddfd0a57"} err="failed to get container status \"8dcfe552980a30fdcb621b3034b6e2afb955d77d9be9763829f46845ddfd0a57\": rpc error: code = NotFound desc = could not find container \"8dcfe552980a30fdcb621b3034b6e2afb955d77d9be9763829f46845ddfd0a57\": container with ID starting with 8dcfe552980a30fdcb621b3034b6e2afb955d77d9be9763829f46845ddfd0a57 not found: ID does not exist" Nov 22 08:31:15 crc kubenswrapper[4789]: I1122 08:31:15.975158 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e70319d4-4101-4ceb-80a3-b615df927a18" path="/var/lib/kubelet/pods/e70319d4-4101-4ceb-80a3-b615df927a18/volumes" Nov 22 08:31:17 crc kubenswrapper[4789]: I1122 08:31:17.278570 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-t5l9z"] Nov 22 08:31:17 crc kubenswrapper[4789]: I1122 08:31:17.278888 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-t5l9z" podUID="01a8d978-74e0-41a2-8f45-81c6d611d570" containerName="registry-server" containerID="cri-o://9bf53ba8c3fe35cc355f92ed91df5e251b55325fe7e741f023fe2e0e347ba99d" gracePeriod=2 Nov 22 08:31:17 crc kubenswrapper[4789]: I1122 08:31:17.700207 4789 generic.go:334] "Generic (PLEG): container finished" podID="01a8d978-74e0-41a2-8f45-81c6d611d570" containerID="9bf53ba8c3fe35cc355f92ed91df5e251b55325fe7e741f023fe2e0e347ba99d" exitCode=0 Nov 22 08:31:17 crc kubenswrapper[4789]: I1122 08:31:17.700305 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t5l9z" event={"ID":"01a8d978-74e0-41a2-8f45-81c6d611d570","Type":"ContainerDied","Data":"9bf53ba8c3fe35cc355f92ed91df5e251b55325fe7e741f023fe2e0e347ba99d"} Nov 22 08:31:17 crc kubenswrapper[4789]: I1122 08:31:17.700665 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t5l9z" event={"ID":"01a8d978-74e0-41a2-8f45-81c6d611d570","Type":"ContainerDied","Data":"9e975a41e0857c0d0cd9f487486fcf9d0e50a9c9ac7d627588bd8130e3c7b0ca"} Nov 22 08:31:17 crc kubenswrapper[4789]: I1122 08:31:17.700685 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9e975a41e0857c0d0cd9f487486fcf9d0e50a9c9ac7d627588bd8130e3c7b0ca" Nov 22 08:31:17 crc kubenswrapper[4789]: I1122 08:31:17.708533 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-t5l9z" Nov 22 08:31:17 crc kubenswrapper[4789]: I1122 08:31:17.846365 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dsbcd\" (UniqueName: \"kubernetes.io/projected/01a8d978-74e0-41a2-8f45-81c6d611d570-kube-api-access-dsbcd\") pod \"01a8d978-74e0-41a2-8f45-81c6d611d570\" (UID: \"01a8d978-74e0-41a2-8f45-81c6d611d570\") " Nov 22 08:31:17 crc kubenswrapper[4789]: I1122 08:31:17.846441 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/01a8d978-74e0-41a2-8f45-81c6d611d570-utilities\") pod \"01a8d978-74e0-41a2-8f45-81c6d611d570\" (UID: \"01a8d978-74e0-41a2-8f45-81c6d611d570\") " Nov 22 08:31:17 crc kubenswrapper[4789]: I1122 08:31:17.846506 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/01a8d978-74e0-41a2-8f45-81c6d611d570-catalog-content\") pod \"01a8d978-74e0-41a2-8f45-81c6d611d570\" (UID: \"01a8d978-74e0-41a2-8f45-81c6d611d570\") " Nov 22 08:31:17 crc kubenswrapper[4789]: I1122 08:31:17.847828 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/01a8d978-74e0-41a2-8f45-81c6d611d570-utilities" (OuterVolumeSpecName: "utilities") pod "01a8d978-74e0-41a2-8f45-81c6d611d570" (UID: "01a8d978-74e0-41a2-8f45-81c6d611d570"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:31:17 crc kubenswrapper[4789]: I1122 08:31:17.853889 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01a8d978-74e0-41a2-8f45-81c6d611d570-kube-api-access-dsbcd" (OuterVolumeSpecName: "kube-api-access-dsbcd") pod "01a8d978-74e0-41a2-8f45-81c6d611d570" (UID: "01a8d978-74e0-41a2-8f45-81c6d611d570"). InnerVolumeSpecName "kube-api-access-dsbcd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:31:17 crc kubenswrapper[4789]: I1122 08:31:17.869057 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/01a8d978-74e0-41a2-8f45-81c6d611d570-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "01a8d978-74e0-41a2-8f45-81c6d611d570" (UID: "01a8d978-74e0-41a2-8f45-81c6d611d570"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:31:17 crc kubenswrapper[4789]: I1122 08:31:17.948316 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dsbcd\" (UniqueName: \"kubernetes.io/projected/01a8d978-74e0-41a2-8f45-81c6d611d570-kube-api-access-dsbcd\") on node \"crc\" DevicePath \"\"" Nov 22 08:31:17 crc kubenswrapper[4789]: I1122 08:31:17.948356 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/01a8d978-74e0-41a2-8f45-81c6d611d570-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 08:31:17 crc kubenswrapper[4789]: I1122 08:31:17.948371 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/01a8d978-74e0-41a2-8f45-81c6d611d570-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 08:31:18 crc kubenswrapper[4789]: I1122 08:31:18.712824 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-t5l9z" Nov 22 08:31:18 crc kubenswrapper[4789]: I1122 08:31:18.742863 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-t5l9z"] Nov 22 08:31:18 crc kubenswrapper[4789]: I1122 08:31:18.761686 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-t5l9z"] Nov 22 08:31:19 crc kubenswrapper[4789]: I1122 08:31:19.976349 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01a8d978-74e0-41a2-8f45-81c6d611d570" path="/var/lib/kubelet/pods/01a8d978-74e0-41a2-8f45-81c6d611d570/volumes" Nov 22 08:33:05 crc kubenswrapper[4789]: I1122 08:33:05.372352 4789 patch_prober.go:28] interesting pod/machine-config-daemon-pxvrm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 08:33:05 crc kubenswrapper[4789]: I1122 08:33:05.373247 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 08:33:35 crc kubenswrapper[4789]: I1122 08:33:35.372854 4789 patch_prober.go:28] interesting pod/machine-config-daemon-pxvrm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 08:33:35 crc kubenswrapper[4789]: I1122 08:33:35.373500 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 08:34:05 crc kubenswrapper[4789]: I1122 08:34:05.372132 4789 patch_prober.go:28] interesting pod/machine-config-daemon-pxvrm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 08:34:05 crc kubenswrapper[4789]: I1122 08:34:05.372730 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 08:34:05 crc kubenswrapper[4789]: I1122 08:34:05.372840 4789 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" Nov 22 08:34:05 crc kubenswrapper[4789]: I1122 08:34:05.373557 4789 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c7898b5c08854d5df6033d34564729f053b077e4fcadc5a1ea77706f1981be3b"} pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 22 08:34:05 crc kubenswrapper[4789]: I1122 08:34:05.373608 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" containerID="cri-o://c7898b5c08854d5df6033d34564729f053b077e4fcadc5a1ea77706f1981be3b" gracePeriod=600 Nov 22 08:34:05 crc kubenswrapper[4789]: E1122 08:34:05.494031 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 08:34:06 crc kubenswrapper[4789]: I1122 08:34:06.266327 4789 generic.go:334] "Generic (PLEG): container finished" podID="58aa8071-7ecc-4692-8789-2db04ec70647" containerID="c7898b5c08854d5df6033d34564729f053b077e4fcadc5a1ea77706f1981be3b" exitCode=0 Nov 22 08:34:06 crc kubenswrapper[4789]: I1122 08:34:06.266388 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" event={"ID":"58aa8071-7ecc-4692-8789-2db04ec70647","Type":"ContainerDied","Data":"c7898b5c08854d5df6033d34564729f053b077e4fcadc5a1ea77706f1981be3b"} Nov 22 08:34:06 crc kubenswrapper[4789]: I1122 08:34:06.266449 4789 scope.go:117] "RemoveContainer" containerID="aaa33370060895e49b5a7fda013e1521882a1ec6a21f755348d821f46b72a57a" Nov 22 08:34:06 crc kubenswrapper[4789]: I1122 08:34:06.267835 4789 scope.go:117] "RemoveContainer" containerID="c7898b5c08854d5df6033d34564729f053b077e4fcadc5a1ea77706f1981be3b" Nov 22 08:34:06 crc kubenswrapper[4789]: E1122 08:34:06.269655 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 08:34:18 crc kubenswrapper[4789]: I1122 08:34:18.965799 4789 scope.go:117] "RemoveContainer" containerID="c7898b5c08854d5df6033d34564729f053b077e4fcadc5a1ea77706f1981be3b" Nov 22 08:34:18 crc kubenswrapper[4789]: E1122 08:34:18.966672 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 08:34:30 crc kubenswrapper[4789]: I1122 08:34:30.966188 4789 scope.go:117] "RemoveContainer" containerID="c7898b5c08854d5df6033d34564729f053b077e4fcadc5a1ea77706f1981be3b" Nov 22 08:34:30 crc kubenswrapper[4789]: E1122 08:34:30.967557 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 08:34:45 crc kubenswrapper[4789]: I1122 08:34:45.965959 4789 scope.go:117] "RemoveContainer" containerID="c7898b5c08854d5df6033d34564729f053b077e4fcadc5a1ea77706f1981be3b" Nov 22 08:34:45 crc kubenswrapper[4789]: E1122 08:34:45.967234 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 08:34:49 crc kubenswrapper[4789]: I1122 08:34:49.188062 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-tvdg4"] Nov 22 08:34:49 crc kubenswrapper[4789]: E1122 08:34:49.197228 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e70319d4-4101-4ceb-80a3-b615df927a18" containerName="extract-utilities" Nov 22 08:34:49 crc kubenswrapper[4789]: I1122 08:34:49.197280 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="e70319d4-4101-4ceb-80a3-b615df927a18" containerName="extract-utilities" Nov 22 08:34:49 crc kubenswrapper[4789]: E1122 08:34:49.197304 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="261de000-bcb1-4c6a-858c-ea3254254d0c" containerName="registry-server" Nov 22 08:34:49 crc kubenswrapper[4789]: I1122 08:34:49.197314 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="261de000-bcb1-4c6a-858c-ea3254254d0c" containerName="registry-server" Nov 22 08:34:49 crc kubenswrapper[4789]: E1122 08:34:49.197341 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="261de000-bcb1-4c6a-858c-ea3254254d0c" containerName="extract-content" Nov 22 08:34:49 crc kubenswrapper[4789]: I1122 08:34:49.197350 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="261de000-bcb1-4c6a-858c-ea3254254d0c" containerName="extract-content" Nov 22 08:34:49 crc kubenswrapper[4789]: E1122 08:34:49.197378 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01a8d978-74e0-41a2-8f45-81c6d611d570" containerName="registry-server" Nov 22 08:34:49 crc kubenswrapper[4789]: I1122 08:34:49.197388 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="01a8d978-74e0-41a2-8f45-81c6d611d570" containerName="registry-server" Nov 22 08:34:49 crc kubenswrapper[4789]: E1122 08:34:49.197404 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e70319d4-4101-4ceb-80a3-b615df927a18" containerName="registry-server" Nov 22 08:34:49 crc kubenswrapper[4789]: I1122 08:34:49.197412 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="e70319d4-4101-4ceb-80a3-b615df927a18" containerName="registry-server" Nov 22 08:34:49 crc kubenswrapper[4789]: E1122 08:34:49.197425 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01a8d978-74e0-41a2-8f45-81c6d611d570" containerName="extract-content" Nov 22 08:34:49 crc kubenswrapper[4789]: I1122 08:34:49.197433 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="01a8d978-74e0-41a2-8f45-81c6d611d570" containerName="extract-content" Nov 22 08:34:49 crc kubenswrapper[4789]: E1122 08:34:49.197452 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="261de000-bcb1-4c6a-858c-ea3254254d0c" containerName="extract-utilities" Nov 22 08:34:49 crc kubenswrapper[4789]: I1122 08:34:49.197461 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="261de000-bcb1-4c6a-858c-ea3254254d0c" containerName="extract-utilities" Nov 22 08:34:49 crc kubenswrapper[4789]: E1122 08:34:49.197480 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01a8d978-74e0-41a2-8f45-81c6d611d570" containerName="extract-utilities" Nov 22 08:34:49 crc kubenswrapper[4789]: I1122 08:34:49.197489 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="01a8d978-74e0-41a2-8f45-81c6d611d570" containerName="extract-utilities" Nov 22 08:34:49 crc kubenswrapper[4789]: E1122 08:34:49.197511 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e70319d4-4101-4ceb-80a3-b615df927a18" containerName="extract-content" Nov 22 08:34:49 crc kubenswrapper[4789]: I1122 08:34:49.197518 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="e70319d4-4101-4ceb-80a3-b615df927a18" containerName="extract-content" Nov 22 08:34:49 crc kubenswrapper[4789]: I1122 08:34:49.199146 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="261de000-bcb1-4c6a-858c-ea3254254d0c" containerName="registry-server" Nov 22 08:34:49 crc kubenswrapper[4789]: I1122 08:34:49.199181 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="e70319d4-4101-4ceb-80a3-b615df927a18" containerName="registry-server" Nov 22 08:34:49 crc kubenswrapper[4789]: I1122 08:34:49.199344 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="01a8d978-74e0-41a2-8f45-81c6d611d570" containerName="registry-server" Nov 22 08:34:49 crc kubenswrapper[4789]: I1122 08:34:49.205163 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tvdg4" Nov 22 08:34:49 crc kubenswrapper[4789]: I1122 08:34:49.209039 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tvdg4"] Nov 22 08:34:49 crc kubenswrapper[4789]: I1122 08:34:49.281633 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5bmtl\" (UniqueName: \"kubernetes.io/projected/fae8448e-393d-4700-ae60-e73ddf5c1053-kube-api-access-5bmtl\") pod \"certified-operators-tvdg4\" (UID: \"fae8448e-393d-4700-ae60-e73ddf5c1053\") " pod="openshift-marketplace/certified-operators-tvdg4" Nov 22 08:34:49 crc kubenswrapper[4789]: I1122 08:34:49.281694 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fae8448e-393d-4700-ae60-e73ddf5c1053-catalog-content\") pod \"certified-operators-tvdg4\" (UID: \"fae8448e-393d-4700-ae60-e73ddf5c1053\") " pod="openshift-marketplace/certified-operators-tvdg4" Nov 22 08:34:49 crc kubenswrapper[4789]: I1122 08:34:49.281824 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fae8448e-393d-4700-ae60-e73ddf5c1053-utilities\") pod \"certified-operators-tvdg4\" (UID: \"fae8448e-393d-4700-ae60-e73ddf5c1053\") " pod="openshift-marketplace/certified-operators-tvdg4" Nov 22 08:34:49 crc kubenswrapper[4789]: I1122 08:34:49.383226 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5bmtl\" (UniqueName: \"kubernetes.io/projected/fae8448e-393d-4700-ae60-e73ddf5c1053-kube-api-access-5bmtl\") pod \"certified-operators-tvdg4\" (UID: \"fae8448e-393d-4700-ae60-e73ddf5c1053\") " pod="openshift-marketplace/certified-operators-tvdg4" Nov 22 08:34:49 crc kubenswrapper[4789]: I1122 08:34:49.383288 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fae8448e-393d-4700-ae60-e73ddf5c1053-catalog-content\") pod \"certified-operators-tvdg4\" (UID: \"fae8448e-393d-4700-ae60-e73ddf5c1053\") " pod="openshift-marketplace/certified-operators-tvdg4" Nov 22 08:34:49 crc kubenswrapper[4789]: I1122 08:34:49.383358 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fae8448e-393d-4700-ae60-e73ddf5c1053-utilities\") pod \"certified-operators-tvdg4\" (UID: \"fae8448e-393d-4700-ae60-e73ddf5c1053\") " pod="openshift-marketplace/certified-operators-tvdg4" Nov 22 08:34:49 crc kubenswrapper[4789]: I1122 08:34:49.383886 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fae8448e-393d-4700-ae60-e73ddf5c1053-catalog-content\") pod \"certified-operators-tvdg4\" (UID: \"fae8448e-393d-4700-ae60-e73ddf5c1053\") " pod="openshift-marketplace/certified-operators-tvdg4" Nov 22 08:34:49 crc kubenswrapper[4789]: I1122 08:34:49.383900 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fae8448e-393d-4700-ae60-e73ddf5c1053-utilities\") pod \"certified-operators-tvdg4\" (UID: \"fae8448e-393d-4700-ae60-e73ddf5c1053\") " pod="openshift-marketplace/certified-operators-tvdg4" Nov 22 08:34:49 crc kubenswrapper[4789]: I1122 08:34:49.406950 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5bmtl\" (UniqueName: \"kubernetes.io/projected/fae8448e-393d-4700-ae60-e73ddf5c1053-kube-api-access-5bmtl\") pod \"certified-operators-tvdg4\" (UID: \"fae8448e-393d-4700-ae60-e73ddf5c1053\") " pod="openshift-marketplace/certified-operators-tvdg4" Nov 22 08:34:49 crc kubenswrapper[4789]: I1122 08:34:49.544180 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tvdg4" Nov 22 08:34:50 crc kubenswrapper[4789]: I1122 08:34:50.038827 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tvdg4"] Nov 22 08:34:50 crc kubenswrapper[4789]: I1122 08:34:50.672895 4789 generic.go:334] "Generic (PLEG): container finished" podID="fae8448e-393d-4700-ae60-e73ddf5c1053" containerID="440315f1369b446fcc72a3ec6145b7c853253bf0c69c80cbc4253fada8c1cf33" exitCode=0 Nov 22 08:34:50 crc kubenswrapper[4789]: I1122 08:34:50.673010 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tvdg4" event={"ID":"fae8448e-393d-4700-ae60-e73ddf5c1053","Type":"ContainerDied","Data":"440315f1369b446fcc72a3ec6145b7c853253bf0c69c80cbc4253fada8c1cf33"} Nov 22 08:34:50 crc kubenswrapper[4789]: I1122 08:34:50.673390 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tvdg4" event={"ID":"fae8448e-393d-4700-ae60-e73ddf5c1053","Type":"ContainerStarted","Data":"15a1ae39a0d49263c395dc1823248fe391c6bc9945bffb18e10b0ddaa5437e06"} Nov 22 08:34:53 crc kubenswrapper[4789]: I1122 08:34:53.705204 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tvdg4" event={"ID":"fae8448e-393d-4700-ae60-e73ddf5c1053","Type":"ContainerStarted","Data":"92111ede29fc20f3f1181547f323d0c89f3cbd911a7923e22b63da4d8d961b87"} Nov 22 08:34:54 crc kubenswrapper[4789]: I1122 08:34:54.717048 4789 generic.go:334] "Generic (PLEG): container finished" podID="fae8448e-393d-4700-ae60-e73ddf5c1053" containerID="92111ede29fc20f3f1181547f323d0c89f3cbd911a7923e22b63da4d8d961b87" exitCode=0 Nov 22 08:34:54 crc kubenswrapper[4789]: I1122 08:34:54.717091 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tvdg4" event={"ID":"fae8448e-393d-4700-ae60-e73ddf5c1053","Type":"ContainerDied","Data":"92111ede29fc20f3f1181547f323d0c89f3cbd911a7923e22b63da4d8d961b87"} Nov 22 08:34:57 crc kubenswrapper[4789]: I1122 08:34:57.754468 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tvdg4" event={"ID":"fae8448e-393d-4700-ae60-e73ddf5c1053","Type":"ContainerStarted","Data":"3e465f9512a39a10e60d7793c1457358d9d15bf95d4f9e4e6f273e175e714502"} Nov 22 08:34:57 crc kubenswrapper[4789]: I1122 08:34:57.772967 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-tvdg4" podStartSLOduration=2.715124318 podStartE2EDuration="8.772949263s" podCreationTimestamp="2025-11-22 08:34:49 +0000 UTC" firstStartedPulling="2025-11-22 08:34:50.675422708 +0000 UTC m=+2304.909822991" lastFinishedPulling="2025-11-22 08:34:56.733247663 +0000 UTC m=+2310.967647936" observedRunningTime="2025-11-22 08:34:57.770622049 +0000 UTC m=+2312.005022332" watchObservedRunningTime="2025-11-22 08:34:57.772949263 +0000 UTC m=+2312.007349536" Nov 22 08:34:57 crc kubenswrapper[4789]: I1122 08:34:57.972213 4789 scope.go:117] "RemoveContainer" containerID="c7898b5c08854d5df6033d34564729f053b077e4fcadc5a1ea77706f1981be3b" Nov 22 08:34:57 crc kubenswrapper[4789]: E1122 08:34:57.972559 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 08:34:59 crc kubenswrapper[4789]: I1122 08:34:59.544414 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-tvdg4" Nov 22 08:34:59 crc kubenswrapper[4789]: I1122 08:34:59.544981 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-tvdg4" Nov 22 08:34:59 crc kubenswrapper[4789]: I1122 08:34:59.597492 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-tvdg4" Nov 22 08:35:08 crc kubenswrapper[4789]: I1122 08:35:08.965420 4789 scope.go:117] "RemoveContainer" containerID="c7898b5c08854d5df6033d34564729f053b077e4fcadc5a1ea77706f1981be3b" Nov 22 08:35:08 crc kubenswrapper[4789]: E1122 08:35:08.966102 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 08:35:09 crc kubenswrapper[4789]: I1122 08:35:09.594392 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-tvdg4" Nov 22 08:35:09 crc kubenswrapper[4789]: I1122 08:35:09.641549 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-tvdg4"] Nov 22 08:35:09 crc kubenswrapper[4789]: I1122 08:35:09.853964 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-tvdg4" podUID="fae8448e-393d-4700-ae60-e73ddf5c1053" containerName="registry-server" containerID="cri-o://3e465f9512a39a10e60d7793c1457358d9d15bf95d4f9e4e6f273e175e714502" gracePeriod=2 Nov 22 08:35:10 crc kubenswrapper[4789]: I1122 08:35:10.311689 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tvdg4" Nov 22 08:35:10 crc kubenswrapper[4789]: I1122 08:35:10.502035 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fae8448e-393d-4700-ae60-e73ddf5c1053-catalog-content\") pod \"fae8448e-393d-4700-ae60-e73ddf5c1053\" (UID: \"fae8448e-393d-4700-ae60-e73ddf5c1053\") " Nov 22 08:35:10 crc kubenswrapper[4789]: I1122 08:35:10.502969 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5bmtl\" (UniqueName: \"kubernetes.io/projected/fae8448e-393d-4700-ae60-e73ddf5c1053-kube-api-access-5bmtl\") pod \"fae8448e-393d-4700-ae60-e73ddf5c1053\" (UID: \"fae8448e-393d-4700-ae60-e73ddf5c1053\") " Nov 22 08:35:10 crc kubenswrapper[4789]: I1122 08:35:10.503146 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fae8448e-393d-4700-ae60-e73ddf5c1053-utilities\") pod \"fae8448e-393d-4700-ae60-e73ddf5c1053\" (UID: \"fae8448e-393d-4700-ae60-e73ddf5c1053\") " Nov 22 08:35:10 crc kubenswrapper[4789]: I1122 08:35:10.504148 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fae8448e-393d-4700-ae60-e73ddf5c1053-utilities" (OuterVolumeSpecName: "utilities") pod "fae8448e-393d-4700-ae60-e73ddf5c1053" (UID: "fae8448e-393d-4700-ae60-e73ddf5c1053"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:35:10 crc kubenswrapper[4789]: I1122 08:35:10.510211 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fae8448e-393d-4700-ae60-e73ddf5c1053-kube-api-access-5bmtl" (OuterVolumeSpecName: "kube-api-access-5bmtl") pod "fae8448e-393d-4700-ae60-e73ddf5c1053" (UID: "fae8448e-393d-4700-ae60-e73ddf5c1053"). InnerVolumeSpecName "kube-api-access-5bmtl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:35:10 crc kubenswrapper[4789]: I1122 08:35:10.551635 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fae8448e-393d-4700-ae60-e73ddf5c1053-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fae8448e-393d-4700-ae60-e73ddf5c1053" (UID: "fae8448e-393d-4700-ae60-e73ddf5c1053"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:35:10 crc kubenswrapper[4789]: I1122 08:35:10.605004 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5bmtl\" (UniqueName: \"kubernetes.io/projected/fae8448e-393d-4700-ae60-e73ddf5c1053-kube-api-access-5bmtl\") on node \"crc\" DevicePath \"\"" Nov 22 08:35:10 crc kubenswrapper[4789]: I1122 08:35:10.605042 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fae8448e-393d-4700-ae60-e73ddf5c1053-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 08:35:10 crc kubenswrapper[4789]: I1122 08:35:10.605052 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fae8448e-393d-4700-ae60-e73ddf5c1053-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 08:35:10 crc kubenswrapper[4789]: I1122 08:35:10.866185 4789 generic.go:334] "Generic (PLEG): container finished" podID="fae8448e-393d-4700-ae60-e73ddf5c1053" containerID="3e465f9512a39a10e60d7793c1457358d9d15bf95d4f9e4e6f273e175e714502" exitCode=0 Nov 22 08:35:10 crc kubenswrapper[4789]: I1122 08:35:10.866261 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tvdg4" event={"ID":"fae8448e-393d-4700-ae60-e73ddf5c1053","Type":"ContainerDied","Data":"3e465f9512a39a10e60d7793c1457358d9d15bf95d4f9e4e6f273e175e714502"} Nov 22 08:35:10 crc kubenswrapper[4789]: I1122 08:35:10.866300 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tvdg4" Nov 22 08:35:10 crc kubenswrapper[4789]: I1122 08:35:10.866333 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tvdg4" event={"ID":"fae8448e-393d-4700-ae60-e73ddf5c1053","Type":"ContainerDied","Data":"15a1ae39a0d49263c395dc1823248fe391c6bc9945bffb18e10b0ddaa5437e06"} Nov 22 08:35:10 crc kubenswrapper[4789]: I1122 08:35:10.866365 4789 scope.go:117] "RemoveContainer" containerID="3e465f9512a39a10e60d7793c1457358d9d15bf95d4f9e4e6f273e175e714502" Nov 22 08:35:10 crc kubenswrapper[4789]: I1122 08:35:10.912629 4789 scope.go:117] "RemoveContainer" containerID="92111ede29fc20f3f1181547f323d0c89f3cbd911a7923e22b63da4d8d961b87" Nov 22 08:35:10 crc kubenswrapper[4789]: I1122 08:35:10.914651 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-tvdg4"] Nov 22 08:35:10 crc kubenswrapper[4789]: I1122 08:35:10.923743 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-tvdg4"] Nov 22 08:35:10 crc kubenswrapper[4789]: E1122 08:35:10.935017 4789 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfae8448e_393d_4700_ae60_e73ddf5c1053.slice\": RecentStats: unable to find data in memory cache]" Nov 22 08:35:10 crc kubenswrapper[4789]: I1122 08:35:10.950673 4789 scope.go:117] "RemoveContainer" containerID="440315f1369b446fcc72a3ec6145b7c853253bf0c69c80cbc4253fada8c1cf33" Nov 22 08:35:10 crc kubenswrapper[4789]: I1122 08:35:10.981010 4789 scope.go:117] "RemoveContainer" containerID="3e465f9512a39a10e60d7793c1457358d9d15bf95d4f9e4e6f273e175e714502" Nov 22 08:35:10 crc kubenswrapper[4789]: E1122 08:35:10.981632 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3e465f9512a39a10e60d7793c1457358d9d15bf95d4f9e4e6f273e175e714502\": container with ID starting with 3e465f9512a39a10e60d7793c1457358d9d15bf95d4f9e4e6f273e175e714502 not found: ID does not exist" containerID="3e465f9512a39a10e60d7793c1457358d9d15bf95d4f9e4e6f273e175e714502" Nov 22 08:35:10 crc kubenswrapper[4789]: I1122 08:35:10.981692 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3e465f9512a39a10e60d7793c1457358d9d15bf95d4f9e4e6f273e175e714502"} err="failed to get container status \"3e465f9512a39a10e60d7793c1457358d9d15bf95d4f9e4e6f273e175e714502\": rpc error: code = NotFound desc = could not find container \"3e465f9512a39a10e60d7793c1457358d9d15bf95d4f9e4e6f273e175e714502\": container with ID starting with 3e465f9512a39a10e60d7793c1457358d9d15bf95d4f9e4e6f273e175e714502 not found: ID does not exist" Nov 22 08:35:10 crc kubenswrapper[4789]: I1122 08:35:10.981725 4789 scope.go:117] "RemoveContainer" containerID="92111ede29fc20f3f1181547f323d0c89f3cbd911a7923e22b63da4d8d961b87" Nov 22 08:35:10 crc kubenswrapper[4789]: E1122 08:35:10.982307 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"92111ede29fc20f3f1181547f323d0c89f3cbd911a7923e22b63da4d8d961b87\": container with ID starting with 92111ede29fc20f3f1181547f323d0c89f3cbd911a7923e22b63da4d8d961b87 not found: ID does not exist" containerID="92111ede29fc20f3f1181547f323d0c89f3cbd911a7923e22b63da4d8d961b87" Nov 22 08:35:10 crc kubenswrapper[4789]: I1122 08:35:10.982343 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"92111ede29fc20f3f1181547f323d0c89f3cbd911a7923e22b63da4d8d961b87"} err="failed to get container status \"92111ede29fc20f3f1181547f323d0c89f3cbd911a7923e22b63da4d8d961b87\": rpc error: code = NotFound desc = could not find container \"92111ede29fc20f3f1181547f323d0c89f3cbd911a7923e22b63da4d8d961b87\": container with ID starting with 92111ede29fc20f3f1181547f323d0c89f3cbd911a7923e22b63da4d8d961b87 not found: ID does not exist" Nov 22 08:35:10 crc kubenswrapper[4789]: I1122 08:35:10.982374 4789 scope.go:117] "RemoveContainer" containerID="440315f1369b446fcc72a3ec6145b7c853253bf0c69c80cbc4253fada8c1cf33" Nov 22 08:35:10 crc kubenswrapper[4789]: E1122 08:35:10.982659 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"440315f1369b446fcc72a3ec6145b7c853253bf0c69c80cbc4253fada8c1cf33\": container with ID starting with 440315f1369b446fcc72a3ec6145b7c853253bf0c69c80cbc4253fada8c1cf33 not found: ID does not exist" containerID="440315f1369b446fcc72a3ec6145b7c853253bf0c69c80cbc4253fada8c1cf33" Nov 22 08:35:10 crc kubenswrapper[4789]: I1122 08:35:10.982685 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"440315f1369b446fcc72a3ec6145b7c853253bf0c69c80cbc4253fada8c1cf33"} err="failed to get container status \"440315f1369b446fcc72a3ec6145b7c853253bf0c69c80cbc4253fada8c1cf33\": rpc error: code = NotFound desc = could not find container \"440315f1369b446fcc72a3ec6145b7c853253bf0c69c80cbc4253fada8c1cf33\": container with ID starting with 440315f1369b446fcc72a3ec6145b7c853253bf0c69c80cbc4253fada8c1cf33 not found: ID does not exist" Nov 22 08:35:11 crc kubenswrapper[4789]: I1122 08:35:11.977331 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fae8448e-393d-4700-ae60-e73ddf5c1053" path="/var/lib/kubelet/pods/fae8448e-393d-4700-ae60-e73ddf5c1053/volumes" Nov 22 08:35:21 crc kubenswrapper[4789]: I1122 08:35:21.965489 4789 scope.go:117] "RemoveContainer" containerID="c7898b5c08854d5df6033d34564729f053b077e4fcadc5a1ea77706f1981be3b" Nov 22 08:35:21 crc kubenswrapper[4789]: E1122 08:35:21.966284 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 08:35:33 crc kubenswrapper[4789]: I1122 08:35:33.965685 4789 scope.go:117] "RemoveContainer" containerID="c7898b5c08854d5df6033d34564729f053b077e4fcadc5a1ea77706f1981be3b" Nov 22 08:35:33 crc kubenswrapper[4789]: E1122 08:35:33.966873 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 08:35:47 crc kubenswrapper[4789]: I1122 08:35:47.965894 4789 scope.go:117] "RemoveContainer" containerID="c7898b5c08854d5df6033d34564729f053b077e4fcadc5a1ea77706f1981be3b" Nov 22 08:35:47 crc kubenswrapper[4789]: E1122 08:35:47.966958 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 08:36:01 crc kubenswrapper[4789]: I1122 08:36:01.965363 4789 scope.go:117] "RemoveContainer" containerID="c7898b5c08854d5df6033d34564729f053b077e4fcadc5a1ea77706f1981be3b" Nov 22 08:36:01 crc kubenswrapper[4789]: E1122 08:36:01.965993 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 08:36:09 crc kubenswrapper[4789]: I1122 08:36:09.496955 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kdxvv"] Nov 22 08:36:09 crc kubenswrapper[4789]: I1122 08:36:09.507081 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-xdxkl"] Nov 22 08:36:09 crc kubenswrapper[4789]: I1122 08:36:09.514091 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8q4rs"] Nov 22 08:36:09 crc kubenswrapper[4789]: I1122 08:36:09.520174 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-8rxgw"] Nov 22 08:36:09 crc kubenswrapper[4789]: I1122 08:36:09.527736 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qp94q"] Nov 22 08:36:09 crc kubenswrapper[4789]: I1122 08:36:09.534252 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-k2c8r"] Nov 22 08:36:09 crc kubenswrapper[4789]: I1122 08:36:09.539705 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-gt7gt"] Nov 22 08:36:09 crc kubenswrapper[4789]: I1122 08:36:09.545629 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xqcfq"] Nov 22 08:36:09 crc kubenswrapper[4789]: I1122 08:36:09.551728 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-8rxgw"] Nov 22 08:36:09 crc kubenswrapper[4789]: I1122 08:36:09.557472 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kdxvv"] Nov 22 08:36:09 crc kubenswrapper[4789]: I1122 08:36:09.562995 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7qscl"] Nov 22 08:36:09 crc kubenswrapper[4789]: I1122 08:36:09.568770 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8q4rs"] Nov 22 08:36:09 crc kubenswrapper[4789]: I1122 08:36:09.574636 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-75ls5"] Nov 22 08:36:09 crc kubenswrapper[4789]: I1122 08:36:09.580863 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qp94q"] Nov 22 08:36:09 crc kubenswrapper[4789]: I1122 08:36:09.587043 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-75ls5"] Nov 22 08:36:09 crc kubenswrapper[4789]: I1122 08:36:09.593511 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-k2c8r"] Nov 22 08:36:09 crc kubenswrapper[4789]: I1122 08:36:09.599416 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xqcfq"] Nov 22 08:36:09 crc kubenswrapper[4789]: I1122 08:36:09.605524 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-gt7gt"] Nov 22 08:36:09 crc kubenswrapper[4789]: I1122 08:36:09.611342 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7qscl"] Nov 22 08:36:09 crc kubenswrapper[4789]: I1122 08:36:09.617160 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-xdxkl"] Nov 22 08:36:09 crc kubenswrapper[4789]: I1122 08:36:09.979124 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0eba14cc-decf-4b92-adbb-3cf60adc26f3" path="/var/lib/kubelet/pods/0eba14cc-decf-4b92-adbb-3cf60adc26f3/volumes" Nov 22 08:36:09 crc kubenswrapper[4789]: I1122 08:36:09.980106 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0f452ff9-9dbd-4504-a089-86c16fcd690d" path="/var/lib/kubelet/pods/0f452ff9-9dbd-4504-a089-86c16fcd690d/volumes" Nov 22 08:36:09 crc kubenswrapper[4789]: I1122 08:36:09.980687 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="276e9f73-f469-44c1-b346-3cca0768cc87" path="/var/lib/kubelet/pods/276e9f73-f469-44c1-b346-3cca0768cc87/volumes" Nov 22 08:36:09 crc kubenswrapper[4789]: I1122 08:36:09.981333 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="393d8a04-e1ff-4707-bf88-964105ed2008" path="/var/lib/kubelet/pods/393d8a04-e1ff-4707-bf88-964105ed2008/volumes" Nov 22 08:36:09 crc kubenswrapper[4789]: I1122 08:36:09.982516 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="424c42d2-3148-4308-91a1-9935d0ab2441" path="/var/lib/kubelet/pods/424c42d2-3148-4308-91a1-9935d0ab2441/volumes" Nov 22 08:36:09 crc kubenswrapper[4789]: I1122 08:36:09.983191 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4852655b-677f-4933-b297-f7e78ec4c037" path="/var/lib/kubelet/pods/4852655b-677f-4933-b297-f7e78ec4c037/volumes" Nov 22 08:36:09 crc kubenswrapper[4789]: I1122 08:36:09.983865 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4cac03ab-292f-4255-aec2-76120f6f8c24" path="/var/lib/kubelet/pods/4cac03ab-292f-4255-aec2-76120f6f8c24/volumes" Nov 22 08:36:09 crc kubenswrapper[4789]: I1122 08:36:09.985075 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8470bb71-39d5-4a92-b60e-856624688535" path="/var/lib/kubelet/pods/8470bb71-39d5-4a92-b60e-856624688535/volumes" Nov 22 08:36:09 crc kubenswrapper[4789]: I1122 08:36:09.985677 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bcf20d51-f782-4c2a-b093-489e9cbd300e" path="/var/lib/kubelet/pods/bcf20d51-f782-4c2a-b093-489e9cbd300e/volumes" Nov 22 08:36:09 crc kubenswrapper[4789]: I1122 08:36:09.986380 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fb964d35-f23a-4c63-a480-eeeea1e2f305" path="/var/lib/kubelet/pods/fb964d35-f23a-4c63-a480-eeeea1e2f305/volumes" Nov 22 08:36:14 crc kubenswrapper[4789]: I1122 08:36:14.965593 4789 scope.go:117] "RemoveContainer" containerID="c7898b5c08854d5df6033d34564729f053b077e4fcadc5a1ea77706f1981be3b" Nov 22 08:36:14 crc kubenswrapper[4789]: E1122 08:36:14.966273 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 08:36:15 crc kubenswrapper[4789]: I1122 08:36:15.568575 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qmwns"] Nov 22 08:36:15 crc kubenswrapper[4789]: E1122 08:36:15.569456 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fae8448e-393d-4700-ae60-e73ddf5c1053" containerName="extract-utilities" Nov 22 08:36:15 crc kubenswrapper[4789]: I1122 08:36:15.569489 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="fae8448e-393d-4700-ae60-e73ddf5c1053" containerName="extract-utilities" Nov 22 08:36:15 crc kubenswrapper[4789]: E1122 08:36:15.569527 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fae8448e-393d-4700-ae60-e73ddf5c1053" containerName="extract-content" Nov 22 08:36:15 crc kubenswrapper[4789]: I1122 08:36:15.569539 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="fae8448e-393d-4700-ae60-e73ddf5c1053" containerName="extract-content" Nov 22 08:36:15 crc kubenswrapper[4789]: E1122 08:36:15.569569 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fae8448e-393d-4700-ae60-e73ddf5c1053" containerName="registry-server" Nov 22 08:36:15 crc kubenswrapper[4789]: I1122 08:36:15.569581 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="fae8448e-393d-4700-ae60-e73ddf5c1053" containerName="registry-server" Nov 22 08:36:15 crc kubenswrapper[4789]: I1122 08:36:15.569922 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="fae8448e-393d-4700-ae60-e73ddf5c1053" containerName="registry-server" Nov 22 08:36:15 crc kubenswrapper[4789]: I1122 08:36:15.571015 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qmwns" Nov 22 08:36:15 crc kubenswrapper[4789]: I1122 08:36:15.573273 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 22 08:36:15 crc kubenswrapper[4789]: I1122 08:36:15.573520 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-glmr7" Nov 22 08:36:15 crc kubenswrapper[4789]: I1122 08:36:15.573586 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 22 08:36:15 crc kubenswrapper[4789]: I1122 08:36:15.574138 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 22 08:36:15 crc kubenswrapper[4789]: I1122 08:36:15.576009 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qmwns"] Nov 22 08:36:15 crc kubenswrapper[4789]: I1122 08:36:15.578240 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 22 08:36:15 crc kubenswrapper[4789]: I1122 08:36:15.603159 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6f03e6f2-254d-4e82-bdda-04a73ae3b36a-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-qmwns\" (UID: \"6f03e6f2-254d-4e82-bdda-04a73ae3b36a\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qmwns" Nov 22 08:36:15 crc kubenswrapper[4789]: I1122 08:36:15.603337 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6f03e6f2-254d-4e82-bdda-04a73ae3b36a-ceph\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-qmwns\" (UID: \"6f03e6f2-254d-4e82-bdda-04a73ae3b36a\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qmwns" Nov 22 08:36:15 crc kubenswrapper[4789]: I1122 08:36:15.603402 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xdpkf\" (UniqueName: \"kubernetes.io/projected/6f03e6f2-254d-4e82-bdda-04a73ae3b36a-kube-api-access-xdpkf\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-qmwns\" (UID: \"6f03e6f2-254d-4e82-bdda-04a73ae3b36a\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qmwns" Nov 22 08:36:15 crc kubenswrapper[4789]: I1122 08:36:15.603440 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f03e6f2-254d-4e82-bdda-04a73ae3b36a-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-qmwns\" (UID: \"6f03e6f2-254d-4e82-bdda-04a73ae3b36a\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qmwns" Nov 22 08:36:15 crc kubenswrapper[4789]: I1122 08:36:15.603561 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6f03e6f2-254d-4e82-bdda-04a73ae3b36a-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-qmwns\" (UID: \"6f03e6f2-254d-4e82-bdda-04a73ae3b36a\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qmwns" Nov 22 08:36:15 crc kubenswrapper[4789]: I1122 08:36:15.705272 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f03e6f2-254d-4e82-bdda-04a73ae3b36a-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-qmwns\" (UID: \"6f03e6f2-254d-4e82-bdda-04a73ae3b36a\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qmwns" Nov 22 08:36:15 crc kubenswrapper[4789]: I1122 08:36:15.705353 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6f03e6f2-254d-4e82-bdda-04a73ae3b36a-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-qmwns\" (UID: \"6f03e6f2-254d-4e82-bdda-04a73ae3b36a\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qmwns" Nov 22 08:36:15 crc kubenswrapper[4789]: I1122 08:36:15.705407 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6f03e6f2-254d-4e82-bdda-04a73ae3b36a-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-qmwns\" (UID: \"6f03e6f2-254d-4e82-bdda-04a73ae3b36a\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qmwns" Nov 22 08:36:15 crc kubenswrapper[4789]: I1122 08:36:15.705464 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6f03e6f2-254d-4e82-bdda-04a73ae3b36a-ceph\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-qmwns\" (UID: \"6f03e6f2-254d-4e82-bdda-04a73ae3b36a\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qmwns" Nov 22 08:36:15 crc kubenswrapper[4789]: I1122 08:36:15.705492 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xdpkf\" (UniqueName: \"kubernetes.io/projected/6f03e6f2-254d-4e82-bdda-04a73ae3b36a-kube-api-access-xdpkf\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-qmwns\" (UID: \"6f03e6f2-254d-4e82-bdda-04a73ae3b36a\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qmwns" Nov 22 08:36:15 crc kubenswrapper[4789]: I1122 08:36:15.711488 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6f03e6f2-254d-4e82-bdda-04a73ae3b36a-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-qmwns\" (UID: \"6f03e6f2-254d-4e82-bdda-04a73ae3b36a\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qmwns" Nov 22 08:36:15 crc kubenswrapper[4789]: I1122 08:36:15.711677 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6f03e6f2-254d-4e82-bdda-04a73ae3b36a-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-qmwns\" (UID: \"6f03e6f2-254d-4e82-bdda-04a73ae3b36a\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qmwns" Nov 22 08:36:15 crc kubenswrapper[4789]: I1122 08:36:15.711848 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f03e6f2-254d-4e82-bdda-04a73ae3b36a-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-qmwns\" (UID: \"6f03e6f2-254d-4e82-bdda-04a73ae3b36a\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qmwns" Nov 22 08:36:15 crc kubenswrapper[4789]: I1122 08:36:15.711864 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6f03e6f2-254d-4e82-bdda-04a73ae3b36a-ceph\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-qmwns\" (UID: \"6f03e6f2-254d-4e82-bdda-04a73ae3b36a\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qmwns" Nov 22 08:36:15 crc kubenswrapper[4789]: I1122 08:36:15.726358 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xdpkf\" (UniqueName: \"kubernetes.io/projected/6f03e6f2-254d-4e82-bdda-04a73ae3b36a-kube-api-access-xdpkf\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-qmwns\" (UID: \"6f03e6f2-254d-4e82-bdda-04a73ae3b36a\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qmwns" Nov 22 08:36:15 crc kubenswrapper[4789]: I1122 08:36:15.893305 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qmwns" Nov 22 08:36:16 crc kubenswrapper[4789]: I1122 08:36:16.373193 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qmwns"] Nov 22 08:36:16 crc kubenswrapper[4789]: I1122 08:36:16.385200 4789 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 22 08:36:16 crc kubenswrapper[4789]: I1122 08:36:16.409882 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qmwns" event={"ID":"6f03e6f2-254d-4e82-bdda-04a73ae3b36a","Type":"ContainerStarted","Data":"67784c1ec198829ada2b2c19ee6c919b3e88e11596aedd329a01bd28505a57db"} Nov 22 08:36:17 crc kubenswrapper[4789]: I1122 08:36:17.419647 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qmwns" event={"ID":"6f03e6f2-254d-4e82-bdda-04a73ae3b36a","Type":"ContainerStarted","Data":"8c083142358a379594e03cfa0549680767b7a8e16ccfd40c179d6f1de557b0bb"} Nov 22 08:36:17 crc kubenswrapper[4789]: I1122 08:36:17.442737 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qmwns" podStartSLOduration=1.9945842040000001 podStartE2EDuration="2.442712902s" podCreationTimestamp="2025-11-22 08:36:15 +0000 UTC" firstStartedPulling="2025-11-22 08:36:16.384970933 +0000 UTC m=+2390.619371206" lastFinishedPulling="2025-11-22 08:36:16.833099631 +0000 UTC m=+2391.067499904" observedRunningTime="2025-11-22 08:36:17.438920489 +0000 UTC m=+2391.673320782" watchObservedRunningTime="2025-11-22 08:36:17.442712902 +0000 UTC m=+2391.677113175" Nov 22 08:36:26 crc kubenswrapper[4789]: I1122 08:36:26.685461 4789 scope.go:117] "RemoveContainer" containerID="645d7595e012c68741c83ef8491e3fc0e745a63b28c053fe594370ee4c2d89b5" Nov 22 08:36:26 crc kubenswrapper[4789]: I1122 08:36:26.747159 4789 scope.go:117] "RemoveContainer" containerID="099f7b3f545e289261b5f69258034452bfecf664007fa741897156878c82d262" Nov 22 08:36:26 crc kubenswrapper[4789]: I1122 08:36:26.775103 4789 scope.go:117] "RemoveContainer" containerID="4d4f6ad56c7b19b2010140fc50c9e595f63e55d53556011d40962dbd1a9d123d" Nov 22 08:36:26 crc kubenswrapper[4789]: I1122 08:36:26.826992 4789 scope.go:117] "RemoveContainer" containerID="e1de92c1a996b7edeb1227765e69a94a410b28c11e3c8a73369fab521842f428" Nov 22 08:36:26 crc kubenswrapper[4789]: I1122 08:36:26.879323 4789 scope.go:117] "RemoveContainer" containerID="f3cc3f695f3c7ebe4b186d126800d634b396ab3b012164684186dce8bcc05b10" Nov 22 08:36:26 crc kubenswrapper[4789]: I1122 08:36:26.923590 4789 scope.go:117] "RemoveContainer" containerID="a60cf380354af9c3e52384cb95aab1c1637904d531e3dc1cbe439dd62de69622" Nov 22 08:36:26 crc kubenswrapper[4789]: I1122 08:36:26.970652 4789 scope.go:117] "RemoveContainer" containerID="5741378a3aff31bf12bdcc33ca2eb5dda6d02fa7bd0ecc769b25e098f493849c" Nov 22 08:36:26 crc kubenswrapper[4789]: I1122 08:36:26.996562 4789 scope.go:117] "RemoveContainer" containerID="7c88ffd385b6dc5704d986c3f01a32015907ab893d1061705ab6115b81533966" Nov 22 08:36:27 crc kubenswrapper[4789]: I1122 08:36:27.040204 4789 scope.go:117] "RemoveContainer" containerID="1357b63657791ed648b6839cf628c9f025201f14b21c296b033237e330e7ac34" Nov 22 08:36:27 crc kubenswrapper[4789]: I1122 08:36:27.113959 4789 scope.go:117] "RemoveContainer" containerID="a029cef305ad2ebf2b3aef186e8ffd588cb9295debda093f4487b22fe6bc9363" Nov 22 08:36:27 crc kubenswrapper[4789]: I1122 08:36:27.971663 4789 scope.go:117] "RemoveContainer" containerID="c7898b5c08854d5df6033d34564729f053b077e4fcadc5a1ea77706f1981be3b" Nov 22 08:36:27 crc kubenswrapper[4789]: E1122 08:36:27.971958 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 08:36:28 crc kubenswrapper[4789]: I1122 08:36:28.517776 4789 generic.go:334] "Generic (PLEG): container finished" podID="6f03e6f2-254d-4e82-bdda-04a73ae3b36a" containerID="8c083142358a379594e03cfa0549680767b7a8e16ccfd40c179d6f1de557b0bb" exitCode=0 Nov 22 08:36:28 crc kubenswrapper[4789]: I1122 08:36:28.517857 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qmwns" event={"ID":"6f03e6f2-254d-4e82-bdda-04a73ae3b36a","Type":"ContainerDied","Data":"8c083142358a379594e03cfa0549680767b7a8e16ccfd40c179d6f1de557b0bb"} Nov 22 08:36:29 crc kubenswrapper[4789]: I1122 08:36:29.898982 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qmwns" Nov 22 08:36:29 crc kubenswrapper[4789]: I1122 08:36:29.961439 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6f03e6f2-254d-4e82-bdda-04a73ae3b36a-ceph\") pod \"6f03e6f2-254d-4e82-bdda-04a73ae3b36a\" (UID: \"6f03e6f2-254d-4e82-bdda-04a73ae3b36a\") " Nov 22 08:36:29 crc kubenswrapper[4789]: I1122 08:36:29.961518 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6f03e6f2-254d-4e82-bdda-04a73ae3b36a-inventory\") pod \"6f03e6f2-254d-4e82-bdda-04a73ae3b36a\" (UID: \"6f03e6f2-254d-4e82-bdda-04a73ae3b36a\") " Nov 22 08:36:29 crc kubenswrapper[4789]: I1122 08:36:29.961547 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6f03e6f2-254d-4e82-bdda-04a73ae3b36a-ssh-key\") pod \"6f03e6f2-254d-4e82-bdda-04a73ae3b36a\" (UID: \"6f03e6f2-254d-4e82-bdda-04a73ae3b36a\") " Nov 22 08:36:29 crc kubenswrapper[4789]: I1122 08:36:29.961588 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f03e6f2-254d-4e82-bdda-04a73ae3b36a-repo-setup-combined-ca-bundle\") pod \"6f03e6f2-254d-4e82-bdda-04a73ae3b36a\" (UID: \"6f03e6f2-254d-4e82-bdda-04a73ae3b36a\") " Nov 22 08:36:29 crc kubenswrapper[4789]: I1122 08:36:29.961657 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xdpkf\" (UniqueName: \"kubernetes.io/projected/6f03e6f2-254d-4e82-bdda-04a73ae3b36a-kube-api-access-xdpkf\") pod \"6f03e6f2-254d-4e82-bdda-04a73ae3b36a\" (UID: \"6f03e6f2-254d-4e82-bdda-04a73ae3b36a\") " Nov 22 08:36:29 crc kubenswrapper[4789]: I1122 08:36:29.987054 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f03e6f2-254d-4e82-bdda-04a73ae3b36a-ceph" (OuterVolumeSpecName: "ceph") pod "6f03e6f2-254d-4e82-bdda-04a73ae3b36a" (UID: "6f03e6f2-254d-4e82-bdda-04a73ae3b36a"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:36:29 crc kubenswrapper[4789]: I1122 08:36:29.992971 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f03e6f2-254d-4e82-bdda-04a73ae3b36a-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "6f03e6f2-254d-4e82-bdda-04a73ae3b36a" (UID: "6f03e6f2-254d-4e82-bdda-04a73ae3b36a"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:36:30 crc kubenswrapper[4789]: I1122 08:36:30.019968 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6f03e6f2-254d-4e82-bdda-04a73ae3b36a-kube-api-access-xdpkf" (OuterVolumeSpecName: "kube-api-access-xdpkf") pod "6f03e6f2-254d-4e82-bdda-04a73ae3b36a" (UID: "6f03e6f2-254d-4e82-bdda-04a73ae3b36a"). InnerVolumeSpecName "kube-api-access-xdpkf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:36:30 crc kubenswrapper[4789]: I1122 08:36:30.053984 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f03e6f2-254d-4e82-bdda-04a73ae3b36a-inventory" (OuterVolumeSpecName: "inventory") pod "6f03e6f2-254d-4e82-bdda-04a73ae3b36a" (UID: "6f03e6f2-254d-4e82-bdda-04a73ae3b36a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:36:30 crc kubenswrapper[4789]: I1122 08:36:30.072950 4789 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6f03e6f2-254d-4e82-bdda-04a73ae3b36a-ceph\") on node \"crc\" DevicePath \"\"" Nov 22 08:36:30 crc kubenswrapper[4789]: I1122 08:36:30.072989 4789 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6f03e6f2-254d-4e82-bdda-04a73ae3b36a-inventory\") on node \"crc\" DevicePath \"\"" Nov 22 08:36:30 crc kubenswrapper[4789]: I1122 08:36:30.072999 4789 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f03e6f2-254d-4e82-bdda-04a73ae3b36a-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:36:30 crc kubenswrapper[4789]: I1122 08:36:30.073009 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xdpkf\" (UniqueName: \"kubernetes.io/projected/6f03e6f2-254d-4e82-bdda-04a73ae3b36a-kube-api-access-xdpkf\") on node \"crc\" DevicePath \"\"" Nov 22 08:36:30 crc kubenswrapper[4789]: I1122 08:36:30.114381 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f03e6f2-254d-4e82-bdda-04a73ae3b36a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "6f03e6f2-254d-4e82-bdda-04a73ae3b36a" (UID: "6f03e6f2-254d-4e82-bdda-04a73ae3b36a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:36:30 crc kubenswrapper[4789]: I1122 08:36:30.174648 4789 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6f03e6f2-254d-4e82-bdda-04a73ae3b36a-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 22 08:36:30 crc kubenswrapper[4789]: I1122 08:36:30.533594 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qmwns" event={"ID":"6f03e6f2-254d-4e82-bdda-04a73ae3b36a","Type":"ContainerDied","Data":"67784c1ec198829ada2b2c19ee6c919b3e88e11596aedd329a01bd28505a57db"} Nov 22 08:36:30 crc kubenswrapper[4789]: I1122 08:36:30.533838 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="67784c1ec198829ada2b2c19ee6c919b3e88e11596aedd329a01bd28505a57db" Nov 22 08:36:30 crc kubenswrapper[4789]: I1122 08:36:30.533886 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qmwns" Nov 22 08:36:30 crc kubenswrapper[4789]: I1122 08:36:30.607784 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bxq52"] Nov 22 08:36:30 crc kubenswrapper[4789]: E1122 08:36:30.608241 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f03e6f2-254d-4e82-bdda-04a73ae3b36a" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 22 08:36:30 crc kubenswrapper[4789]: I1122 08:36:30.608267 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f03e6f2-254d-4e82-bdda-04a73ae3b36a" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 22 08:36:30 crc kubenswrapper[4789]: I1122 08:36:30.608484 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f03e6f2-254d-4e82-bdda-04a73ae3b36a" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 22 08:36:30 crc kubenswrapper[4789]: I1122 08:36:30.609318 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bxq52" Nov 22 08:36:30 crc kubenswrapper[4789]: I1122 08:36:30.611769 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 22 08:36:30 crc kubenswrapper[4789]: I1122 08:36:30.611955 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 22 08:36:30 crc kubenswrapper[4789]: I1122 08:36:30.612346 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-glmr7" Nov 22 08:36:30 crc kubenswrapper[4789]: I1122 08:36:30.612621 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 22 08:36:30 crc kubenswrapper[4789]: I1122 08:36:30.615319 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 22 08:36:30 crc kubenswrapper[4789]: I1122 08:36:30.622256 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bxq52"] Nov 22 08:36:30 crc kubenswrapper[4789]: I1122 08:36:30.682458 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fmzwd\" (UniqueName: \"kubernetes.io/projected/61e96af2-fd7c-4c68-959b-db900528ab2e-kube-api-access-fmzwd\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-bxq52\" (UID: \"61e96af2-fd7c-4c68-959b-db900528ab2e\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bxq52" Nov 22 08:36:30 crc kubenswrapper[4789]: I1122 08:36:30.682874 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/61e96af2-fd7c-4c68-959b-db900528ab2e-ceph\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-bxq52\" (UID: \"61e96af2-fd7c-4c68-959b-db900528ab2e\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bxq52" Nov 22 08:36:30 crc kubenswrapper[4789]: I1122 08:36:30.683055 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/61e96af2-fd7c-4c68-959b-db900528ab2e-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-bxq52\" (UID: \"61e96af2-fd7c-4c68-959b-db900528ab2e\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bxq52" Nov 22 08:36:30 crc kubenswrapper[4789]: I1122 08:36:30.683120 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61e96af2-fd7c-4c68-959b-db900528ab2e-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-bxq52\" (UID: \"61e96af2-fd7c-4c68-959b-db900528ab2e\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bxq52" Nov 22 08:36:30 crc kubenswrapper[4789]: I1122 08:36:30.683242 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/61e96af2-fd7c-4c68-959b-db900528ab2e-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-bxq52\" (UID: \"61e96af2-fd7c-4c68-959b-db900528ab2e\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bxq52" Nov 22 08:36:30 crc kubenswrapper[4789]: I1122 08:36:30.784997 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/61e96af2-fd7c-4c68-959b-db900528ab2e-ceph\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-bxq52\" (UID: \"61e96af2-fd7c-4c68-959b-db900528ab2e\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bxq52" Nov 22 08:36:30 crc kubenswrapper[4789]: I1122 08:36:30.785078 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/61e96af2-fd7c-4c68-959b-db900528ab2e-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-bxq52\" (UID: \"61e96af2-fd7c-4c68-959b-db900528ab2e\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bxq52" Nov 22 08:36:30 crc kubenswrapper[4789]: I1122 08:36:30.785104 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61e96af2-fd7c-4c68-959b-db900528ab2e-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-bxq52\" (UID: \"61e96af2-fd7c-4c68-959b-db900528ab2e\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bxq52" Nov 22 08:36:30 crc kubenswrapper[4789]: I1122 08:36:30.785134 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/61e96af2-fd7c-4c68-959b-db900528ab2e-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-bxq52\" (UID: \"61e96af2-fd7c-4c68-959b-db900528ab2e\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bxq52" Nov 22 08:36:30 crc kubenswrapper[4789]: I1122 08:36:30.785161 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fmzwd\" (UniqueName: \"kubernetes.io/projected/61e96af2-fd7c-4c68-959b-db900528ab2e-kube-api-access-fmzwd\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-bxq52\" (UID: \"61e96af2-fd7c-4c68-959b-db900528ab2e\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bxq52" Nov 22 08:36:30 crc kubenswrapper[4789]: I1122 08:36:30.789652 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/61e96af2-fd7c-4c68-959b-db900528ab2e-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-bxq52\" (UID: \"61e96af2-fd7c-4c68-959b-db900528ab2e\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bxq52" Nov 22 08:36:30 crc kubenswrapper[4789]: I1122 08:36:30.789929 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/61e96af2-fd7c-4c68-959b-db900528ab2e-ceph\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-bxq52\" (UID: \"61e96af2-fd7c-4c68-959b-db900528ab2e\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bxq52" Nov 22 08:36:30 crc kubenswrapper[4789]: I1122 08:36:30.790314 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/61e96af2-fd7c-4c68-959b-db900528ab2e-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-bxq52\" (UID: \"61e96af2-fd7c-4c68-959b-db900528ab2e\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bxq52" Nov 22 08:36:30 crc kubenswrapper[4789]: I1122 08:36:30.790873 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61e96af2-fd7c-4c68-959b-db900528ab2e-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-bxq52\" (UID: \"61e96af2-fd7c-4c68-959b-db900528ab2e\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bxq52" Nov 22 08:36:30 crc kubenswrapper[4789]: I1122 08:36:30.802196 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fmzwd\" (UniqueName: \"kubernetes.io/projected/61e96af2-fd7c-4c68-959b-db900528ab2e-kube-api-access-fmzwd\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-bxq52\" (UID: \"61e96af2-fd7c-4c68-959b-db900528ab2e\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bxq52" Nov 22 08:36:30 crc kubenswrapper[4789]: I1122 08:36:30.925518 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bxq52" Nov 22 08:36:31 crc kubenswrapper[4789]: I1122 08:36:31.440193 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bxq52"] Nov 22 08:36:31 crc kubenswrapper[4789]: W1122 08:36:31.450466 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod61e96af2_fd7c_4c68_959b_db900528ab2e.slice/crio-98ebde52536c286b7432ff1fee112ad3475f7be08439cb832dbe2a361f0f8fab WatchSource:0}: Error finding container 98ebde52536c286b7432ff1fee112ad3475f7be08439cb832dbe2a361f0f8fab: Status 404 returned error can't find the container with id 98ebde52536c286b7432ff1fee112ad3475f7be08439cb832dbe2a361f0f8fab Nov 22 08:36:31 crc kubenswrapper[4789]: I1122 08:36:31.544482 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bxq52" event={"ID":"61e96af2-fd7c-4c68-959b-db900528ab2e","Type":"ContainerStarted","Data":"98ebde52536c286b7432ff1fee112ad3475f7be08439cb832dbe2a361f0f8fab"} Nov 22 08:36:32 crc kubenswrapper[4789]: I1122 08:36:32.552633 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bxq52" event={"ID":"61e96af2-fd7c-4c68-959b-db900528ab2e","Type":"ContainerStarted","Data":"fd8ad258d174db3b61d2a60dfb35dfc9279cb4d31ee2e506969a3e70870d283b"} Nov 22 08:36:32 crc kubenswrapper[4789]: I1122 08:36:32.576207 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bxq52" podStartSLOduration=2.06003045 podStartE2EDuration="2.576189873s" podCreationTimestamp="2025-11-22 08:36:30 +0000 UTC" firstStartedPulling="2025-11-22 08:36:31.453278202 +0000 UTC m=+2405.687678475" lastFinishedPulling="2025-11-22 08:36:31.969437625 +0000 UTC m=+2406.203837898" observedRunningTime="2025-11-22 08:36:32.572656578 +0000 UTC m=+2406.807056851" watchObservedRunningTime="2025-11-22 08:36:32.576189873 +0000 UTC m=+2406.810590146" Nov 22 08:36:39 crc kubenswrapper[4789]: I1122 08:36:39.965671 4789 scope.go:117] "RemoveContainer" containerID="c7898b5c08854d5df6033d34564729f053b077e4fcadc5a1ea77706f1981be3b" Nov 22 08:36:39 crc kubenswrapper[4789]: E1122 08:36:39.966444 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 08:36:51 crc kubenswrapper[4789]: I1122 08:36:51.965041 4789 scope.go:117] "RemoveContainer" containerID="c7898b5c08854d5df6033d34564729f053b077e4fcadc5a1ea77706f1981be3b" Nov 22 08:36:51 crc kubenswrapper[4789]: E1122 08:36:51.965725 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 08:37:03 crc kubenswrapper[4789]: I1122 08:37:03.965926 4789 scope.go:117] "RemoveContainer" containerID="c7898b5c08854d5df6033d34564729f053b077e4fcadc5a1ea77706f1981be3b" Nov 22 08:37:03 crc kubenswrapper[4789]: E1122 08:37:03.967304 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 08:37:15 crc kubenswrapper[4789]: I1122 08:37:15.964598 4789 scope.go:117] "RemoveContainer" containerID="c7898b5c08854d5df6033d34564729f053b077e4fcadc5a1ea77706f1981be3b" Nov 22 08:37:15 crc kubenswrapper[4789]: E1122 08:37:15.965467 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 08:37:26 crc kubenswrapper[4789]: I1122 08:37:26.965364 4789 scope.go:117] "RemoveContainer" containerID="c7898b5c08854d5df6033d34564729f053b077e4fcadc5a1ea77706f1981be3b" Nov 22 08:37:26 crc kubenswrapper[4789]: E1122 08:37:26.966279 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 08:37:27 crc kubenswrapper[4789]: I1122 08:37:27.282898 4789 scope.go:117] "RemoveContainer" containerID="207d6f9bbb91889ddc8585c03b2852eacc08d7623d46d65aec7db7e3be62e8c3" Nov 22 08:37:27 crc kubenswrapper[4789]: I1122 08:37:27.306897 4789 scope.go:117] "RemoveContainer" containerID="9bf53ba8c3fe35cc355f92ed91df5e251b55325fe7e741f023fe2e0e347ba99d" Nov 22 08:37:27 crc kubenswrapper[4789]: I1122 08:37:27.359120 4789 scope.go:117] "RemoveContainer" containerID="3e17bcb1c3576a168118fbdebf1a03387615ebdb139dabad7f395d11f78069f3" Nov 22 08:37:39 crc kubenswrapper[4789]: I1122 08:37:39.965315 4789 scope.go:117] "RemoveContainer" containerID="c7898b5c08854d5df6033d34564729f053b077e4fcadc5a1ea77706f1981be3b" Nov 22 08:37:39 crc kubenswrapper[4789]: E1122 08:37:39.966138 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 08:37:51 crc kubenswrapper[4789]: I1122 08:37:51.965336 4789 scope.go:117] "RemoveContainer" containerID="c7898b5c08854d5df6033d34564729f053b077e4fcadc5a1ea77706f1981be3b" Nov 22 08:37:51 crc kubenswrapper[4789]: E1122 08:37:51.966148 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 08:38:04 crc kubenswrapper[4789]: I1122 08:38:04.966158 4789 scope.go:117] "RemoveContainer" containerID="c7898b5c08854d5df6033d34564729f053b077e4fcadc5a1ea77706f1981be3b" Nov 22 08:38:04 crc kubenswrapper[4789]: E1122 08:38:04.966848 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 08:38:11 crc kubenswrapper[4789]: I1122 08:38:11.406487 4789 generic.go:334] "Generic (PLEG): container finished" podID="61e96af2-fd7c-4c68-959b-db900528ab2e" containerID="fd8ad258d174db3b61d2a60dfb35dfc9279cb4d31ee2e506969a3e70870d283b" exitCode=0 Nov 22 08:38:11 crc kubenswrapper[4789]: I1122 08:38:11.406575 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bxq52" event={"ID":"61e96af2-fd7c-4c68-959b-db900528ab2e","Type":"ContainerDied","Data":"fd8ad258d174db3b61d2a60dfb35dfc9279cb4d31ee2e506969a3e70870d283b"} Nov 22 08:38:12 crc kubenswrapper[4789]: I1122 08:38:12.829858 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bxq52" Nov 22 08:38:12 crc kubenswrapper[4789]: I1122 08:38:12.934677 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/61e96af2-fd7c-4c68-959b-db900528ab2e-ssh-key\") pod \"61e96af2-fd7c-4c68-959b-db900528ab2e\" (UID: \"61e96af2-fd7c-4c68-959b-db900528ab2e\") " Nov 22 08:38:12 crc kubenswrapper[4789]: I1122 08:38:12.934794 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fmzwd\" (UniqueName: \"kubernetes.io/projected/61e96af2-fd7c-4c68-959b-db900528ab2e-kube-api-access-fmzwd\") pod \"61e96af2-fd7c-4c68-959b-db900528ab2e\" (UID: \"61e96af2-fd7c-4c68-959b-db900528ab2e\") " Nov 22 08:38:12 crc kubenswrapper[4789]: I1122 08:38:12.934894 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61e96af2-fd7c-4c68-959b-db900528ab2e-bootstrap-combined-ca-bundle\") pod \"61e96af2-fd7c-4c68-959b-db900528ab2e\" (UID: \"61e96af2-fd7c-4c68-959b-db900528ab2e\") " Nov 22 08:38:12 crc kubenswrapper[4789]: I1122 08:38:12.934961 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/61e96af2-fd7c-4c68-959b-db900528ab2e-inventory\") pod \"61e96af2-fd7c-4c68-959b-db900528ab2e\" (UID: \"61e96af2-fd7c-4c68-959b-db900528ab2e\") " Nov 22 08:38:12 crc kubenswrapper[4789]: I1122 08:38:12.934994 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/61e96af2-fd7c-4c68-959b-db900528ab2e-ceph\") pod \"61e96af2-fd7c-4c68-959b-db900528ab2e\" (UID: \"61e96af2-fd7c-4c68-959b-db900528ab2e\") " Nov 22 08:38:12 crc kubenswrapper[4789]: I1122 08:38:12.940910 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/61e96af2-fd7c-4c68-959b-db900528ab2e-kube-api-access-fmzwd" (OuterVolumeSpecName: "kube-api-access-fmzwd") pod "61e96af2-fd7c-4c68-959b-db900528ab2e" (UID: "61e96af2-fd7c-4c68-959b-db900528ab2e"). InnerVolumeSpecName "kube-api-access-fmzwd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:38:12 crc kubenswrapper[4789]: I1122 08:38:12.941508 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61e96af2-fd7c-4c68-959b-db900528ab2e-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "61e96af2-fd7c-4c68-959b-db900528ab2e" (UID: "61e96af2-fd7c-4c68-959b-db900528ab2e"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:38:12 crc kubenswrapper[4789]: I1122 08:38:12.942126 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61e96af2-fd7c-4c68-959b-db900528ab2e-ceph" (OuterVolumeSpecName: "ceph") pod "61e96af2-fd7c-4c68-959b-db900528ab2e" (UID: "61e96af2-fd7c-4c68-959b-db900528ab2e"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:38:12 crc kubenswrapper[4789]: I1122 08:38:12.964056 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61e96af2-fd7c-4c68-959b-db900528ab2e-inventory" (OuterVolumeSpecName: "inventory") pod "61e96af2-fd7c-4c68-959b-db900528ab2e" (UID: "61e96af2-fd7c-4c68-959b-db900528ab2e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:38:12 crc kubenswrapper[4789]: I1122 08:38:12.966015 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61e96af2-fd7c-4c68-959b-db900528ab2e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "61e96af2-fd7c-4c68-959b-db900528ab2e" (UID: "61e96af2-fd7c-4c68-959b-db900528ab2e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:38:13 crc kubenswrapper[4789]: I1122 08:38:13.036776 4789 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61e96af2-fd7c-4c68-959b-db900528ab2e-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:38:13 crc kubenswrapper[4789]: I1122 08:38:13.037193 4789 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/61e96af2-fd7c-4c68-959b-db900528ab2e-inventory\") on node \"crc\" DevicePath \"\"" Nov 22 08:38:13 crc kubenswrapper[4789]: I1122 08:38:13.037268 4789 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/61e96af2-fd7c-4c68-959b-db900528ab2e-ceph\") on node \"crc\" DevicePath \"\"" Nov 22 08:38:13 crc kubenswrapper[4789]: I1122 08:38:13.037287 4789 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/61e96af2-fd7c-4c68-959b-db900528ab2e-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 22 08:38:13 crc kubenswrapper[4789]: I1122 08:38:13.037332 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fmzwd\" (UniqueName: \"kubernetes.io/projected/61e96af2-fd7c-4c68-959b-db900528ab2e-kube-api-access-fmzwd\") on node \"crc\" DevicePath \"\"" Nov 22 08:38:13 crc kubenswrapper[4789]: I1122 08:38:13.427232 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bxq52" event={"ID":"61e96af2-fd7c-4c68-959b-db900528ab2e","Type":"ContainerDied","Data":"98ebde52536c286b7432ff1fee112ad3475f7be08439cb832dbe2a361f0f8fab"} Nov 22 08:38:13 crc kubenswrapper[4789]: I1122 08:38:13.427281 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="98ebde52536c286b7432ff1fee112ad3475f7be08439cb832dbe2a361f0f8fab" Nov 22 08:38:13 crc kubenswrapper[4789]: I1122 08:38:13.427301 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bxq52" Nov 22 08:38:13 crc kubenswrapper[4789]: I1122 08:38:13.538621 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-lwqsb"] Nov 22 08:38:13 crc kubenswrapper[4789]: E1122 08:38:13.539056 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61e96af2-fd7c-4c68-959b-db900528ab2e" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 22 08:38:13 crc kubenswrapper[4789]: I1122 08:38:13.539090 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="61e96af2-fd7c-4c68-959b-db900528ab2e" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 22 08:38:13 crc kubenswrapper[4789]: I1122 08:38:13.539310 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="61e96af2-fd7c-4c68-959b-db900528ab2e" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 22 08:38:13 crc kubenswrapper[4789]: I1122 08:38:13.540053 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-lwqsb" Nov 22 08:38:13 crc kubenswrapper[4789]: I1122 08:38:13.542103 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-glmr7" Nov 22 08:38:13 crc kubenswrapper[4789]: I1122 08:38:13.542280 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 22 08:38:13 crc kubenswrapper[4789]: I1122 08:38:13.542490 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 22 08:38:13 crc kubenswrapper[4789]: I1122 08:38:13.542814 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 22 08:38:13 crc kubenswrapper[4789]: I1122 08:38:13.544380 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 22 08:38:13 crc kubenswrapper[4789]: I1122 08:38:13.545388 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/981cbfb1-edd3-449f-865e-b958220cf470-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-lwqsb\" (UID: \"981cbfb1-edd3-449f-865e-b958220cf470\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-lwqsb" Nov 22 08:38:13 crc kubenswrapper[4789]: I1122 08:38:13.545461 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/981cbfb1-edd3-449f-865e-b958220cf470-ceph\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-lwqsb\" (UID: \"981cbfb1-edd3-449f-865e-b958220cf470\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-lwqsb" Nov 22 08:38:13 crc kubenswrapper[4789]: I1122 08:38:13.545499 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-74p4z\" (UniqueName: \"kubernetes.io/projected/981cbfb1-edd3-449f-865e-b958220cf470-kube-api-access-74p4z\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-lwqsb\" (UID: \"981cbfb1-edd3-449f-865e-b958220cf470\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-lwqsb" Nov 22 08:38:13 crc kubenswrapper[4789]: I1122 08:38:13.545586 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/981cbfb1-edd3-449f-865e-b958220cf470-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-lwqsb\" (UID: \"981cbfb1-edd3-449f-865e-b958220cf470\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-lwqsb" Nov 22 08:38:13 crc kubenswrapper[4789]: I1122 08:38:13.548683 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-lwqsb"] Nov 22 08:38:13 crc kubenswrapper[4789]: I1122 08:38:13.647406 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/981cbfb1-edd3-449f-865e-b958220cf470-ceph\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-lwqsb\" (UID: \"981cbfb1-edd3-449f-865e-b958220cf470\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-lwqsb" Nov 22 08:38:13 crc kubenswrapper[4789]: I1122 08:38:13.647469 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-74p4z\" (UniqueName: \"kubernetes.io/projected/981cbfb1-edd3-449f-865e-b958220cf470-kube-api-access-74p4z\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-lwqsb\" (UID: \"981cbfb1-edd3-449f-865e-b958220cf470\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-lwqsb" Nov 22 08:38:13 crc kubenswrapper[4789]: I1122 08:38:13.647496 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/981cbfb1-edd3-449f-865e-b958220cf470-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-lwqsb\" (UID: \"981cbfb1-edd3-449f-865e-b958220cf470\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-lwqsb" Nov 22 08:38:13 crc kubenswrapper[4789]: I1122 08:38:13.647639 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/981cbfb1-edd3-449f-865e-b958220cf470-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-lwqsb\" (UID: \"981cbfb1-edd3-449f-865e-b958220cf470\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-lwqsb" Nov 22 08:38:13 crc kubenswrapper[4789]: I1122 08:38:13.651099 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/981cbfb1-edd3-449f-865e-b958220cf470-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-lwqsb\" (UID: \"981cbfb1-edd3-449f-865e-b958220cf470\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-lwqsb" Nov 22 08:38:13 crc kubenswrapper[4789]: I1122 08:38:13.651484 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/981cbfb1-edd3-449f-865e-b958220cf470-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-lwqsb\" (UID: \"981cbfb1-edd3-449f-865e-b958220cf470\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-lwqsb" Nov 22 08:38:13 crc kubenswrapper[4789]: I1122 08:38:13.651649 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/981cbfb1-edd3-449f-865e-b958220cf470-ceph\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-lwqsb\" (UID: \"981cbfb1-edd3-449f-865e-b958220cf470\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-lwqsb" Nov 22 08:38:13 crc kubenswrapper[4789]: I1122 08:38:13.665398 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-74p4z\" (UniqueName: \"kubernetes.io/projected/981cbfb1-edd3-449f-865e-b958220cf470-kube-api-access-74p4z\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-lwqsb\" (UID: \"981cbfb1-edd3-449f-865e-b958220cf470\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-lwqsb" Nov 22 08:38:13 crc kubenswrapper[4789]: I1122 08:38:13.865407 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-lwqsb" Nov 22 08:38:14 crc kubenswrapper[4789]: I1122 08:38:14.401245 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-lwqsb"] Nov 22 08:38:14 crc kubenswrapper[4789]: I1122 08:38:14.442170 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-lwqsb" event={"ID":"981cbfb1-edd3-449f-865e-b958220cf470","Type":"ContainerStarted","Data":"f2528ca0bb5bc5f0ddea2a5e51e4f6461b5256b87986ec380053a6738f536150"} Nov 22 08:38:15 crc kubenswrapper[4789]: I1122 08:38:15.450065 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-lwqsb" event={"ID":"981cbfb1-edd3-449f-865e-b958220cf470","Type":"ContainerStarted","Data":"cd76604fe8734b21dfd166ac93788cbaa956f42dff513677da0504d04f637f96"} Nov 22 08:38:15 crc kubenswrapper[4789]: I1122 08:38:15.467953 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-lwqsb" podStartSLOduration=1.872658801 podStartE2EDuration="2.467933739s" podCreationTimestamp="2025-11-22 08:38:13 +0000 UTC" firstStartedPulling="2025-11-22 08:38:14.410422501 +0000 UTC m=+2508.644822774" lastFinishedPulling="2025-11-22 08:38:15.005697439 +0000 UTC m=+2509.240097712" observedRunningTime="2025-11-22 08:38:15.466456929 +0000 UTC m=+2509.700857202" watchObservedRunningTime="2025-11-22 08:38:15.467933739 +0000 UTC m=+2509.702334012" Nov 22 08:38:19 crc kubenswrapper[4789]: I1122 08:38:19.965070 4789 scope.go:117] "RemoveContainer" containerID="c7898b5c08854d5df6033d34564729f053b077e4fcadc5a1ea77706f1981be3b" Nov 22 08:38:19 crc kubenswrapper[4789]: E1122 08:38:19.965956 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 08:38:33 crc kubenswrapper[4789]: I1122 08:38:33.968296 4789 scope.go:117] "RemoveContainer" containerID="c7898b5c08854d5df6033d34564729f053b077e4fcadc5a1ea77706f1981be3b" Nov 22 08:38:33 crc kubenswrapper[4789]: E1122 08:38:33.970542 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 08:38:41 crc kubenswrapper[4789]: I1122 08:38:41.682590 4789 generic.go:334] "Generic (PLEG): container finished" podID="981cbfb1-edd3-449f-865e-b958220cf470" containerID="cd76604fe8734b21dfd166ac93788cbaa956f42dff513677da0504d04f637f96" exitCode=0 Nov 22 08:38:41 crc kubenswrapper[4789]: I1122 08:38:41.682729 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-lwqsb" event={"ID":"981cbfb1-edd3-449f-865e-b958220cf470","Type":"ContainerDied","Data":"cd76604fe8734b21dfd166ac93788cbaa956f42dff513677da0504d04f637f96"} Nov 22 08:38:43 crc kubenswrapper[4789]: I1122 08:38:43.111897 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-lwqsb" Nov 22 08:38:43 crc kubenswrapper[4789]: I1122 08:38:43.155938 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-74p4z\" (UniqueName: \"kubernetes.io/projected/981cbfb1-edd3-449f-865e-b958220cf470-kube-api-access-74p4z\") pod \"981cbfb1-edd3-449f-865e-b958220cf470\" (UID: \"981cbfb1-edd3-449f-865e-b958220cf470\") " Nov 22 08:38:43 crc kubenswrapper[4789]: I1122 08:38:43.156025 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/981cbfb1-edd3-449f-865e-b958220cf470-ceph\") pod \"981cbfb1-edd3-449f-865e-b958220cf470\" (UID: \"981cbfb1-edd3-449f-865e-b958220cf470\") " Nov 22 08:38:43 crc kubenswrapper[4789]: I1122 08:38:43.156060 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/981cbfb1-edd3-449f-865e-b958220cf470-inventory\") pod \"981cbfb1-edd3-449f-865e-b958220cf470\" (UID: \"981cbfb1-edd3-449f-865e-b958220cf470\") " Nov 22 08:38:43 crc kubenswrapper[4789]: I1122 08:38:43.156086 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/981cbfb1-edd3-449f-865e-b958220cf470-ssh-key\") pod \"981cbfb1-edd3-449f-865e-b958220cf470\" (UID: \"981cbfb1-edd3-449f-865e-b958220cf470\") " Nov 22 08:38:43 crc kubenswrapper[4789]: I1122 08:38:43.163897 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/981cbfb1-edd3-449f-865e-b958220cf470-ceph" (OuterVolumeSpecName: "ceph") pod "981cbfb1-edd3-449f-865e-b958220cf470" (UID: "981cbfb1-edd3-449f-865e-b958220cf470"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:38:43 crc kubenswrapper[4789]: I1122 08:38:43.164571 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/981cbfb1-edd3-449f-865e-b958220cf470-kube-api-access-74p4z" (OuterVolumeSpecName: "kube-api-access-74p4z") pod "981cbfb1-edd3-449f-865e-b958220cf470" (UID: "981cbfb1-edd3-449f-865e-b958220cf470"). InnerVolumeSpecName "kube-api-access-74p4z". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:38:43 crc kubenswrapper[4789]: I1122 08:38:43.185783 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/981cbfb1-edd3-449f-865e-b958220cf470-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "981cbfb1-edd3-449f-865e-b958220cf470" (UID: "981cbfb1-edd3-449f-865e-b958220cf470"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:38:43 crc kubenswrapper[4789]: I1122 08:38:43.190563 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/981cbfb1-edd3-449f-865e-b958220cf470-inventory" (OuterVolumeSpecName: "inventory") pod "981cbfb1-edd3-449f-865e-b958220cf470" (UID: "981cbfb1-edd3-449f-865e-b958220cf470"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:38:43 crc kubenswrapper[4789]: I1122 08:38:43.259199 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-74p4z\" (UniqueName: \"kubernetes.io/projected/981cbfb1-edd3-449f-865e-b958220cf470-kube-api-access-74p4z\") on node \"crc\" DevicePath \"\"" Nov 22 08:38:43 crc kubenswrapper[4789]: I1122 08:38:43.259265 4789 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/981cbfb1-edd3-449f-865e-b958220cf470-ceph\") on node \"crc\" DevicePath \"\"" Nov 22 08:38:43 crc kubenswrapper[4789]: I1122 08:38:43.259282 4789 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/981cbfb1-edd3-449f-865e-b958220cf470-inventory\") on node \"crc\" DevicePath \"\"" Nov 22 08:38:43 crc kubenswrapper[4789]: I1122 08:38:43.259292 4789 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/981cbfb1-edd3-449f-865e-b958220cf470-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 22 08:38:43 crc kubenswrapper[4789]: I1122 08:38:43.700951 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-lwqsb" event={"ID":"981cbfb1-edd3-449f-865e-b958220cf470","Type":"ContainerDied","Data":"f2528ca0bb5bc5f0ddea2a5e51e4f6461b5256b87986ec380053a6738f536150"} Nov 22 08:38:43 crc kubenswrapper[4789]: I1122 08:38:43.700991 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f2528ca0bb5bc5f0ddea2a5e51e4f6461b5256b87986ec380053a6738f536150" Nov 22 08:38:43 crc kubenswrapper[4789]: I1122 08:38:43.701008 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-lwqsb" Nov 22 08:38:43 crc kubenswrapper[4789]: I1122 08:38:43.817938 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-n4w68"] Nov 22 08:38:43 crc kubenswrapper[4789]: E1122 08:38:43.818328 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="981cbfb1-edd3-449f-865e-b958220cf470" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 22 08:38:43 crc kubenswrapper[4789]: I1122 08:38:43.818346 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="981cbfb1-edd3-449f-865e-b958220cf470" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 22 08:38:43 crc kubenswrapper[4789]: I1122 08:38:43.818540 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="981cbfb1-edd3-449f-865e-b958220cf470" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 22 08:38:43 crc kubenswrapper[4789]: I1122 08:38:43.819148 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-n4w68" Nov 22 08:38:43 crc kubenswrapper[4789]: I1122 08:38:43.822240 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 22 08:38:43 crc kubenswrapper[4789]: I1122 08:38:43.822801 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 22 08:38:43 crc kubenswrapper[4789]: I1122 08:38:43.823182 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 22 08:38:43 crc kubenswrapper[4789]: I1122 08:38:43.823431 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 22 08:38:43 crc kubenswrapper[4789]: I1122 08:38:43.837891 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-glmr7" Nov 22 08:38:43 crc kubenswrapper[4789]: I1122 08:38:43.846142 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-n4w68"] Nov 22 08:38:43 crc kubenswrapper[4789]: I1122 08:38:43.877859 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b8b49b82-f67f-4416-86f1-474272d6fdee-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-n4w68\" (UID: \"b8b49b82-f67f-4416-86f1-474272d6fdee\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-n4w68" Nov 22 08:38:43 crc kubenswrapper[4789]: I1122 08:38:43.877948 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vb7v4\" (UniqueName: \"kubernetes.io/projected/b8b49b82-f67f-4416-86f1-474272d6fdee-kube-api-access-vb7v4\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-n4w68\" (UID: \"b8b49b82-f67f-4416-86f1-474272d6fdee\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-n4w68" Nov 22 08:38:43 crc kubenswrapper[4789]: I1122 08:38:43.878047 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b8b49b82-f67f-4416-86f1-474272d6fdee-ceph\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-n4w68\" (UID: \"b8b49b82-f67f-4416-86f1-474272d6fdee\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-n4w68" Nov 22 08:38:43 crc kubenswrapper[4789]: I1122 08:38:43.878143 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b8b49b82-f67f-4416-86f1-474272d6fdee-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-n4w68\" (UID: \"b8b49b82-f67f-4416-86f1-474272d6fdee\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-n4w68" Nov 22 08:38:43 crc kubenswrapper[4789]: I1122 08:38:43.979475 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b8b49b82-f67f-4416-86f1-474272d6fdee-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-n4w68\" (UID: \"b8b49b82-f67f-4416-86f1-474272d6fdee\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-n4w68" Nov 22 08:38:43 crc kubenswrapper[4789]: I1122 08:38:43.980167 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vb7v4\" (UniqueName: \"kubernetes.io/projected/b8b49b82-f67f-4416-86f1-474272d6fdee-kube-api-access-vb7v4\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-n4w68\" (UID: \"b8b49b82-f67f-4416-86f1-474272d6fdee\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-n4w68" Nov 22 08:38:43 crc kubenswrapper[4789]: I1122 08:38:43.980415 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b8b49b82-f67f-4416-86f1-474272d6fdee-ceph\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-n4w68\" (UID: \"b8b49b82-f67f-4416-86f1-474272d6fdee\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-n4w68" Nov 22 08:38:43 crc kubenswrapper[4789]: I1122 08:38:43.980664 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b8b49b82-f67f-4416-86f1-474272d6fdee-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-n4w68\" (UID: \"b8b49b82-f67f-4416-86f1-474272d6fdee\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-n4w68" Nov 22 08:38:43 crc kubenswrapper[4789]: I1122 08:38:43.983668 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b8b49b82-f67f-4416-86f1-474272d6fdee-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-n4w68\" (UID: \"b8b49b82-f67f-4416-86f1-474272d6fdee\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-n4w68" Nov 22 08:38:43 crc kubenswrapper[4789]: I1122 08:38:43.984310 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b8b49b82-f67f-4416-86f1-474272d6fdee-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-n4w68\" (UID: \"b8b49b82-f67f-4416-86f1-474272d6fdee\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-n4w68" Nov 22 08:38:43 crc kubenswrapper[4789]: I1122 08:38:43.985229 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b8b49b82-f67f-4416-86f1-474272d6fdee-ceph\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-n4w68\" (UID: \"b8b49b82-f67f-4416-86f1-474272d6fdee\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-n4w68" Nov 22 08:38:43 crc kubenswrapper[4789]: I1122 08:38:43.998787 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vb7v4\" (UniqueName: \"kubernetes.io/projected/b8b49b82-f67f-4416-86f1-474272d6fdee-kube-api-access-vb7v4\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-n4w68\" (UID: \"b8b49b82-f67f-4416-86f1-474272d6fdee\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-n4w68" Nov 22 08:38:44 crc kubenswrapper[4789]: I1122 08:38:44.137030 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-n4w68" Nov 22 08:38:44 crc kubenswrapper[4789]: I1122 08:38:44.681330 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-n4w68"] Nov 22 08:38:44 crc kubenswrapper[4789]: I1122 08:38:44.714241 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-n4w68" event={"ID":"b8b49b82-f67f-4416-86f1-474272d6fdee","Type":"ContainerStarted","Data":"6a86ae79e25ced6d1e6555819126e1ca1c059be767c9e0bad797b6b8fdb4d407"} Nov 22 08:38:45 crc kubenswrapper[4789]: I1122 08:38:45.964946 4789 scope.go:117] "RemoveContainer" containerID="c7898b5c08854d5df6033d34564729f053b077e4fcadc5a1ea77706f1981be3b" Nov 22 08:38:45 crc kubenswrapper[4789]: E1122 08:38:45.966126 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 08:38:46 crc kubenswrapper[4789]: I1122 08:38:46.730518 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-n4w68" event={"ID":"b8b49b82-f67f-4416-86f1-474272d6fdee","Type":"ContainerStarted","Data":"ac9f9a9d871d9f6ae3b876b21822475a5848ce3af92f9084b37150742d3901da"} Nov 22 08:38:46 crc kubenswrapper[4789]: I1122 08:38:46.760665 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-n4w68" podStartSLOduration=2.719732205 podStartE2EDuration="3.760642693s" podCreationTimestamp="2025-11-22 08:38:43 +0000 UTC" firstStartedPulling="2025-11-22 08:38:44.688611692 +0000 UTC m=+2538.923011965" lastFinishedPulling="2025-11-22 08:38:45.72952218 +0000 UTC m=+2539.963922453" observedRunningTime="2025-11-22 08:38:46.757240161 +0000 UTC m=+2540.991640434" watchObservedRunningTime="2025-11-22 08:38:46.760642693 +0000 UTC m=+2540.995042966" Nov 22 08:38:51 crc kubenswrapper[4789]: I1122 08:38:51.772247 4789 generic.go:334] "Generic (PLEG): container finished" podID="b8b49b82-f67f-4416-86f1-474272d6fdee" containerID="ac9f9a9d871d9f6ae3b876b21822475a5848ce3af92f9084b37150742d3901da" exitCode=0 Nov 22 08:38:51 crc kubenswrapper[4789]: I1122 08:38:51.772362 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-n4w68" event={"ID":"b8b49b82-f67f-4416-86f1-474272d6fdee","Type":"ContainerDied","Data":"ac9f9a9d871d9f6ae3b876b21822475a5848ce3af92f9084b37150742d3901da"} Nov 22 08:38:53 crc kubenswrapper[4789]: I1122 08:38:53.190538 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-n4w68" Nov 22 08:38:53 crc kubenswrapper[4789]: I1122 08:38:53.265363 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vb7v4\" (UniqueName: \"kubernetes.io/projected/b8b49b82-f67f-4416-86f1-474272d6fdee-kube-api-access-vb7v4\") pod \"b8b49b82-f67f-4416-86f1-474272d6fdee\" (UID: \"b8b49b82-f67f-4416-86f1-474272d6fdee\") " Nov 22 08:38:53 crc kubenswrapper[4789]: I1122 08:38:53.265491 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b8b49b82-f67f-4416-86f1-474272d6fdee-ceph\") pod \"b8b49b82-f67f-4416-86f1-474272d6fdee\" (UID: \"b8b49b82-f67f-4416-86f1-474272d6fdee\") " Nov 22 08:38:53 crc kubenswrapper[4789]: I1122 08:38:53.265547 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b8b49b82-f67f-4416-86f1-474272d6fdee-ssh-key\") pod \"b8b49b82-f67f-4416-86f1-474272d6fdee\" (UID: \"b8b49b82-f67f-4416-86f1-474272d6fdee\") " Nov 22 08:38:53 crc kubenswrapper[4789]: I1122 08:38:53.265578 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b8b49b82-f67f-4416-86f1-474272d6fdee-inventory\") pod \"b8b49b82-f67f-4416-86f1-474272d6fdee\" (UID: \"b8b49b82-f67f-4416-86f1-474272d6fdee\") " Nov 22 08:38:53 crc kubenswrapper[4789]: I1122 08:38:53.271471 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b8b49b82-f67f-4416-86f1-474272d6fdee-kube-api-access-vb7v4" (OuterVolumeSpecName: "kube-api-access-vb7v4") pod "b8b49b82-f67f-4416-86f1-474272d6fdee" (UID: "b8b49b82-f67f-4416-86f1-474272d6fdee"). InnerVolumeSpecName "kube-api-access-vb7v4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:38:53 crc kubenswrapper[4789]: I1122 08:38:53.272125 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8b49b82-f67f-4416-86f1-474272d6fdee-ceph" (OuterVolumeSpecName: "ceph") pod "b8b49b82-f67f-4416-86f1-474272d6fdee" (UID: "b8b49b82-f67f-4416-86f1-474272d6fdee"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:38:53 crc kubenswrapper[4789]: E1122 08:38:53.288163 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b8b49b82-f67f-4416-86f1-474272d6fdee-ssh-key podName:b8b49b82-f67f-4416-86f1-474272d6fdee nodeName:}" failed. No retries permitted until 2025-11-22 08:38:53.788126338 +0000 UTC m=+2548.022526611 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "ssh-key" (UniqueName: "kubernetes.io/secret/b8b49b82-f67f-4416-86f1-474272d6fdee-ssh-key") pod "b8b49b82-f67f-4416-86f1-474272d6fdee" (UID: "b8b49b82-f67f-4416-86f1-474272d6fdee") : error deleting /var/lib/kubelet/pods/b8b49b82-f67f-4416-86f1-474272d6fdee/volume-subpaths: remove /var/lib/kubelet/pods/b8b49b82-f67f-4416-86f1-474272d6fdee/volume-subpaths: no such file or directory Nov 22 08:38:53 crc kubenswrapper[4789]: I1122 08:38:53.290986 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8b49b82-f67f-4416-86f1-474272d6fdee-inventory" (OuterVolumeSpecName: "inventory") pod "b8b49b82-f67f-4416-86f1-474272d6fdee" (UID: "b8b49b82-f67f-4416-86f1-474272d6fdee"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:38:53 crc kubenswrapper[4789]: I1122 08:38:53.367936 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vb7v4\" (UniqueName: \"kubernetes.io/projected/b8b49b82-f67f-4416-86f1-474272d6fdee-kube-api-access-vb7v4\") on node \"crc\" DevicePath \"\"" Nov 22 08:38:53 crc kubenswrapper[4789]: I1122 08:38:53.367972 4789 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b8b49b82-f67f-4416-86f1-474272d6fdee-ceph\") on node \"crc\" DevicePath \"\"" Nov 22 08:38:53 crc kubenswrapper[4789]: I1122 08:38:53.367981 4789 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b8b49b82-f67f-4416-86f1-474272d6fdee-inventory\") on node \"crc\" DevicePath \"\"" Nov 22 08:38:53 crc kubenswrapper[4789]: I1122 08:38:53.793805 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-n4w68" event={"ID":"b8b49b82-f67f-4416-86f1-474272d6fdee","Type":"ContainerDied","Data":"6a86ae79e25ced6d1e6555819126e1ca1c059be767c9e0bad797b6b8fdb4d407"} Nov 22 08:38:53 crc kubenswrapper[4789]: I1122 08:38:53.793845 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-n4w68" Nov 22 08:38:53 crc kubenswrapper[4789]: I1122 08:38:53.793861 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6a86ae79e25ced6d1e6555819126e1ca1c059be767c9e0bad797b6b8fdb4d407" Nov 22 08:38:53 crc kubenswrapper[4789]: I1122 08:38:53.865955 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-rrdbw"] Nov 22 08:38:53 crc kubenswrapper[4789]: E1122 08:38:53.866356 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8b49b82-f67f-4416-86f1-474272d6fdee" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 22 08:38:53 crc kubenswrapper[4789]: I1122 08:38:53.866375 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8b49b82-f67f-4416-86f1-474272d6fdee" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 22 08:38:53 crc kubenswrapper[4789]: I1122 08:38:53.866579 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8b49b82-f67f-4416-86f1-474272d6fdee" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 22 08:38:53 crc kubenswrapper[4789]: I1122 08:38:53.867259 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-rrdbw" Nov 22 08:38:53 crc kubenswrapper[4789]: I1122 08:38:53.875298 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b8b49b82-f67f-4416-86f1-474272d6fdee-ssh-key\") pod \"b8b49b82-f67f-4416-86f1-474272d6fdee\" (UID: \"b8b49b82-f67f-4416-86f1-474272d6fdee\") " Nov 22 08:38:53 crc kubenswrapper[4789]: I1122 08:38:53.876512 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-rrdbw"] Nov 22 08:38:53 crc kubenswrapper[4789]: I1122 08:38:53.879919 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8b49b82-f67f-4416-86f1-474272d6fdee-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b8b49b82-f67f-4416-86f1-474272d6fdee" (UID: "b8b49b82-f67f-4416-86f1-474272d6fdee"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:38:53 crc kubenswrapper[4789]: I1122 08:38:53.977269 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3798e707-7719-49a1-b644-ce53243a5131-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-rrdbw\" (UID: \"3798e707-7719-49a1-b644-ce53243a5131\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-rrdbw" Nov 22 08:38:53 crc kubenswrapper[4789]: I1122 08:38:53.977337 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3798e707-7719-49a1-b644-ce53243a5131-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-rrdbw\" (UID: \"3798e707-7719-49a1-b644-ce53243a5131\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-rrdbw" Nov 22 08:38:53 crc kubenswrapper[4789]: I1122 08:38:53.977365 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zqq59\" (UniqueName: \"kubernetes.io/projected/3798e707-7719-49a1-b644-ce53243a5131-kube-api-access-zqq59\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-rrdbw\" (UID: \"3798e707-7719-49a1-b644-ce53243a5131\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-rrdbw" Nov 22 08:38:53 crc kubenswrapper[4789]: I1122 08:38:53.978008 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3798e707-7719-49a1-b644-ce53243a5131-ceph\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-rrdbw\" (UID: \"3798e707-7719-49a1-b644-ce53243a5131\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-rrdbw" Nov 22 08:38:53 crc kubenswrapper[4789]: I1122 08:38:53.978177 4789 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b8b49b82-f67f-4416-86f1-474272d6fdee-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 22 08:38:54 crc kubenswrapper[4789]: I1122 08:38:54.083003 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3798e707-7719-49a1-b644-ce53243a5131-ceph\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-rrdbw\" (UID: \"3798e707-7719-49a1-b644-ce53243a5131\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-rrdbw" Nov 22 08:38:54 crc kubenswrapper[4789]: I1122 08:38:54.084084 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3798e707-7719-49a1-b644-ce53243a5131-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-rrdbw\" (UID: \"3798e707-7719-49a1-b644-ce53243a5131\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-rrdbw" Nov 22 08:38:54 crc kubenswrapper[4789]: I1122 08:38:54.084173 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3798e707-7719-49a1-b644-ce53243a5131-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-rrdbw\" (UID: \"3798e707-7719-49a1-b644-ce53243a5131\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-rrdbw" Nov 22 08:38:54 crc kubenswrapper[4789]: I1122 08:38:54.084200 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zqq59\" (UniqueName: \"kubernetes.io/projected/3798e707-7719-49a1-b644-ce53243a5131-kube-api-access-zqq59\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-rrdbw\" (UID: \"3798e707-7719-49a1-b644-ce53243a5131\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-rrdbw" Nov 22 08:38:54 crc kubenswrapper[4789]: I1122 08:38:54.088166 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3798e707-7719-49a1-b644-ce53243a5131-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-rrdbw\" (UID: \"3798e707-7719-49a1-b644-ce53243a5131\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-rrdbw" Nov 22 08:38:54 crc kubenswrapper[4789]: I1122 08:38:54.088227 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3798e707-7719-49a1-b644-ce53243a5131-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-rrdbw\" (UID: \"3798e707-7719-49a1-b644-ce53243a5131\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-rrdbw" Nov 22 08:38:54 crc kubenswrapper[4789]: I1122 08:38:54.088182 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3798e707-7719-49a1-b644-ce53243a5131-ceph\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-rrdbw\" (UID: \"3798e707-7719-49a1-b644-ce53243a5131\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-rrdbw" Nov 22 08:38:54 crc kubenswrapper[4789]: I1122 08:38:54.099505 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zqq59\" (UniqueName: \"kubernetes.io/projected/3798e707-7719-49a1-b644-ce53243a5131-kube-api-access-zqq59\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-rrdbw\" (UID: \"3798e707-7719-49a1-b644-ce53243a5131\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-rrdbw" Nov 22 08:38:54 crc kubenswrapper[4789]: I1122 08:38:54.231401 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-rrdbw" Nov 22 08:38:54 crc kubenswrapper[4789]: I1122 08:38:54.761051 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-rrdbw"] Nov 22 08:38:54 crc kubenswrapper[4789]: W1122 08:38:54.763634 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3798e707_7719_49a1_b644_ce53243a5131.slice/crio-f364bbfd3d4e849b2a592aea71b98bf3ed8d854dc925c5af6a91696c74c9c4c2 WatchSource:0}: Error finding container f364bbfd3d4e849b2a592aea71b98bf3ed8d854dc925c5af6a91696c74c9c4c2: Status 404 returned error can't find the container with id f364bbfd3d4e849b2a592aea71b98bf3ed8d854dc925c5af6a91696c74c9c4c2 Nov 22 08:38:54 crc kubenswrapper[4789]: I1122 08:38:54.804141 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-rrdbw" event={"ID":"3798e707-7719-49a1-b644-ce53243a5131","Type":"ContainerStarted","Data":"f364bbfd3d4e849b2a592aea71b98bf3ed8d854dc925c5af6a91696c74c9c4c2"} Nov 22 08:38:55 crc kubenswrapper[4789]: I1122 08:38:55.814578 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-rrdbw" event={"ID":"3798e707-7719-49a1-b644-ce53243a5131","Type":"ContainerStarted","Data":"170d11184d3ea6ff516ff74cb83d7a2196644d1396562b75b5e942488cb15365"} Nov 22 08:38:55 crc kubenswrapper[4789]: I1122 08:38:55.835374 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-rrdbw" podStartSLOduration=2.416601089 podStartE2EDuration="2.835351691s" podCreationTimestamp="2025-11-22 08:38:53 +0000 UTC" firstStartedPulling="2025-11-22 08:38:54.76592405 +0000 UTC m=+2549.000324323" lastFinishedPulling="2025-11-22 08:38:55.184674652 +0000 UTC m=+2549.419074925" observedRunningTime="2025-11-22 08:38:55.832923725 +0000 UTC m=+2550.067323998" watchObservedRunningTime="2025-11-22 08:38:55.835351691 +0000 UTC m=+2550.069751964" Nov 22 08:38:58 crc kubenswrapper[4789]: I1122 08:38:58.965900 4789 scope.go:117] "RemoveContainer" containerID="c7898b5c08854d5df6033d34564729f053b077e4fcadc5a1ea77706f1981be3b" Nov 22 08:38:58 crc kubenswrapper[4789]: E1122 08:38:58.966776 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 08:39:09 crc kubenswrapper[4789]: I1122 08:39:09.965282 4789 scope.go:117] "RemoveContainer" containerID="c7898b5c08854d5df6033d34564729f053b077e4fcadc5a1ea77706f1981be3b" Nov 22 08:39:11 crc kubenswrapper[4789]: I1122 08:39:11.141122 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" event={"ID":"58aa8071-7ecc-4692-8789-2db04ec70647","Type":"ContainerStarted","Data":"95ee0f2b0a62b96c30cc7400004c7e7b084559a4811441f88de89dadf3d054d1"} Nov 22 08:39:32 crc kubenswrapper[4789]: I1122 08:39:32.313304 4789 generic.go:334] "Generic (PLEG): container finished" podID="3798e707-7719-49a1-b644-ce53243a5131" containerID="170d11184d3ea6ff516ff74cb83d7a2196644d1396562b75b5e942488cb15365" exitCode=0 Nov 22 08:39:32 crc kubenswrapper[4789]: I1122 08:39:32.313441 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-rrdbw" event={"ID":"3798e707-7719-49a1-b644-ce53243a5131","Type":"ContainerDied","Data":"170d11184d3ea6ff516ff74cb83d7a2196644d1396562b75b5e942488cb15365"} Nov 22 08:39:33 crc kubenswrapper[4789]: I1122 08:39:33.703678 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-rrdbw" Nov 22 08:39:33 crc kubenswrapper[4789]: I1122 08:39:33.825453 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3798e707-7719-49a1-b644-ce53243a5131-ssh-key\") pod \"3798e707-7719-49a1-b644-ce53243a5131\" (UID: \"3798e707-7719-49a1-b644-ce53243a5131\") " Nov 22 08:39:33 crc kubenswrapper[4789]: I1122 08:39:33.825514 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3798e707-7719-49a1-b644-ce53243a5131-inventory\") pod \"3798e707-7719-49a1-b644-ce53243a5131\" (UID: \"3798e707-7719-49a1-b644-ce53243a5131\") " Nov 22 08:39:33 crc kubenswrapper[4789]: I1122 08:39:33.825578 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3798e707-7719-49a1-b644-ce53243a5131-ceph\") pod \"3798e707-7719-49a1-b644-ce53243a5131\" (UID: \"3798e707-7719-49a1-b644-ce53243a5131\") " Nov 22 08:39:33 crc kubenswrapper[4789]: I1122 08:39:33.825651 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zqq59\" (UniqueName: \"kubernetes.io/projected/3798e707-7719-49a1-b644-ce53243a5131-kube-api-access-zqq59\") pod \"3798e707-7719-49a1-b644-ce53243a5131\" (UID: \"3798e707-7719-49a1-b644-ce53243a5131\") " Nov 22 08:39:33 crc kubenswrapper[4789]: I1122 08:39:33.832387 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3798e707-7719-49a1-b644-ce53243a5131-kube-api-access-zqq59" (OuterVolumeSpecName: "kube-api-access-zqq59") pod "3798e707-7719-49a1-b644-ce53243a5131" (UID: "3798e707-7719-49a1-b644-ce53243a5131"). InnerVolumeSpecName "kube-api-access-zqq59". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:39:33 crc kubenswrapper[4789]: I1122 08:39:33.832772 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3798e707-7719-49a1-b644-ce53243a5131-ceph" (OuterVolumeSpecName: "ceph") pod "3798e707-7719-49a1-b644-ce53243a5131" (UID: "3798e707-7719-49a1-b644-ce53243a5131"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:39:33 crc kubenswrapper[4789]: I1122 08:39:33.852670 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3798e707-7719-49a1-b644-ce53243a5131-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "3798e707-7719-49a1-b644-ce53243a5131" (UID: "3798e707-7719-49a1-b644-ce53243a5131"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:39:33 crc kubenswrapper[4789]: I1122 08:39:33.853003 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3798e707-7719-49a1-b644-ce53243a5131-inventory" (OuterVolumeSpecName: "inventory") pod "3798e707-7719-49a1-b644-ce53243a5131" (UID: "3798e707-7719-49a1-b644-ce53243a5131"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:39:33 crc kubenswrapper[4789]: I1122 08:39:33.928036 4789 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3798e707-7719-49a1-b644-ce53243a5131-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 22 08:39:33 crc kubenswrapper[4789]: I1122 08:39:33.928101 4789 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3798e707-7719-49a1-b644-ce53243a5131-inventory\") on node \"crc\" DevicePath \"\"" Nov 22 08:39:33 crc kubenswrapper[4789]: I1122 08:39:33.928112 4789 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3798e707-7719-49a1-b644-ce53243a5131-ceph\") on node \"crc\" DevicePath \"\"" Nov 22 08:39:33 crc kubenswrapper[4789]: I1122 08:39:33.928122 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zqq59\" (UniqueName: \"kubernetes.io/projected/3798e707-7719-49a1-b644-ce53243a5131-kube-api-access-zqq59\") on node \"crc\" DevicePath \"\"" Nov 22 08:39:34 crc kubenswrapper[4789]: I1122 08:39:34.331108 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-rrdbw" event={"ID":"3798e707-7719-49a1-b644-ce53243a5131","Type":"ContainerDied","Data":"f364bbfd3d4e849b2a592aea71b98bf3ed8d854dc925c5af6a91696c74c9c4c2"} Nov 22 08:39:34 crc kubenswrapper[4789]: I1122 08:39:34.331153 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f364bbfd3d4e849b2a592aea71b98bf3ed8d854dc925c5af6a91696c74c9c4c2" Nov 22 08:39:34 crc kubenswrapper[4789]: I1122 08:39:34.331159 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-rrdbw" Nov 22 08:39:34 crc kubenswrapper[4789]: I1122 08:39:34.420083 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-mdh5v"] Nov 22 08:39:34 crc kubenswrapper[4789]: E1122 08:39:34.420487 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3798e707-7719-49a1-b644-ce53243a5131" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 22 08:39:34 crc kubenswrapper[4789]: I1122 08:39:34.420507 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="3798e707-7719-49a1-b644-ce53243a5131" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 22 08:39:34 crc kubenswrapper[4789]: I1122 08:39:34.420716 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="3798e707-7719-49a1-b644-ce53243a5131" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 22 08:39:34 crc kubenswrapper[4789]: I1122 08:39:34.421338 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-mdh5v" Nov 22 08:39:34 crc kubenswrapper[4789]: I1122 08:39:34.423374 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 22 08:39:34 crc kubenswrapper[4789]: I1122 08:39:34.423569 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 22 08:39:34 crc kubenswrapper[4789]: I1122 08:39:34.423847 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-glmr7" Nov 22 08:39:34 crc kubenswrapper[4789]: I1122 08:39:34.424129 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 22 08:39:34 crc kubenswrapper[4789]: I1122 08:39:34.424299 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 22 08:39:34 crc kubenswrapper[4789]: I1122 08:39:34.433230 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-mdh5v"] Nov 22 08:39:34 crc kubenswrapper[4789]: I1122 08:39:34.537169 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-slpnf\" (UniqueName: \"kubernetes.io/projected/a05e1e58-64e0-4b6c-bab9-b4ad7a8eea2b-kube-api-access-slpnf\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-mdh5v\" (UID: \"a05e1e58-64e0-4b6c-bab9-b4ad7a8eea2b\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-mdh5v" Nov 22 08:39:34 crc kubenswrapper[4789]: I1122 08:39:34.537242 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/a05e1e58-64e0-4b6c-bab9-b4ad7a8eea2b-ceph\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-mdh5v\" (UID: \"a05e1e58-64e0-4b6c-bab9-b4ad7a8eea2b\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-mdh5v" Nov 22 08:39:34 crc kubenswrapper[4789]: I1122 08:39:34.538275 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a05e1e58-64e0-4b6c-bab9-b4ad7a8eea2b-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-mdh5v\" (UID: \"a05e1e58-64e0-4b6c-bab9-b4ad7a8eea2b\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-mdh5v" Nov 22 08:39:34 crc kubenswrapper[4789]: I1122 08:39:34.538591 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a05e1e58-64e0-4b6c-bab9-b4ad7a8eea2b-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-mdh5v\" (UID: \"a05e1e58-64e0-4b6c-bab9-b4ad7a8eea2b\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-mdh5v" Nov 22 08:39:34 crc kubenswrapper[4789]: I1122 08:39:34.639965 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-slpnf\" (UniqueName: \"kubernetes.io/projected/a05e1e58-64e0-4b6c-bab9-b4ad7a8eea2b-kube-api-access-slpnf\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-mdh5v\" (UID: \"a05e1e58-64e0-4b6c-bab9-b4ad7a8eea2b\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-mdh5v" Nov 22 08:39:34 crc kubenswrapper[4789]: I1122 08:39:34.640022 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/a05e1e58-64e0-4b6c-bab9-b4ad7a8eea2b-ceph\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-mdh5v\" (UID: \"a05e1e58-64e0-4b6c-bab9-b4ad7a8eea2b\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-mdh5v" Nov 22 08:39:34 crc kubenswrapper[4789]: I1122 08:39:34.640068 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a05e1e58-64e0-4b6c-bab9-b4ad7a8eea2b-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-mdh5v\" (UID: \"a05e1e58-64e0-4b6c-bab9-b4ad7a8eea2b\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-mdh5v" Nov 22 08:39:34 crc kubenswrapper[4789]: I1122 08:39:34.640122 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a05e1e58-64e0-4b6c-bab9-b4ad7a8eea2b-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-mdh5v\" (UID: \"a05e1e58-64e0-4b6c-bab9-b4ad7a8eea2b\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-mdh5v" Nov 22 08:39:34 crc kubenswrapper[4789]: I1122 08:39:34.644659 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a05e1e58-64e0-4b6c-bab9-b4ad7a8eea2b-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-mdh5v\" (UID: \"a05e1e58-64e0-4b6c-bab9-b4ad7a8eea2b\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-mdh5v" Nov 22 08:39:34 crc kubenswrapper[4789]: I1122 08:39:34.644672 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a05e1e58-64e0-4b6c-bab9-b4ad7a8eea2b-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-mdh5v\" (UID: \"a05e1e58-64e0-4b6c-bab9-b4ad7a8eea2b\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-mdh5v" Nov 22 08:39:34 crc kubenswrapper[4789]: I1122 08:39:34.644840 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/a05e1e58-64e0-4b6c-bab9-b4ad7a8eea2b-ceph\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-mdh5v\" (UID: \"a05e1e58-64e0-4b6c-bab9-b4ad7a8eea2b\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-mdh5v" Nov 22 08:39:34 crc kubenswrapper[4789]: I1122 08:39:34.656811 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-slpnf\" (UniqueName: \"kubernetes.io/projected/a05e1e58-64e0-4b6c-bab9-b4ad7a8eea2b-kube-api-access-slpnf\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-mdh5v\" (UID: \"a05e1e58-64e0-4b6c-bab9-b4ad7a8eea2b\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-mdh5v" Nov 22 08:39:34 crc kubenswrapper[4789]: I1122 08:39:34.738634 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-mdh5v" Nov 22 08:39:35 crc kubenswrapper[4789]: I1122 08:39:35.224787 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-mdh5v"] Nov 22 08:39:35 crc kubenswrapper[4789]: I1122 08:39:35.341187 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-mdh5v" event={"ID":"a05e1e58-64e0-4b6c-bab9-b4ad7a8eea2b","Type":"ContainerStarted","Data":"6a9336fe598652078e117e018b1140bfb3b51cdf8bfd64851418a321fb28b44e"} Nov 22 08:39:36 crc kubenswrapper[4789]: I1122 08:39:36.349710 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-mdh5v" event={"ID":"a05e1e58-64e0-4b6c-bab9-b4ad7a8eea2b","Type":"ContainerStarted","Data":"0861b6ce5ae3ac8839c56d187fd682ca1f51ad2728fa4c9c3f3f187dc4a94efe"} Nov 22 08:39:40 crc kubenswrapper[4789]: I1122 08:39:40.387926 4789 generic.go:334] "Generic (PLEG): container finished" podID="a05e1e58-64e0-4b6c-bab9-b4ad7a8eea2b" containerID="0861b6ce5ae3ac8839c56d187fd682ca1f51ad2728fa4c9c3f3f187dc4a94efe" exitCode=0 Nov 22 08:39:40 crc kubenswrapper[4789]: I1122 08:39:40.388010 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-mdh5v" event={"ID":"a05e1e58-64e0-4b6c-bab9-b4ad7a8eea2b","Type":"ContainerDied","Data":"0861b6ce5ae3ac8839c56d187fd682ca1f51ad2728fa4c9c3f3f187dc4a94efe"} Nov 22 08:39:41 crc kubenswrapper[4789]: I1122 08:39:41.788890 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-mdh5v" Nov 22 08:39:41 crc kubenswrapper[4789]: I1122 08:39:41.875656 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/a05e1e58-64e0-4b6c-bab9-b4ad7a8eea2b-ceph\") pod \"a05e1e58-64e0-4b6c-bab9-b4ad7a8eea2b\" (UID: \"a05e1e58-64e0-4b6c-bab9-b4ad7a8eea2b\") " Nov 22 08:39:41 crc kubenswrapper[4789]: I1122 08:39:41.875808 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a05e1e58-64e0-4b6c-bab9-b4ad7a8eea2b-inventory\") pod \"a05e1e58-64e0-4b6c-bab9-b4ad7a8eea2b\" (UID: \"a05e1e58-64e0-4b6c-bab9-b4ad7a8eea2b\") " Nov 22 08:39:41 crc kubenswrapper[4789]: I1122 08:39:41.875830 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a05e1e58-64e0-4b6c-bab9-b4ad7a8eea2b-ssh-key\") pod \"a05e1e58-64e0-4b6c-bab9-b4ad7a8eea2b\" (UID: \"a05e1e58-64e0-4b6c-bab9-b4ad7a8eea2b\") " Nov 22 08:39:41 crc kubenswrapper[4789]: I1122 08:39:41.875856 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-slpnf\" (UniqueName: \"kubernetes.io/projected/a05e1e58-64e0-4b6c-bab9-b4ad7a8eea2b-kube-api-access-slpnf\") pod \"a05e1e58-64e0-4b6c-bab9-b4ad7a8eea2b\" (UID: \"a05e1e58-64e0-4b6c-bab9-b4ad7a8eea2b\") " Nov 22 08:39:41 crc kubenswrapper[4789]: I1122 08:39:41.881934 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a05e1e58-64e0-4b6c-bab9-b4ad7a8eea2b-kube-api-access-slpnf" (OuterVolumeSpecName: "kube-api-access-slpnf") pod "a05e1e58-64e0-4b6c-bab9-b4ad7a8eea2b" (UID: "a05e1e58-64e0-4b6c-bab9-b4ad7a8eea2b"). InnerVolumeSpecName "kube-api-access-slpnf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:39:41 crc kubenswrapper[4789]: I1122 08:39:41.882062 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a05e1e58-64e0-4b6c-bab9-b4ad7a8eea2b-ceph" (OuterVolumeSpecName: "ceph") pod "a05e1e58-64e0-4b6c-bab9-b4ad7a8eea2b" (UID: "a05e1e58-64e0-4b6c-bab9-b4ad7a8eea2b"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:39:41 crc kubenswrapper[4789]: I1122 08:39:41.900539 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a05e1e58-64e0-4b6c-bab9-b4ad7a8eea2b-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "a05e1e58-64e0-4b6c-bab9-b4ad7a8eea2b" (UID: "a05e1e58-64e0-4b6c-bab9-b4ad7a8eea2b"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:39:41 crc kubenswrapper[4789]: I1122 08:39:41.901076 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a05e1e58-64e0-4b6c-bab9-b4ad7a8eea2b-inventory" (OuterVolumeSpecName: "inventory") pod "a05e1e58-64e0-4b6c-bab9-b4ad7a8eea2b" (UID: "a05e1e58-64e0-4b6c-bab9-b4ad7a8eea2b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:39:41 crc kubenswrapper[4789]: I1122 08:39:41.978194 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-slpnf\" (UniqueName: \"kubernetes.io/projected/a05e1e58-64e0-4b6c-bab9-b4ad7a8eea2b-kube-api-access-slpnf\") on node \"crc\" DevicePath \"\"" Nov 22 08:39:41 crc kubenswrapper[4789]: I1122 08:39:41.978229 4789 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/a05e1e58-64e0-4b6c-bab9-b4ad7a8eea2b-ceph\") on node \"crc\" DevicePath \"\"" Nov 22 08:39:41 crc kubenswrapper[4789]: I1122 08:39:41.978242 4789 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a05e1e58-64e0-4b6c-bab9-b4ad7a8eea2b-inventory\") on node \"crc\" DevicePath \"\"" Nov 22 08:39:41 crc kubenswrapper[4789]: I1122 08:39:41.978253 4789 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a05e1e58-64e0-4b6c-bab9-b4ad7a8eea2b-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 22 08:39:42 crc kubenswrapper[4789]: I1122 08:39:42.413161 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-mdh5v" event={"ID":"a05e1e58-64e0-4b6c-bab9-b4ad7a8eea2b","Type":"ContainerDied","Data":"6a9336fe598652078e117e018b1140bfb3b51cdf8bfd64851418a321fb28b44e"} Nov 22 08:39:42 crc kubenswrapper[4789]: I1122 08:39:42.413286 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6a9336fe598652078e117e018b1140bfb3b51cdf8bfd64851418a321fb28b44e" Nov 22 08:39:42 crc kubenswrapper[4789]: I1122 08:39:42.413248 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-mdh5v" Nov 22 08:39:42 crc kubenswrapper[4789]: I1122 08:39:42.479881 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vt7vg"] Nov 22 08:39:42 crc kubenswrapper[4789]: E1122 08:39:42.480244 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a05e1e58-64e0-4b6c-bab9-b4ad7a8eea2b" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Nov 22 08:39:42 crc kubenswrapper[4789]: I1122 08:39:42.480263 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="a05e1e58-64e0-4b6c-bab9-b4ad7a8eea2b" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Nov 22 08:39:42 crc kubenswrapper[4789]: I1122 08:39:42.480429 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="a05e1e58-64e0-4b6c-bab9-b4ad7a8eea2b" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Nov 22 08:39:42 crc kubenswrapper[4789]: I1122 08:39:42.481048 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vt7vg" Nov 22 08:39:42 crc kubenswrapper[4789]: I1122 08:39:42.483070 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 22 08:39:42 crc kubenswrapper[4789]: I1122 08:39:42.483863 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 22 08:39:42 crc kubenswrapper[4789]: I1122 08:39:42.486590 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 22 08:39:42 crc kubenswrapper[4789]: I1122 08:39:42.486857 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rkf88\" (UniqueName: \"kubernetes.io/projected/2f51ca85-18d1-429b-99ce-2785256d7bb8-kube-api-access-rkf88\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-vt7vg\" (UID: \"2f51ca85-18d1-429b-99ce-2785256d7bb8\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vt7vg" Nov 22 08:39:42 crc kubenswrapper[4789]: I1122 08:39:42.486917 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2f51ca85-18d1-429b-99ce-2785256d7bb8-ceph\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-vt7vg\" (UID: \"2f51ca85-18d1-429b-99ce-2785256d7bb8\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vt7vg" Nov 22 08:39:42 crc kubenswrapper[4789]: I1122 08:39:42.487035 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2f51ca85-18d1-429b-99ce-2785256d7bb8-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-vt7vg\" (UID: \"2f51ca85-18d1-429b-99ce-2785256d7bb8\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vt7vg" Nov 22 08:39:42 crc kubenswrapper[4789]: I1122 08:39:42.487091 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2f51ca85-18d1-429b-99ce-2785256d7bb8-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-vt7vg\" (UID: \"2f51ca85-18d1-429b-99ce-2785256d7bb8\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vt7vg" Nov 22 08:39:42 crc kubenswrapper[4789]: I1122 08:39:42.487242 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-glmr7" Nov 22 08:39:42 crc kubenswrapper[4789]: I1122 08:39:42.487321 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 22 08:39:42 crc kubenswrapper[4789]: I1122 08:39:42.497809 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vt7vg"] Nov 22 08:39:42 crc kubenswrapper[4789]: I1122 08:39:42.588809 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2f51ca85-18d1-429b-99ce-2785256d7bb8-ceph\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-vt7vg\" (UID: \"2f51ca85-18d1-429b-99ce-2785256d7bb8\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vt7vg" Nov 22 08:39:42 crc kubenswrapper[4789]: I1122 08:39:42.588975 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2f51ca85-18d1-429b-99ce-2785256d7bb8-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-vt7vg\" (UID: \"2f51ca85-18d1-429b-99ce-2785256d7bb8\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vt7vg" Nov 22 08:39:42 crc kubenswrapper[4789]: I1122 08:39:42.589042 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2f51ca85-18d1-429b-99ce-2785256d7bb8-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-vt7vg\" (UID: \"2f51ca85-18d1-429b-99ce-2785256d7bb8\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vt7vg" Nov 22 08:39:42 crc kubenswrapper[4789]: I1122 08:39:42.589175 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rkf88\" (UniqueName: \"kubernetes.io/projected/2f51ca85-18d1-429b-99ce-2785256d7bb8-kube-api-access-rkf88\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-vt7vg\" (UID: \"2f51ca85-18d1-429b-99ce-2785256d7bb8\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vt7vg" Nov 22 08:39:42 crc kubenswrapper[4789]: I1122 08:39:42.593894 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2f51ca85-18d1-429b-99ce-2785256d7bb8-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-vt7vg\" (UID: \"2f51ca85-18d1-429b-99ce-2785256d7bb8\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vt7vg" Nov 22 08:39:42 crc kubenswrapper[4789]: I1122 08:39:42.596125 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2f51ca85-18d1-429b-99ce-2785256d7bb8-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-vt7vg\" (UID: \"2f51ca85-18d1-429b-99ce-2785256d7bb8\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vt7vg" Nov 22 08:39:42 crc kubenswrapper[4789]: I1122 08:39:42.602120 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2f51ca85-18d1-429b-99ce-2785256d7bb8-ceph\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-vt7vg\" (UID: \"2f51ca85-18d1-429b-99ce-2785256d7bb8\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vt7vg" Nov 22 08:39:42 crc kubenswrapper[4789]: I1122 08:39:42.605805 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rkf88\" (UniqueName: \"kubernetes.io/projected/2f51ca85-18d1-429b-99ce-2785256d7bb8-kube-api-access-rkf88\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-vt7vg\" (UID: \"2f51ca85-18d1-429b-99ce-2785256d7bb8\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vt7vg" Nov 22 08:39:42 crc kubenswrapper[4789]: I1122 08:39:42.815082 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vt7vg" Nov 22 08:39:43 crc kubenswrapper[4789]: I1122 08:39:43.312181 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vt7vg"] Nov 22 08:39:43 crc kubenswrapper[4789]: I1122 08:39:43.422171 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vt7vg" event={"ID":"2f51ca85-18d1-429b-99ce-2785256d7bb8","Type":"ContainerStarted","Data":"90f233513a9076391a34bc763be4bb2950ebcd25fb746442c53984ca5403f891"} Nov 22 08:39:44 crc kubenswrapper[4789]: I1122 08:39:44.431379 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vt7vg" event={"ID":"2f51ca85-18d1-429b-99ce-2785256d7bb8","Type":"ContainerStarted","Data":"e25ffd6d2f90a3f7d095f3acac2970b5c15f95b6dfcb6c723152bc86e6b63fe6"} Nov 22 08:39:44 crc kubenswrapper[4789]: I1122 08:39:44.452049 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vt7vg" podStartSLOduration=1.926575416 podStartE2EDuration="2.452028231s" podCreationTimestamp="2025-11-22 08:39:42 +0000 UTC" firstStartedPulling="2025-11-22 08:39:43.319817158 +0000 UTC m=+2597.554217431" lastFinishedPulling="2025-11-22 08:39:43.845269963 +0000 UTC m=+2598.079670246" observedRunningTime="2025-11-22 08:39:44.447519299 +0000 UTC m=+2598.681919582" watchObservedRunningTime="2025-11-22 08:39:44.452028231 +0000 UTC m=+2598.686428494" Nov 22 08:40:26 crc kubenswrapper[4789]: I1122 08:40:26.784162 4789 generic.go:334] "Generic (PLEG): container finished" podID="2f51ca85-18d1-429b-99ce-2785256d7bb8" containerID="e25ffd6d2f90a3f7d095f3acac2970b5c15f95b6dfcb6c723152bc86e6b63fe6" exitCode=0 Nov 22 08:40:26 crc kubenswrapper[4789]: I1122 08:40:26.784245 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vt7vg" event={"ID":"2f51ca85-18d1-429b-99ce-2785256d7bb8","Type":"ContainerDied","Data":"e25ffd6d2f90a3f7d095f3acac2970b5c15f95b6dfcb6c723152bc86e6b63fe6"} Nov 22 08:40:28 crc kubenswrapper[4789]: I1122 08:40:28.180075 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vt7vg" Nov 22 08:40:28 crc kubenswrapper[4789]: I1122 08:40:28.287242 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2f51ca85-18d1-429b-99ce-2785256d7bb8-ssh-key\") pod \"2f51ca85-18d1-429b-99ce-2785256d7bb8\" (UID: \"2f51ca85-18d1-429b-99ce-2785256d7bb8\") " Nov 22 08:40:28 crc kubenswrapper[4789]: I1122 08:40:28.287688 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2f51ca85-18d1-429b-99ce-2785256d7bb8-inventory\") pod \"2f51ca85-18d1-429b-99ce-2785256d7bb8\" (UID: \"2f51ca85-18d1-429b-99ce-2785256d7bb8\") " Nov 22 08:40:28 crc kubenswrapper[4789]: I1122 08:40:28.287809 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rkf88\" (UniqueName: \"kubernetes.io/projected/2f51ca85-18d1-429b-99ce-2785256d7bb8-kube-api-access-rkf88\") pod \"2f51ca85-18d1-429b-99ce-2785256d7bb8\" (UID: \"2f51ca85-18d1-429b-99ce-2785256d7bb8\") " Nov 22 08:40:28 crc kubenswrapper[4789]: I1122 08:40:28.288168 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2f51ca85-18d1-429b-99ce-2785256d7bb8-ceph\") pod \"2f51ca85-18d1-429b-99ce-2785256d7bb8\" (UID: \"2f51ca85-18d1-429b-99ce-2785256d7bb8\") " Nov 22 08:40:28 crc kubenswrapper[4789]: I1122 08:40:28.293694 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f51ca85-18d1-429b-99ce-2785256d7bb8-ceph" (OuterVolumeSpecName: "ceph") pod "2f51ca85-18d1-429b-99ce-2785256d7bb8" (UID: "2f51ca85-18d1-429b-99ce-2785256d7bb8"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:40:28 crc kubenswrapper[4789]: I1122 08:40:28.295963 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f51ca85-18d1-429b-99ce-2785256d7bb8-kube-api-access-rkf88" (OuterVolumeSpecName: "kube-api-access-rkf88") pod "2f51ca85-18d1-429b-99ce-2785256d7bb8" (UID: "2f51ca85-18d1-429b-99ce-2785256d7bb8"). InnerVolumeSpecName "kube-api-access-rkf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:40:28 crc kubenswrapper[4789]: I1122 08:40:28.324345 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f51ca85-18d1-429b-99ce-2785256d7bb8-inventory" (OuterVolumeSpecName: "inventory") pod "2f51ca85-18d1-429b-99ce-2785256d7bb8" (UID: "2f51ca85-18d1-429b-99ce-2785256d7bb8"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:40:28 crc kubenswrapper[4789]: I1122 08:40:28.330679 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f51ca85-18d1-429b-99ce-2785256d7bb8-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2f51ca85-18d1-429b-99ce-2785256d7bb8" (UID: "2f51ca85-18d1-429b-99ce-2785256d7bb8"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:40:28 crc kubenswrapper[4789]: I1122 08:40:28.390682 4789 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2f51ca85-18d1-429b-99ce-2785256d7bb8-ceph\") on node \"crc\" DevicePath \"\"" Nov 22 08:40:28 crc kubenswrapper[4789]: I1122 08:40:28.390725 4789 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2f51ca85-18d1-429b-99ce-2785256d7bb8-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 22 08:40:28 crc kubenswrapper[4789]: I1122 08:40:28.390737 4789 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2f51ca85-18d1-429b-99ce-2785256d7bb8-inventory\") on node \"crc\" DevicePath \"\"" Nov 22 08:40:28 crc kubenswrapper[4789]: I1122 08:40:28.390760 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rkf88\" (UniqueName: \"kubernetes.io/projected/2f51ca85-18d1-429b-99ce-2785256d7bb8-kube-api-access-rkf88\") on node \"crc\" DevicePath \"\"" Nov 22 08:40:28 crc kubenswrapper[4789]: I1122 08:40:28.806361 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vt7vg" event={"ID":"2f51ca85-18d1-429b-99ce-2785256d7bb8","Type":"ContainerDied","Data":"90f233513a9076391a34bc763be4bb2950ebcd25fb746442c53984ca5403f891"} Nov 22 08:40:28 crc kubenswrapper[4789]: I1122 08:40:28.806403 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="90f233513a9076391a34bc763be4bb2950ebcd25fb746442c53984ca5403f891" Nov 22 08:40:28 crc kubenswrapper[4789]: I1122 08:40:28.806420 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vt7vg" Nov 22 08:40:28 crc kubenswrapper[4789]: I1122 08:40:28.880926 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-h22cd"] Nov 22 08:40:28 crc kubenswrapper[4789]: E1122 08:40:28.881350 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f51ca85-18d1-429b-99ce-2785256d7bb8" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 22 08:40:28 crc kubenswrapper[4789]: I1122 08:40:28.881371 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f51ca85-18d1-429b-99ce-2785256d7bb8" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 22 08:40:28 crc kubenswrapper[4789]: I1122 08:40:28.881534 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f51ca85-18d1-429b-99ce-2785256d7bb8" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 22 08:40:28 crc kubenswrapper[4789]: I1122 08:40:28.882199 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-h22cd" Nov 22 08:40:28 crc kubenswrapper[4789]: I1122 08:40:28.884054 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 22 08:40:28 crc kubenswrapper[4789]: I1122 08:40:28.884698 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 22 08:40:28 crc kubenswrapper[4789]: I1122 08:40:28.884888 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 22 08:40:28 crc kubenswrapper[4789]: I1122 08:40:28.885110 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 22 08:40:28 crc kubenswrapper[4789]: I1122 08:40:28.885774 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-glmr7" Nov 22 08:40:28 crc kubenswrapper[4789]: I1122 08:40:28.890307 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-h22cd"] Nov 22 08:40:28 crc kubenswrapper[4789]: I1122 08:40:28.901910 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3901b861-f71d-4272-ba5c-2fc28c9e7972-ceph\") pod \"ssh-known-hosts-edpm-deployment-h22cd\" (UID: \"3901b861-f71d-4272-ba5c-2fc28c9e7972\") " pod="openstack/ssh-known-hosts-edpm-deployment-h22cd" Nov 22 08:40:28 crc kubenswrapper[4789]: I1122 08:40:28.902123 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/3901b861-f71d-4272-ba5c-2fc28c9e7972-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-h22cd\" (UID: \"3901b861-f71d-4272-ba5c-2fc28c9e7972\") " pod="openstack/ssh-known-hosts-edpm-deployment-h22cd" Nov 22 08:40:28 crc kubenswrapper[4789]: I1122 08:40:28.902280 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-627qp\" (UniqueName: \"kubernetes.io/projected/3901b861-f71d-4272-ba5c-2fc28c9e7972-kube-api-access-627qp\") pod \"ssh-known-hosts-edpm-deployment-h22cd\" (UID: \"3901b861-f71d-4272-ba5c-2fc28c9e7972\") " pod="openstack/ssh-known-hosts-edpm-deployment-h22cd" Nov 22 08:40:28 crc kubenswrapper[4789]: I1122 08:40:28.902331 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3901b861-f71d-4272-ba5c-2fc28c9e7972-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-h22cd\" (UID: \"3901b861-f71d-4272-ba5c-2fc28c9e7972\") " pod="openstack/ssh-known-hosts-edpm-deployment-h22cd" Nov 22 08:40:29 crc kubenswrapper[4789]: I1122 08:40:29.004098 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-627qp\" (UniqueName: \"kubernetes.io/projected/3901b861-f71d-4272-ba5c-2fc28c9e7972-kube-api-access-627qp\") pod \"ssh-known-hosts-edpm-deployment-h22cd\" (UID: \"3901b861-f71d-4272-ba5c-2fc28c9e7972\") " pod="openstack/ssh-known-hosts-edpm-deployment-h22cd" Nov 22 08:40:29 crc kubenswrapper[4789]: I1122 08:40:29.004202 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3901b861-f71d-4272-ba5c-2fc28c9e7972-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-h22cd\" (UID: \"3901b861-f71d-4272-ba5c-2fc28c9e7972\") " pod="openstack/ssh-known-hosts-edpm-deployment-h22cd" Nov 22 08:40:29 crc kubenswrapper[4789]: I1122 08:40:29.004275 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3901b861-f71d-4272-ba5c-2fc28c9e7972-ceph\") pod \"ssh-known-hosts-edpm-deployment-h22cd\" (UID: \"3901b861-f71d-4272-ba5c-2fc28c9e7972\") " pod="openstack/ssh-known-hosts-edpm-deployment-h22cd" Nov 22 08:40:29 crc kubenswrapper[4789]: I1122 08:40:29.004423 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/3901b861-f71d-4272-ba5c-2fc28c9e7972-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-h22cd\" (UID: \"3901b861-f71d-4272-ba5c-2fc28c9e7972\") " pod="openstack/ssh-known-hosts-edpm-deployment-h22cd" Nov 22 08:40:29 crc kubenswrapper[4789]: I1122 08:40:29.009704 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3901b861-f71d-4272-ba5c-2fc28c9e7972-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-h22cd\" (UID: \"3901b861-f71d-4272-ba5c-2fc28c9e7972\") " pod="openstack/ssh-known-hosts-edpm-deployment-h22cd" Nov 22 08:40:29 crc kubenswrapper[4789]: I1122 08:40:29.009899 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/3901b861-f71d-4272-ba5c-2fc28c9e7972-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-h22cd\" (UID: \"3901b861-f71d-4272-ba5c-2fc28c9e7972\") " pod="openstack/ssh-known-hosts-edpm-deployment-h22cd" Nov 22 08:40:29 crc kubenswrapper[4789]: I1122 08:40:29.010613 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3901b861-f71d-4272-ba5c-2fc28c9e7972-ceph\") pod \"ssh-known-hosts-edpm-deployment-h22cd\" (UID: \"3901b861-f71d-4272-ba5c-2fc28c9e7972\") " pod="openstack/ssh-known-hosts-edpm-deployment-h22cd" Nov 22 08:40:29 crc kubenswrapper[4789]: I1122 08:40:29.023457 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-627qp\" (UniqueName: \"kubernetes.io/projected/3901b861-f71d-4272-ba5c-2fc28c9e7972-kube-api-access-627qp\") pod \"ssh-known-hosts-edpm-deployment-h22cd\" (UID: \"3901b861-f71d-4272-ba5c-2fc28c9e7972\") " pod="openstack/ssh-known-hosts-edpm-deployment-h22cd" Nov 22 08:40:29 crc kubenswrapper[4789]: I1122 08:40:29.199275 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-h22cd" Nov 22 08:40:29 crc kubenswrapper[4789]: I1122 08:40:29.706929 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-h22cd"] Nov 22 08:40:29 crc kubenswrapper[4789]: I1122 08:40:29.815838 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-h22cd" event={"ID":"3901b861-f71d-4272-ba5c-2fc28c9e7972","Type":"ContainerStarted","Data":"d46bfe70c4c3f5391413f39655650332e97d8e4c5150b3cc2eed0e5dad3a5867"} Nov 22 08:40:30 crc kubenswrapper[4789]: I1122 08:40:30.827441 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-h22cd" event={"ID":"3901b861-f71d-4272-ba5c-2fc28c9e7972","Type":"ContainerStarted","Data":"290d3d9dccae2bf9bef87108ed3547046798a5f908c5bd6eaa2e85caea59751e"} Nov 22 08:40:30 crc kubenswrapper[4789]: I1122 08:40:30.850092 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-h22cd" podStartSLOduration=2.450717366 podStartE2EDuration="2.850067542s" podCreationTimestamp="2025-11-22 08:40:28 +0000 UTC" firstStartedPulling="2025-11-22 08:40:29.711989749 +0000 UTC m=+2643.946390022" lastFinishedPulling="2025-11-22 08:40:30.111339925 +0000 UTC m=+2644.345740198" observedRunningTime="2025-11-22 08:40:30.843568045 +0000 UTC m=+2645.077968338" watchObservedRunningTime="2025-11-22 08:40:30.850067542 +0000 UTC m=+2645.084467835" Nov 22 08:40:39 crc kubenswrapper[4789]: I1122 08:40:39.902673 4789 generic.go:334] "Generic (PLEG): container finished" podID="3901b861-f71d-4272-ba5c-2fc28c9e7972" containerID="290d3d9dccae2bf9bef87108ed3547046798a5f908c5bd6eaa2e85caea59751e" exitCode=0 Nov 22 08:40:39 crc kubenswrapper[4789]: I1122 08:40:39.902815 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-h22cd" event={"ID":"3901b861-f71d-4272-ba5c-2fc28c9e7972","Type":"ContainerDied","Data":"290d3d9dccae2bf9bef87108ed3547046798a5f908c5bd6eaa2e85caea59751e"} Nov 22 08:40:41 crc kubenswrapper[4789]: I1122 08:40:41.319853 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-h22cd" Nov 22 08:40:41 crc kubenswrapper[4789]: I1122 08:40:41.328493 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-627qp\" (UniqueName: \"kubernetes.io/projected/3901b861-f71d-4272-ba5c-2fc28c9e7972-kube-api-access-627qp\") pod \"3901b861-f71d-4272-ba5c-2fc28c9e7972\" (UID: \"3901b861-f71d-4272-ba5c-2fc28c9e7972\") " Nov 22 08:40:41 crc kubenswrapper[4789]: I1122 08:40:41.328564 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/3901b861-f71d-4272-ba5c-2fc28c9e7972-inventory-0\") pod \"3901b861-f71d-4272-ba5c-2fc28c9e7972\" (UID: \"3901b861-f71d-4272-ba5c-2fc28c9e7972\") " Nov 22 08:40:41 crc kubenswrapper[4789]: I1122 08:40:41.328647 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3901b861-f71d-4272-ba5c-2fc28c9e7972-ssh-key-openstack-edpm-ipam\") pod \"3901b861-f71d-4272-ba5c-2fc28c9e7972\" (UID: \"3901b861-f71d-4272-ba5c-2fc28c9e7972\") " Nov 22 08:40:41 crc kubenswrapper[4789]: I1122 08:40:41.328828 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3901b861-f71d-4272-ba5c-2fc28c9e7972-ceph\") pod \"3901b861-f71d-4272-ba5c-2fc28c9e7972\" (UID: \"3901b861-f71d-4272-ba5c-2fc28c9e7972\") " Nov 22 08:40:41 crc kubenswrapper[4789]: I1122 08:40:41.337904 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3901b861-f71d-4272-ba5c-2fc28c9e7972-kube-api-access-627qp" (OuterVolumeSpecName: "kube-api-access-627qp") pod "3901b861-f71d-4272-ba5c-2fc28c9e7972" (UID: "3901b861-f71d-4272-ba5c-2fc28c9e7972"). InnerVolumeSpecName "kube-api-access-627qp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:40:41 crc kubenswrapper[4789]: I1122 08:40:41.343941 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3901b861-f71d-4272-ba5c-2fc28c9e7972-ceph" (OuterVolumeSpecName: "ceph") pod "3901b861-f71d-4272-ba5c-2fc28c9e7972" (UID: "3901b861-f71d-4272-ba5c-2fc28c9e7972"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:40:41 crc kubenswrapper[4789]: I1122 08:40:41.357904 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3901b861-f71d-4272-ba5c-2fc28c9e7972-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "3901b861-f71d-4272-ba5c-2fc28c9e7972" (UID: "3901b861-f71d-4272-ba5c-2fc28c9e7972"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:40:41 crc kubenswrapper[4789]: I1122 08:40:41.367049 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3901b861-f71d-4272-ba5c-2fc28c9e7972-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "3901b861-f71d-4272-ba5c-2fc28c9e7972" (UID: "3901b861-f71d-4272-ba5c-2fc28c9e7972"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:40:41 crc kubenswrapper[4789]: I1122 08:40:41.431823 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-627qp\" (UniqueName: \"kubernetes.io/projected/3901b861-f71d-4272-ba5c-2fc28c9e7972-kube-api-access-627qp\") on node \"crc\" DevicePath \"\"" Nov 22 08:40:41 crc kubenswrapper[4789]: I1122 08:40:41.431873 4789 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/3901b861-f71d-4272-ba5c-2fc28c9e7972-inventory-0\") on node \"crc\" DevicePath \"\"" Nov 22 08:40:41 crc kubenswrapper[4789]: I1122 08:40:41.431885 4789 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3901b861-f71d-4272-ba5c-2fc28c9e7972-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Nov 22 08:40:41 crc kubenswrapper[4789]: I1122 08:40:41.431897 4789 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3901b861-f71d-4272-ba5c-2fc28c9e7972-ceph\") on node \"crc\" DevicePath \"\"" Nov 22 08:40:41 crc kubenswrapper[4789]: I1122 08:40:41.921142 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-h22cd" event={"ID":"3901b861-f71d-4272-ba5c-2fc28c9e7972","Type":"ContainerDied","Data":"d46bfe70c4c3f5391413f39655650332e97d8e4c5150b3cc2eed0e5dad3a5867"} Nov 22 08:40:41 crc kubenswrapper[4789]: I1122 08:40:41.921199 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d46bfe70c4c3f5391413f39655650332e97d8e4c5150b3cc2eed0e5dad3a5867" Nov 22 08:40:41 crc kubenswrapper[4789]: I1122 08:40:41.921254 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-h22cd" Nov 22 08:40:41 crc kubenswrapper[4789]: I1122 08:40:41.994012 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-bmfgh"] Nov 22 08:40:41 crc kubenswrapper[4789]: E1122 08:40:41.994477 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3901b861-f71d-4272-ba5c-2fc28c9e7972" containerName="ssh-known-hosts-edpm-deployment" Nov 22 08:40:41 crc kubenswrapper[4789]: I1122 08:40:41.994499 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="3901b861-f71d-4272-ba5c-2fc28c9e7972" containerName="ssh-known-hosts-edpm-deployment" Nov 22 08:40:41 crc kubenswrapper[4789]: I1122 08:40:41.994714 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="3901b861-f71d-4272-ba5c-2fc28c9e7972" containerName="ssh-known-hosts-edpm-deployment" Nov 22 08:40:41 crc kubenswrapper[4789]: I1122 08:40:41.995673 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-bmfgh" Nov 22 08:40:41 crc kubenswrapper[4789]: I1122 08:40:41.999107 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 22 08:40:41 crc kubenswrapper[4789]: I1122 08:40:41.999281 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 22 08:40:41 crc kubenswrapper[4789]: I1122 08:40:41.999397 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-glmr7" Nov 22 08:40:41 crc kubenswrapper[4789]: I1122 08:40:41.999560 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 22 08:40:42 crc kubenswrapper[4789]: I1122 08:40:42.001220 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 22 08:40:42 crc kubenswrapper[4789]: I1122 08:40:42.008328 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-bmfgh"] Nov 22 08:40:42 crc kubenswrapper[4789]: I1122 08:40:42.040383 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b8ccf801-a404-4a95-bda9-5a5569d45a51-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-bmfgh\" (UID: \"b8ccf801-a404-4a95-bda9-5a5569d45a51\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-bmfgh" Nov 22 08:40:42 crc kubenswrapper[4789]: I1122 08:40:42.040660 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-78rrc\" (UniqueName: \"kubernetes.io/projected/b8ccf801-a404-4a95-bda9-5a5569d45a51-kube-api-access-78rrc\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-bmfgh\" (UID: \"b8ccf801-a404-4a95-bda9-5a5569d45a51\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-bmfgh" Nov 22 08:40:42 crc kubenswrapper[4789]: I1122 08:40:42.040812 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b8ccf801-a404-4a95-bda9-5a5569d45a51-ceph\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-bmfgh\" (UID: \"b8ccf801-a404-4a95-bda9-5a5569d45a51\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-bmfgh" Nov 22 08:40:42 crc kubenswrapper[4789]: I1122 08:40:42.041006 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b8ccf801-a404-4a95-bda9-5a5569d45a51-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-bmfgh\" (UID: \"b8ccf801-a404-4a95-bda9-5a5569d45a51\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-bmfgh" Nov 22 08:40:42 crc kubenswrapper[4789]: I1122 08:40:42.142771 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b8ccf801-a404-4a95-bda9-5a5569d45a51-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-bmfgh\" (UID: \"b8ccf801-a404-4a95-bda9-5a5569d45a51\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-bmfgh" Nov 22 08:40:42 crc kubenswrapper[4789]: I1122 08:40:42.142899 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b8ccf801-a404-4a95-bda9-5a5569d45a51-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-bmfgh\" (UID: \"b8ccf801-a404-4a95-bda9-5a5569d45a51\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-bmfgh" Nov 22 08:40:42 crc kubenswrapper[4789]: I1122 08:40:42.142931 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-78rrc\" (UniqueName: \"kubernetes.io/projected/b8ccf801-a404-4a95-bda9-5a5569d45a51-kube-api-access-78rrc\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-bmfgh\" (UID: \"b8ccf801-a404-4a95-bda9-5a5569d45a51\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-bmfgh" Nov 22 08:40:42 crc kubenswrapper[4789]: I1122 08:40:42.142983 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b8ccf801-a404-4a95-bda9-5a5569d45a51-ceph\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-bmfgh\" (UID: \"b8ccf801-a404-4a95-bda9-5a5569d45a51\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-bmfgh" Nov 22 08:40:42 crc kubenswrapper[4789]: I1122 08:40:42.146511 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b8ccf801-a404-4a95-bda9-5a5569d45a51-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-bmfgh\" (UID: \"b8ccf801-a404-4a95-bda9-5a5569d45a51\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-bmfgh" Nov 22 08:40:42 crc kubenswrapper[4789]: I1122 08:40:42.146527 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b8ccf801-a404-4a95-bda9-5a5569d45a51-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-bmfgh\" (UID: \"b8ccf801-a404-4a95-bda9-5a5569d45a51\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-bmfgh" Nov 22 08:40:42 crc kubenswrapper[4789]: I1122 08:40:42.147811 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b8ccf801-a404-4a95-bda9-5a5569d45a51-ceph\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-bmfgh\" (UID: \"b8ccf801-a404-4a95-bda9-5a5569d45a51\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-bmfgh" Nov 22 08:40:42 crc kubenswrapper[4789]: I1122 08:40:42.161454 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-78rrc\" (UniqueName: \"kubernetes.io/projected/b8ccf801-a404-4a95-bda9-5a5569d45a51-kube-api-access-78rrc\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-bmfgh\" (UID: \"b8ccf801-a404-4a95-bda9-5a5569d45a51\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-bmfgh" Nov 22 08:40:42 crc kubenswrapper[4789]: I1122 08:40:42.330282 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-bmfgh" Nov 22 08:40:42 crc kubenswrapper[4789]: I1122 08:40:42.841654 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-bmfgh"] Nov 22 08:40:42 crc kubenswrapper[4789]: I1122 08:40:42.933173 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-bmfgh" event={"ID":"b8ccf801-a404-4a95-bda9-5a5569d45a51","Type":"ContainerStarted","Data":"86d8cd5ec2351d47463beed5f85e7712d278d2c4666a5097f6493e2402c6dd5f"} Nov 22 08:40:44 crc kubenswrapper[4789]: I1122 08:40:44.951504 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-bmfgh" event={"ID":"b8ccf801-a404-4a95-bda9-5a5569d45a51","Type":"ContainerStarted","Data":"10f0619bc5d4af57fe2bab8a17668be9168fbb24e63fea93796f8b7dd81c2a4d"} Nov 22 08:40:44 crc kubenswrapper[4789]: I1122 08:40:44.982417 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-bmfgh" podStartSLOduration=2.926241905 podStartE2EDuration="3.982394917s" podCreationTimestamp="2025-11-22 08:40:41 +0000 UTC" firstStartedPulling="2025-11-22 08:40:42.843699759 +0000 UTC m=+2657.078100032" lastFinishedPulling="2025-11-22 08:40:43.899852761 +0000 UTC m=+2658.134253044" observedRunningTime="2025-11-22 08:40:44.975145291 +0000 UTC m=+2659.209545584" watchObservedRunningTime="2025-11-22 08:40:44.982394917 +0000 UTC m=+2659.216795200" Nov 22 08:40:52 crc kubenswrapper[4789]: I1122 08:40:52.015849 4789 generic.go:334] "Generic (PLEG): container finished" podID="b8ccf801-a404-4a95-bda9-5a5569d45a51" containerID="10f0619bc5d4af57fe2bab8a17668be9168fbb24e63fea93796f8b7dd81c2a4d" exitCode=0 Nov 22 08:40:52 crc kubenswrapper[4789]: I1122 08:40:52.015925 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-bmfgh" event={"ID":"b8ccf801-a404-4a95-bda9-5a5569d45a51","Type":"ContainerDied","Data":"10f0619bc5d4af57fe2bab8a17668be9168fbb24e63fea93796f8b7dd81c2a4d"} Nov 22 08:40:53 crc kubenswrapper[4789]: I1122 08:40:53.410529 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-bmfgh" Nov 22 08:40:53 crc kubenswrapper[4789]: I1122 08:40:53.551090 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b8ccf801-a404-4a95-bda9-5a5569d45a51-ceph\") pod \"b8ccf801-a404-4a95-bda9-5a5569d45a51\" (UID: \"b8ccf801-a404-4a95-bda9-5a5569d45a51\") " Nov 22 08:40:53 crc kubenswrapper[4789]: I1122 08:40:53.551212 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b8ccf801-a404-4a95-bda9-5a5569d45a51-inventory\") pod \"b8ccf801-a404-4a95-bda9-5a5569d45a51\" (UID: \"b8ccf801-a404-4a95-bda9-5a5569d45a51\") " Nov 22 08:40:53 crc kubenswrapper[4789]: I1122 08:40:53.551256 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b8ccf801-a404-4a95-bda9-5a5569d45a51-ssh-key\") pod \"b8ccf801-a404-4a95-bda9-5a5569d45a51\" (UID: \"b8ccf801-a404-4a95-bda9-5a5569d45a51\") " Nov 22 08:40:53 crc kubenswrapper[4789]: I1122 08:40:53.551307 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-78rrc\" (UniqueName: \"kubernetes.io/projected/b8ccf801-a404-4a95-bda9-5a5569d45a51-kube-api-access-78rrc\") pod \"b8ccf801-a404-4a95-bda9-5a5569d45a51\" (UID: \"b8ccf801-a404-4a95-bda9-5a5569d45a51\") " Nov 22 08:40:53 crc kubenswrapper[4789]: I1122 08:40:53.571177 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b8ccf801-a404-4a95-bda9-5a5569d45a51-kube-api-access-78rrc" (OuterVolumeSpecName: "kube-api-access-78rrc") pod "b8ccf801-a404-4a95-bda9-5a5569d45a51" (UID: "b8ccf801-a404-4a95-bda9-5a5569d45a51"). InnerVolumeSpecName "kube-api-access-78rrc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:40:53 crc kubenswrapper[4789]: I1122 08:40:53.582718 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8ccf801-a404-4a95-bda9-5a5569d45a51-ceph" (OuterVolumeSpecName: "ceph") pod "b8ccf801-a404-4a95-bda9-5a5569d45a51" (UID: "b8ccf801-a404-4a95-bda9-5a5569d45a51"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:40:53 crc kubenswrapper[4789]: I1122 08:40:53.596539 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8ccf801-a404-4a95-bda9-5a5569d45a51-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b8ccf801-a404-4a95-bda9-5a5569d45a51" (UID: "b8ccf801-a404-4a95-bda9-5a5569d45a51"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:40:53 crc kubenswrapper[4789]: I1122 08:40:53.597044 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8ccf801-a404-4a95-bda9-5a5569d45a51-inventory" (OuterVolumeSpecName: "inventory") pod "b8ccf801-a404-4a95-bda9-5a5569d45a51" (UID: "b8ccf801-a404-4a95-bda9-5a5569d45a51"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:40:53 crc kubenswrapper[4789]: I1122 08:40:53.655295 4789 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b8ccf801-a404-4a95-bda9-5a5569d45a51-ceph\") on node \"crc\" DevicePath \"\"" Nov 22 08:40:53 crc kubenswrapper[4789]: I1122 08:40:53.655369 4789 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b8ccf801-a404-4a95-bda9-5a5569d45a51-inventory\") on node \"crc\" DevicePath \"\"" Nov 22 08:40:53 crc kubenswrapper[4789]: I1122 08:40:53.655386 4789 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b8ccf801-a404-4a95-bda9-5a5569d45a51-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 22 08:40:53 crc kubenswrapper[4789]: I1122 08:40:53.655400 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-78rrc\" (UniqueName: \"kubernetes.io/projected/b8ccf801-a404-4a95-bda9-5a5569d45a51-kube-api-access-78rrc\") on node \"crc\" DevicePath \"\"" Nov 22 08:40:54 crc kubenswrapper[4789]: I1122 08:40:54.033254 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-bmfgh" event={"ID":"b8ccf801-a404-4a95-bda9-5a5569d45a51","Type":"ContainerDied","Data":"86d8cd5ec2351d47463beed5f85e7712d278d2c4666a5097f6493e2402c6dd5f"} Nov 22 08:40:54 crc kubenswrapper[4789]: I1122 08:40:54.033529 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="86d8cd5ec2351d47463beed5f85e7712d278d2c4666a5097f6493e2402c6dd5f" Nov 22 08:40:54 crc kubenswrapper[4789]: I1122 08:40:54.033413 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-bmfgh" Nov 22 08:40:54 crc kubenswrapper[4789]: I1122 08:40:54.104309 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-pbfnm"] Nov 22 08:40:54 crc kubenswrapper[4789]: E1122 08:40:54.104680 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8ccf801-a404-4a95-bda9-5a5569d45a51" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 22 08:40:54 crc kubenswrapper[4789]: I1122 08:40:54.104699 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8ccf801-a404-4a95-bda9-5a5569d45a51" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 22 08:40:54 crc kubenswrapper[4789]: I1122 08:40:54.104986 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8ccf801-a404-4a95-bda9-5a5569d45a51" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 22 08:40:54 crc kubenswrapper[4789]: I1122 08:40:54.105708 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-pbfnm" Nov 22 08:40:54 crc kubenswrapper[4789]: I1122 08:40:54.108415 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 22 08:40:54 crc kubenswrapper[4789]: I1122 08:40:54.108585 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 22 08:40:54 crc kubenswrapper[4789]: I1122 08:40:54.108613 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-glmr7" Nov 22 08:40:54 crc kubenswrapper[4789]: I1122 08:40:54.108911 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 22 08:40:54 crc kubenswrapper[4789]: I1122 08:40:54.110079 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 22 08:40:54 crc kubenswrapper[4789]: I1122 08:40:54.115373 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-pbfnm"] Nov 22 08:40:54 crc kubenswrapper[4789]: I1122 08:40:54.266674 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d82458d2-439e-479d-8315-32c793ed9401-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-pbfnm\" (UID: \"d82458d2-439e-479d-8315-32c793ed9401\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-pbfnm" Nov 22 08:40:54 crc kubenswrapper[4789]: I1122 08:40:54.266764 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xgw4l\" (UniqueName: \"kubernetes.io/projected/d82458d2-439e-479d-8315-32c793ed9401-kube-api-access-xgw4l\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-pbfnm\" (UID: \"d82458d2-439e-479d-8315-32c793ed9401\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-pbfnm" Nov 22 08:40:54 crc kubenswrapper[4789]: I1122 08:40:54.266934 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d82458d2-439e-479d-8315-32c793ed9401-ceph\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-pbfnm\" (UID: \"d82458d2-439e-479d-8315-32c793ed9401\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-pbfnm" Nov 22 08:40:54 crc kubenswrapper[4789]: I1122 08:40:54.266961 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d82458d2-439e-479d-8315-32c793ed9401-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-pbfnm\" (UID: \"d82458d2-439e-479d-8315-32c793ed9401\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-pbfnm" Nov 22 08:40:54 crc kubenswrapper[4789]: I1122 08:40:54.369099 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d82458d2-439e-479d-8315-32c793ed9401-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-pbfnm\" (UID: \"d82458d2-439e-479d-8315-32c793ed9401\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-pbfnm" Nov 22 08:40:54 crc kubenswrapper[4789]: I1122 08:40:54.369217 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xgw4l\" (UniqueName: \"kubernetes.io/projected/d82458d2-439e-479d-8315-32c793ed9401-kube-api-access-xgw4l\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-pbfnm\" (UID: \"d82458d2-439e-479d-8315-32c793ed9401\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-pbfnm" Nov 22 08:40:54 crc kubenswrapper[4789]: I1122 08:40:54.369501 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d82458d2-439e-479d-8315-32c793ed9401-ceph\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-pbfnm\" (UID: \"d82458d2-439e-479d-8315-32c793ed9401\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-pbfnm" Nov 22 08:40:54 crc kubenswrapper[4789]: I1122 08:40:54.369557 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d82458d2-439e-479d-8315-32c793ed9401-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-pbfnm\" (UID: \"d82458d2-439e-479d-8315-32c793ed9401\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-pbfnm" Nov 22 08:40:54 crc kubenswrapper[4789]: I1122 08:40:54.373297 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d82458d2-439e-479d-8315-32c793ed9401-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-pbfnm\" (UID: \"d82458d2-439e-479d-8315-32c793ed9401\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-pbfnm" Nov 22 08:40:54 crc kubenswrapper[4789]: I1122 08:40:54.373329 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d82458d2-439e-479d-8315-32c793ed9401-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-pbfnm\" (UID: \"d82458d2-439e-479d-8315-32c793ed9401\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-pbfnm" Nov 22 08:40:54 crc kubenswrapper[4789]: I1122 08:40:54.373377 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d82458d2-439e-479d-8315-32c793ed9401-ceph\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-pbfnm\" (UID: \"d82458d2-439e-479d-8315-32c793ed9401\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-pbfnm" Nov 22 08:40:54 crc kubenswrapper[4789]: I1122 08:40:54.387516 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xgw4l\" (UniqueName: \"kubernetes.io/projected/d82458d2-439e-479d-8315-32c793ed9401-kube-api-access-xgw4l\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-pbfnm\" (UID: \"d82458d2-439e-479d-8315-32c793ed9401\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-pbfnm" Nov 22 08:40:54 crc kubenswrapper[4789]: I1122 08:40:54.423796 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-pbfnm" Nov 22 08:40:54 crc kubenswrapper[4789]: I1122 08:40:54.899591 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-pbfnm"] Nov 22 08:40:55 crc kubenswrapper[4789]: I1122 08:40:55.041371 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-pbfnm" event={"ID":"d82458d2-439e-479d-8315-32c793ed9401","Type":"ContainerStarted","Data":"6e6b69e477805bb480a5d26aafe85d336711ca8100f2d0eaa2c9db58cd057aa3"} Nov 22 08:40:57 crc kubenswrapper[4789]: I1122 08:40:57.057857 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-pbfnm" event={"ID":"d82458d2-439e-479d-8315-32c793ed9401","Type":"ContainerStarted","Data":"3d0e77005b7e07366219586891f9a62c743e1296c627cf019e47a7d7ffae26a8"} Nov 22 08:40:57 crc kubenswrapper[4789]: I1122 08:40:57.075677 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-pbfnm" podStartSLOduration=2.0675465060000002 podStartE2EDuration="3.075658855s" podCreationTimestamp="2025-11-22 08:40:54 +0000 UTC" firstStartedPulling="2025-11-22 08:40:54.906334817 +0000 UTC m=+2669.140735090" lastFinishedPulling="2025-11-22 08:40:55.914447166 +0000 UTC m=+2670.148847439" observedRunningTime="2025-11-22 08:40:57.072270424 +0000 UTC m=+2671.306670697" watchObservedRunningTime="2025-11-22 08:40:57.075658855 +0000 UTC m=+2671.310059118" Nov 22 08:41:06 crc kubenswrapper[4789]: I1122 08:41:06.136111 4789 generic.go:334] "Generic (PLEG): container finished" podID="d82458d2-439e-479d-8315-32c793ed9401" containerID="3d0e77005b7e07366219586891f9a62c743e1296c627cf019e47a7d7ffae26a8" exitCode=0 Nov 22 08:41:06 crc kubenswrapper[4789]: I1122 08:41:06.136200 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-pbfnm" event={"ID":"d82458d2-439e-479d-8315-32c793ed9401","Type":"ContainerDied","Data":"3d0e77005b7e07366219586891f9a62c743e1296c627cf019e47a7d7ffae26a8"} Nov 22 08:41:07 crc kubenswrapper[4789]: I1122 08:41:07.541207 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-pbfnm" Nov 22 08:41:07 crc kubenswrapper[4789]: I1122 08:41:07.700010 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d82458d2-439e-479d-8315-32c793ed9401-ssh-key\") pod \"d82458d2-439e-479d-8315-32c793ed9401\" (UID: \"d82458d2-439e-479d-8315-32c793ed9401\") " Nov 22 08:41:07 crc kubenswrapper[4789]: I1122 08:41:07.700362 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d82458d2-439e-479d-8315-32c793ed9401-ceph\") pod \"d82458d2-439e-479d-8315-32c793ed9401\" (UID: \"d82458d2-439e-479d-8315-32c793ed9401\") " Nov 22 08:41:07 crc kubenswrapper[4789]: I1122 08:41:07.700458 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xgw4l\" (UniqueName: \"kubernetes.io/projected/d82458d2-439e-479d-8315-32c793ed9401-kube-api-access-xgw4l\") pod \"d82458d2-439e-479d-8315-32c793ed9401\" (UID: \"d82458d2-439e-479d-8315-32c793ed9401\") " Nov 22 08:41:07 crc kubenswrapper[4789]: I1122 08:41:07.700528 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d82458d2-439e-479d-8315-32c793ed9401-inventory\") pod \"d82458d2-439e-479d-8315-32c793ed9401\" (UID: \"d82458d2-439e-479d-8315-32c793ed9401\") " Nov 22 08:41:07 crc kubenswrapper[4789]: I1122 08:41:07.705828 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d82458d2-439e-479d-8315-32c793ed9401-kube-api-access-xgw4l" (OuterVolumeSpecName: "kube-api-access-xgw4l") pod "d82458d2-439e-479d-8315-32c793ed9401" (UID: "d82458d2-439e-479d-8315-32c793ed9401"). InnerVolumeSpecName "kube-api-access-xgw4l". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:41:07 crc kubenswrapper[4789]: I1122 08:41:07.707109 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d82458d2-439e-479d-8315-32c793ed9401-ceph" (OuterVolumeSpecName: "ceph") pod "d82458d2-439e-479d-8315-32c793ed9401" (UID: "d82458d2-439e-479d-8315-32c793ed9401"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:41:07 crc kubenswrapper[4789]: I1122 08:41:07.728410 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d82458d2-439e-479d-8315-32c793ed9401-inventory" (OuterVolumeSpecName: "inventory") pod "d82458d2-439e-479d-8315-32c793ed9401" (UID: "d82458d2-439e-479d-8315-32c793ed9401"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:41:07 crc kubenswrapper[4789]: I1122 08:41:07.728806 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d82458d2-439e-479d-8315-32c793ed9401-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d82458d2-439e-479d-8315-32c793ed9401" (UID: "d82458d2-439e-479d-8315-32c793ed9401"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:41:07 crc kubenswrapper[4789]: I1122 08:41:07.802504 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xgw4l\" (UniqueName: \"kubernetes.io/projected/d82458d2-439e-479d-8315-32c793ed9401-kube-api-access-xgw4l\") on node \"crc\" DevicePath \"\"" Nov 22 08:41:07 crc kubenswrapper[4789]: I1122 08:41:07.802534 4789 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d82458d2-439e-479d-8315-32c793ed9401-inventory\") on node \"crc\" DevicePath \"\"" Nov 22 08:41:07 crc kubenswrapper[4789]: I1122 08:41:07.802548 4789 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d82458d2-439e-479d-8315-32c793ed9401-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 22 08:41:07 crc kubenswrapper[4789]: I1122 08:41:07.802560 4789 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d82458d2-439e-479d-8315-32c793ed9401-ceph\") on node \"crc\" DevicePath \"\"" Nov 22 08:41:08 crc kubenswrapper[4789]: I1122 08:41:08.156211 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-pbfnm" event={"ID":"d82458d2-439e-479d-8315-32c793ed9401","Type":"ContainerDied","Data":"6e6b69e477805bb480a5d26aafe85d336711ca8100f2d0eaa2c9db58cd057aa3"} Nov 22 08:41:08 crc kubenswrapper[4789]: I1122 08:41:08.156251 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6e6b69e477805bb480a5d26aafe85d336711ca8100f2d0eaa2c9db58cd057aa3" Nov 22 08:41:08 crc kubenswrapper[4789]: I1122 08:41:08.156324 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-pbfnm" Nov 22 08:41:08 crc kubenswrapper[4789]: I1122 08:41:08.229455 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nvsts"] Nov 22 08:41:08 crc kubenswrapper[4789]: E1122 08:41:08.229816 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d82458d2-439e-479d-8315-32c793ed9401" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 22 08:41:08 crc kubenswrapper[4789]: I1122 08:41:08.229827 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="d82458d2-439e-479d-8315-32c793ed9401" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 22 08:41:08 crc kubenswrapper[4789]: I1122 08:41:08.230012 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="d82458d2-439e-479d-8315-32c793ed9401" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 22 08:41:08 crc kubenswrapper[4789]: I1122 08:41:08.230583 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nvsts" Nov 22 08:41:08 crc kubenswrapper[4789]: I1122 08:41:08.233012 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Nov 22 08:41:08 crc kubenswrapper[4789]: I1122 08:41:08.233275 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 22 08:41:08 crc kubenswrapper[4789]: I1122 08:41:08.233393 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 22 08:41:08 crc kubenswrapper[4789]: I1122 08:41:08.234665 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 22 08:41:08 crc kubenswrapper[4789]: I1122 08:41:08.234729 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Nov 22 08:41:08 crc kubenswrapper[4789]: I1122 08:41:08.235463 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-glmr7" Nov 22 08:41:08 crc kubenswrapper[4789]: I1122 08:41:08.235481 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Nov 22 08:41:08 crc kubenswrapper[4789]: I1122 08:41:08.242506 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nvsts"] Nov 22 08:41:08 crc kubenswrapper[4789]: I1122 08:41:08.244123 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 22 08:41:08 crc kubenswrapper[4789]: I1122 08:41:08.414242 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07bdec58-1883-4ddb-9aa0-b6eca2498ed4-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nvsts\" (UID: \"07bdec58-1883-4ddb-9aa0-b6eca2498ed4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nvsts" Nov 22 08:41:08 crc kubenswrapper[4789]: I1122 08:41:08.414332 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/07bdec58-1883-4ddb-9aa0-b6eca2498ed4-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nvsts\" (UID: \"07bdec58-1883-4ddb-9aa0-b6eca2498ed4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nvsts" Nov 22 08:41:08 crc kubenswrapper[4789]: I1122 08:41:08.414539 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/07bdec58-1883-4ddb-9aa0-b6eca2498ed4-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nvsts\" (UID: \"07bdec58-1883-4ddb-9aa0-b6eca2498ed4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nvsts" Nov 22 08:41:08 crc kubenswrapper[4789]: I1122 08:41:08.414608 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/07bdec58-1883-4ddb-9aa0-b6eca2498ed4-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nvsts\" (UID: \"07bdec58-1883-4ddb-9aa0-b6eca2498ed4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nvsts" Nov 22 08:41:08 crc kubenswrapper[4789]: I1122 08:41:08.414658 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07bdec58-1883-4ddb-9aa0-b6eca2498ed4-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nvsts\" (UID: \"07bdec58-1883-4ddb-9aa0-b6eca2498ed4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nvsts" Nov 22 08:41:08 crc kubenswrapper[4789]: I1122 08:41:08.414870 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07bdec58-1883-4ddb-9aa0-b6eca2498ed4-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nvsts\" (UID: \"07bdec58-1883-4ddb-9aa0-b6eca2498ed4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nvsts" Nov 22 08:41:08 crc kubenswrapper[4789]: I1122 08:41:08.414977 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/07bdec58-1883-4ddb-9aa0-b6eca2498ed4-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nvsts\" (UID: \"07bdec58-1883-4ddb-9aa0-b6eca2498ed4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nvsts" Nov 22 08:41:08 crc kubenswrapper[4789]: I1122 08:41:08.415038 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/07bdec58-1883-4ddb-9aa0-b6eca2498ed4-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nvsts\" (UID: \"07bdec58-1883-4ddb-9aa0-b6eca2498ed4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nvsts" Nov 22 08:41:08 crc kubenswrapper[4789]: I1122 08:41:08.415086 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07bdec58-1883-4ddb-9aa0-b6eca2498ed4-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nvsts\" (UID: \"07bdec58-1883-4ddb-9aa0-b6eca2498ed4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nvsts" Nov 22 08:41:08 crc kubenswrapper[4789]: I1122 08:41:08.415122 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07bdec58-1883-4ddb-9aa0-b6eca2498ed4-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nvsts\" (UID: \"07bdec58-1883-4ddb-9aa0-b6eca2498ed4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nvsts" Nov 22 08:41:08 crc kubenswrapper[4789]: I1122 08:41:08.415254 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z4wkg\" (UniqueName: \"kubernetes.io/projected/07bdec58-1883-4ddb-9aa0-b6eca2498ed4-kube-api-access-z4wkg\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nvsts\" (UID: \"07bdec58-1883-4ddb-9aa0-b6eca2498ed4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nvsts" Nov 22 08:41:08 crc kubenswrapper[4789]: I1122 08:41:08.415327 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07bdec58-1883-4ddb-9aa0-b6eca2498ed4-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nvsts\" (UID: \"07bdec58-1883-4ddb-9aa0-b6eca2498ed4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nvsts" Nov 22 08:41:08 crc kubenswrapper[4789]: I1122 08:41:08.415361 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/07bdec58-1883-4ddb-9aa0-b6eca2498ed4-ceph\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nvsts\" (UID: \"07bdec58-1883-4ddb-9aa0-b6eca2498ed4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nvsts" Nov 22 08:41:08 crc kubenswrapper[4789]: I1122 08:41:08.516316 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/07bdec58-1883-4ddb-9aa0-b6eca2498ed4-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nvsts\" (UID: \"07bdec58-1883-4ddb-9aa0-b6eca2498ed4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nvsts" Nov 22 08:41:08 crc kubenswrapper[4789]: I1122 08:41:08.516381 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/07bdec58-1883-4ddb-9aa0-b6eca2498ed4-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nvsts\" (UID: \"07bdec58-1883-4ddb-9aa0-b6eca2498ed4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nvsts" Nov 22 08:41:08 crc kubenswrapper[4789]: I1122 08:41:08.516414 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07bdec58-1883-4ddb-9aa0-b6eca2498ed4-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nvsts\" (UID: \"07bdec58-1883-4ddb-9aa0-b6eca2498ed4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nvsts" Nov 22 08:41:08 crc kubenswrapper[4789]: I1122 08:41:08.516438 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07bdec58-1883-4ddb-9aa0-b6eca2498ed4-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nvsts\" (UID: \"07bdec58-1883-4ddb-9aa0-b6eca2498ed4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nvsts" Nov 22 08:41:08 crc kubenswrapper[4789]: I1122 08:41:08.516474 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z4wkg\" (UniqueName: \"kubernetes.io/projected/07bdec58-1883-4ddb-9aa0-b6eca2498ed4-kube-api-access-z4wkg\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nvsts\" (UID: \"07bdec58-1883-4ddb-9aa0-b6eca2498ed4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nvsts" Nov 22 08:41:08 crc kubenswrapper[4789]: I1122 08:41:08.516509 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07bdec58-1883-4ddb-9aa0-b6eca2498ed4-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nvsts\" (UID: \"07bdec58-1883-4ddb-9aa0-b6eca2498ed4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nvsts" Nov 22 08:41:08 crc kubenswrapper[4789]: I1122 08:41:08.516533 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/07bdec58-1883-4ddb-9aa0-b6eca2498ed4-ceph\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nvsts\" (UID: \"07bdec58-1883-4ddb-9aa0-b6eca2498ed4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nvsts" Nov 22 08:41:08 crc kubenswrapper[4789]: I1122 08:41:08.516574 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07bdec58-1883-4ddb-9aa0-b6eca2498ed4-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nvsts\" (UID: \"07bdec58-1883-4ddb-9aa0-b6eca2498ed4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nvsts" Nov 22 08:41:08 crc kubenswrapper[4789]: I1122 08:41:08.516613 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/07bdec58-1883-4ddb-9aa0-b6eca2498ed4-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nvsts\" (UID: \"07bdec58-1883-4ddb-9aa0-b6eca2498ed4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nvsts" Nov 22 08:41:08 crc kubenswrapper[4789]: I1122 08:41:08.516661 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/07bdec58-1883-4ddb-9aa0-b6eca2498ed4-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nvsts\" (UID: \"07bdec58-1883-4ddb-9aa0-b6eca2498ed4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nvsts" Nov 22 08:41:08 crc kubenswrapper[4789]: I1122 08:41:08.516691 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/07bdec58-1883-4ddb-9aa0-b6eca2498ed4-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nvsts\" (UID: \"07bdec58-1883-4ddb-9aa0-b6eca2498ed4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nvsts" Nov 22 08:41:08 crc kubenswrapper[4789]: I1122 08:41:08.516718 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07bdec58-1883-4ddb-9aa0-b6eca2498ed4-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nvsts\" (UID: \"07bdec58-1883-4ddb-9aa0-b6eca2498ed4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nvsts" Nov 22 08:41:08 crc kubenswrapper[4789]: I1122 08:41:08.516737 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07bdec58-1883-4ddb-9aa0-b6eca2498ed4-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nvsts\" (UID: \"07bdec58-1883-4ddb-9aa0-b6eca2498ed4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nvsts" Nov 22 08:41:08 crc kubenswrapper[4789]: I1122 08:41:08.522224 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07bdec58-1883-4ddb-9aa0-b6eca2498ed4-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nvsts\" (UID: \"07bdec58-1883-4ddb-9aa0-b6eca2498ed4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nvsts" Nov 22 08:41:08 crc kubenswrapper[4789]: I1122 08:41:08.523605 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/07bdec58-1883-4ddb-9aa0-b6eca2498ed4-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nvsts\" (UID: \"07bdec58-1883-4ddb-9aa0-b6eca2498ed4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nvsts" Nov 22 08:41:08 crc kubenswrapper[4789]: I1122 08:41:08.523827 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/07bdec58-1883-4ddb-9aa0-b6eca2498ed4-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nvsts\" (UID: \"07bdec58-1883-4ddb-9aa0-b6eca2498ed4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nvsts" Nov 22 08:41:08 crc kubenswrapper[4789]: I1122 08:41:08.523957 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07bdec58-1883-4ddb-9aa0-b6eca2498ed4-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nvsts\" (UID: \"07bdec58-1883-4ddb-9aa0-b6eca2498ed4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nvsts" Nov 22 08:41:08 crc kubenswrapper[4789]: I1122 08:41:08.524330 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07bdec58-1883-4ddb-9aa0-b6eca2498ed4-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nvsts\" (UID: \"07bdec58-1883-4ddb-9aa0-b6eca2498ed4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nvsts" Nov 22 08:41:08 crc kubenswrapper[4789]: I1122 08:41:08.524759 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/07bdec58-1883-4ddb-9aa0-b6eca2498ed4-ceph\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nvsts\" (UID: \"07bdec58-1883-4ddb-9aa0-b6eca2498ed4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nvsts" Nov 22 08:41:08 crc kubenswrapper[4789]: I1122 08:41:08.525127 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07bdec58-1883-4ddb-9aa0-b6eca2498ed4-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nvsts\" (UID: \"07bdec58-1883-4ddb-9aa0-b6eca2498ed4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nvsts" Nov 22 08:41:08 crc kubenswrapper[4789]: I1122 08:41:08.525183 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/07bdec58-1883-4ddb-9aa0-b6eca2498ed4-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nvsts\" (UID: \"07bdec58-1883-4ddb-9aa0-b6eca2498ed4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nvsts" Nov 22 08:41:08 crc kubenswrapper[4789]: I1122 08:41:08.525830 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/07bdec58-1883-4ddb-9aa0-b6eca2498ed4-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nvsts\" (UID: \"07bdec58-1883-4ddb-9aa0-b6eca2498ed4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nvsts" Nov 22 08:41:08 crc kubenswrapper[4789]: I1122 08:41:08.527093 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07bdec58-1883-4ddb-9aa0-b6eca2498ed4-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nvsts\" (UID: \"07bdec58-1883-4ddb-9aa0-b6eca2498ed4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nvsts" Nov 22 08:41:08 crc kubenswrapper[4789]: I1122 08:41:08.526376 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07bdec58-1883-4ddb-9aa0-b6eca2498ed4-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nvsts\" (UID: \"07bdec58-1883-4ddb-9aa0-b6eca2498ed4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nvsts" Nov 22 08:41:08 crc kubenswrapper[4789]: I1122 08:41:08.527888 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/07bdec58-1883-4ddb-9aa0-b6eca2498ed4-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nvsts\" (UID: \"07bdec58-1883-4ddb-9aa0-b6eca2498ed4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nvsts" Nov 22 08:41:08 crc kubenswrapper[4789]: I1122 08:41:08.536427 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z4wkg\" (UniqueName: \"kubernetes.io/projected/07bdec58-1883-4ddb-9aa0-b6eca2498ed4-kube-api-access-z4wkg\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nvsts\" (UID: \"07bdec58-1883-4ddb-9aa0-b6eca2498ed4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nvsts" Nov 22 08:41:08 crc kubenswrapper[4789]: I1122 08:41:08.547167 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nvsts" Nov 22 08:41:09 crc kubenswrapper[4789]: I1122 08:41:09.038242 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nvsts"] Nov 22 08:41:09 crc kubenswrapper[4789]: I1122 08:41:09.166171 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nvsts" event={"ID":"07bdec58-1883-4ddb-9aa0-b6eca2498ed4","Type":"ContainerStarted","Data":"1e5a7962ee7ad8d55e57fc2fc4eff25c53c86b1c04f7b0411cd6788490905c05"} Nov 22 08:41:10 crc kubenswrapper[4789]: I1122 08:41:10.174207 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nvsts" event={"ID":"07bdec58-1883-4ddb-9aa0-b6eca2498ed4","Type":"ContainerStarted","Data":"2df9e12e5e0c3ae7293d56b4bb07b0a0e5d2ff2f55b390df73f3b88f72216d34"} Nov 22 08:41:10 crc kubenswrapper[4789]: I1122 08:41:10.202214 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nvsts" podStartSLOduration=1.6045108940000001 podStartE2EDuration="2.202190866s" podCreationTimestamp="2025-11-22 08:41:08 +0000 UTC" firstStartedPulling="2025-11-22 08:41:09.042116828 +0000 UTC m=+2683.276517101" lastFinishedPulling="2025-11-22 08:41:09.63979679 +0000 UTC m=+2683.874197073" observedRunningTime="2025-11-22 08:41:10.201282482 +0000 UTC m=+2684.435682755" watchObservedRunningTime="2025-11-22 08:41:10.202190866 +0000 UTC m=+2684.436591159" Nov 22 08:41:35 crc kubenswrapper[4789]: I1122 08:41:35.372143 4789 patch_prober.go:28] interesting pod/machine-config-daemon-pxvrm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 08:41:35 crc kubenswrapper[4789]: I1122 08:41:35.373388 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 08:41:42 crc kubenswrapper[4789]: I1122 08:41:42.436404 4789 generic.go:334] "Generic (PLEG): container finished" podID="07bdec58-1883-4ddb-9aa0-b6eca2498ed4" containerID="2df9e12e5e0c3ae7293d56b4bb07b0a0e5d2ff2f55b390df73f3b88f72216d34" exitCode=0 Nov 22 08:41:42 crc kubenswrapper[4789]: I1122 08:41:42.436525 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nvsts" event={"ID":"07bdec58-1883-4ddb-9aa0-b6eca2498ed4","Type":"ContainerDied","Data":"2df9e12e5e0c3ae7293d56b4bb07b0a0e5d2ff2f55b390df73f3b88f72216d34"} Nov 22 08:41:43 crc kubenswrapper[4789]: I1122 08:41:43.854578 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nvsts" Nov 22 08:41:43 crc kubenswrapper[4789]: I1122 08:41:43.968762 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/07bdec58-1883-4ddb-9aa0-b6eca2498ed4-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"07bdec58-1883-4ddb-9aa0-b6eca2498ed4\" (UID: \"07bdec58-1883-4ddb-9aa0-b6eca2498ed4\") " Nov 22 08:41:43 crc kubenswrapper[4789]: I1122 08:41:43.969102 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07bdec58-1883-4ddb-9aa0-b6eca2498ed4-ovn-combined-ca-bundle\") pod \"07bdec58-1883-4ddb-9aa0-b6eca2498ed4\" (UID: \"07bdec58-1883-4ddb-9aa0-b6eca2498ed4\") " Nov 22 08:41:43 crc kubenswrapper[4789]: I1122 08:41:43.969189 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07bdec58-1883-4ddb-9aa0-b6eca2498ed4-nova-combined-ca-bundle\") pod \"07bdec58-1883-4ddb-9aa0-b6eca2498ed4\" (UID: \"07bdec58-1883-4ddb-9aa0-b6eca2498ed4\") " Nov 22 08:41:43 crc kubenswrapper[4789]: I1122 08:41:43.969238 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/07bdec58-1883-4ddb-9aa0-b6eca2498ed4-ssh-key\") pod \"07bdec58-1883-4ddb-9aa0-b6eca2498ed4\" (UID: \"07bdec58-1883-4ddb-9aa0-b6eca2498ed4\") " Nov 22 08:41:43 crc kubenswrapper[4789]: I1122 08:41:43.969263 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07bdec58-1883-4ddb-9aa0-b6eca2498ed4-neutron-metadata-combined-ca-bundle\") pod \"07bdec58-1883-4ddb-9aa0-b6eca2498ed4\" (UID: \"07bdec58-1883-4ddb-9aa0-b6eca2498ed4\") " Nov 22 08:41:43 crc kubenswrapper[4789]: I1122 08:41:43.969306 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z4wkg\" (UniqueName: \"kubernetes.io/projected/07bdec58-1883-4ddb-9aa0-b6eca2498ed4-kube-api-access-z4wkg\") pod \"07bdec58-1883-4ddb-9aa0-b6eca2498ed4\" (UID: \"07bdec58-1883-4ddb-9aa0-b6eca2498ed4\") " Nov 22 08:41:43 crc kubenswrapper[4789]: I1122 08:41:43.969327 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/07bdec58-1883-4ddb-9aa0-b6eca2498ed4-openstack-edpm-ipam-ovn-default-certs-0\") pod \"07bdec58-1883-4ddb-9aa0-b6eca2498ed4\" (UID: \"07bdec58-1883-4ddb-9aa0-b6eca2498ed4\") " Nov 22 08:41:43 crc kubenswrapper[4789]: I1122 08:41:43.969360 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07bdec58-1883-4ddb-9aa0-b6eca2498ed4-libvirt-combined-ca-bundle\") pod \"07bdec58-1883-4ddb-9aa0-b6eca2498ed4\" (UID: \"07bdec58-1883-4ddb-9aa0-b6eca2498ed4\") " Nov 22 08:41:43 crc kubenswrapper[4789]: I1122 08:41:43.969385 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/07bdec58-1883-4ddb-9aa0-b6eca2498ed4-inventory\") pod \"07bdec58-1883-4ddb-9aa0-b6eca2498ed4\" (UID: \"07bdec58-1883-4ddb-9aa0-b6eca2498ed4\") " Nov 22 08:41:43 crc kubenswrapper[4789]: I1122 08:41:43.969414 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/07bdec58-1883-4ddb-9aa0-b6eca2498ed4-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"07bdec58-1883-4ddb-9aa0-b6eca2498ed4\" (UID: \"07bdec58-1883-4ddb-9aa0-b6eca2498ed4\") " Nov 22 08:41:43 crc kubenswrapper[4789]: I1122 08:41:43.969457 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/07bdec58-1883-4ddb-9aa0-b6eca2498ed4-ceph\") pod \"07bdec58-1883-4ddb-9aa0-b6eca2498ed4\" (UID: \"07bdec58-1883-4ddb-9aa0-b6eca2498ed4\") " Nov 22 08:41:43 crc kubenswrapper[4789]: I1122 08:41:43.969478 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07bdec58-1883-4ddb-9aa0-b6eca2498ed4-bootstrap-combined-ca-bundle\") pod \"07bdec58-1883-4ddb-9aa0-b6eca2498ed4\" (UID: \"07bdec58-1883-4ddb-9aa0-b6eca2498ed4\") " Nov 22 08:41:43 crc kubenswrapper[4789]: I1122 08:41:43.969527 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07bdec58-1883-4ddb-9aa0-b6eca2498ed4-repo-setup-combined-ca-bundle\") pod \"07bdec58-1883-4ddb-9aa0-b6eca2498ed4\" (UID: \"07bdec58-1883-4ddb-9aa0-b6eca2498ed4\") " Nov 22 08:41:43 crc kubenswrapper[4789]: I1122 08:41:43.975506 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/07bdec58-1883-4ddb-9aa0-b6eca2498ed4-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "07bdec58-1883-4ddb-9aa0-b6eca2498ed4" (UID: "07bdec58-1883-4ddb-9aa0-b6eca2498ed4"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:41:43 crc kubenswrapper[4789]: I1122 08:41:43.976189 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/07bdec58-1883-4ddb-9aa0-b6eca2498ed4-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "07bdec58-1883-4ddb-9aa0-b6eca2498ed4" (UID: "07bdec58-1883-4ddb-9aa0-b6eca2498ed4"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:41:43 crc kubenswrapper[4789]: I1122 08:41:43.976422 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/07bdec58-1883-4ddb-9aa0-b6eca2498ed4-ceph" (OuterVolumeSpecName: "ceph") pod "07bdec58-1883-4ddb-9aa0-b6eca2498ed4" (UID: "07bdec58-1883-4ddb-9aa0-b6eca2498ed4"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:41:43 crc kubenswrapper[4789]: I1122 08:41:43.976496 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/07bdec58-1883-4ddb-9aa0-b6eca2498ed4-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "07bdec58-1883-4ddb-9aa0-b6eca2498ed4" (UID: "07bdec58-1883-4ddb-9aa0-b6eca2498ed4"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:41:43 crc kubenswrapper[4789]: I1122 08:41:43.976714 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/07bdec58-1883-4ddb-9aa0-b6eca2498ed4-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "07bdec58-1883-4ddb-9aa0-b6eca2498ed4" (UID: "07bdec58-1883-4ddb-9aa0-b6eca2498ed4"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:41:43 crc kubenswrapper[4789]: I1122 08:41:43.977585 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/07bdec58-1883-4ddb-9aa0-b6eca2498ed4-kube-api-access-z4wkg" (OuterVolumeSpecName: "kube-api-access-z4wkg") pod "07bdec58-1883-4ddb-9aa0-b6eca2498ed4" (UID: "07bdec58-1883-4ddb-9aa0-b6eca2498ed4"). InnerVolumeSpecName "kube-api-access-z4wkg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:41:43 crc kubenswrapper[4789]: I1122 08:41:43.977578 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/07bdec58-1883-4ddb-9aa0-b6eca2498ed4-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "07bdec58-1883-4ddb-9aa0-b6eca2498ed4" (UID: "07bdec58-1883-4ddb-9aa0-b6eca2498ed4"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:41:43 crc kubenswrapper[4789]: I1122 08:41:43.977708 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/07bdec58-1883-4ddb-9aa0-b6eca2498ed4-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "07bdec58-1883-4ddb-9aa0-b6eca2498ed4" (UID: "07bdec58-1883-4ddb-9aa0-b6eca2498ed4"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:41:43 crc kubenswrapper[4789]: I1122 08:41:43.978322 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/07bdec58-1883-4ddb-9aa0-b6eca2498ed4-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "07bdec58-1883-4ddb-9aa0-b6eca2498ed4" (UID: "07bdec58-1883-4ddb-9aa0-b6eca2498ed4"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:41:43 crc kubenswrapper[4789]: I1122 08:41:43.980394 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/07bdec58-1883-4ddb-9aa0-b6eca2498ed4-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "07bdec58-1883-4ddb-9aa0-b6eca2498ed4" (UID: "07bdec58-1883-4ddb-9aa0-b6eca2498ed4"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:41:43 crc kubenswrapper[4789]: I1122 08:41:43.981623 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/07bdec58-1883-4ddb-9aa0-b6eca2498ed4-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "07bdec58-1883-4ddb-9aa0-b6eca2498ed4" (UID: "07bdec58-1883-4ddb-9aa0-b6eca2498ed4"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:41:44 crc kubenswrapper[4789]: I1122 08:41:44.001203 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/07bdec58-1883-4ddb-9aa0-b6eca2498ed4-inventory" (OuterVolumeSpecName: "inventory") pod "07bdec58-1883-4ddb-9aa0-b6eca2498ed4" (UID: "07bdec58-1883-4ddb-9aa0-b6eca2498ed4"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:41:44 crc kubenswrapper[4789]: I1122 08:41:44.003946 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/07bdec58-1883-4ddb-9aa0-b6eca2498ed4-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "07bdec58-1883-4ddb-9aa0-b6eca2498ed4" (UID: "07bdec58-1883-4ddb-9aa0-b6eca2498ed4"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:41:44 crc kubenswrapper[4789]: I1122 08:41:44.072138 4789 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/07bdec58-1883-4ddb-9aa0-b6eca2498ed4-inventory\") on node \"crc\" DevicePath \"\"" Nov 22 08:41:44 crc kubenswrapper[4789]: I1122 08:41:44.072179 4789 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/07bdec58-1883-4ddb-9aa0-b6eca2498ed4-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 22 08:41:44 crc kubenswrapper[4789]: I1122 08:41:44.072191 4789 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/07bdec58-1883-4ddb-9aa0-b6eca2498ed4-ceph\") on node \"crc\" DevicePath \"\"" Nov 22 08:41:44 crc kubenswrapper[4789]: I1122 08:41:44.072200 4789 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07bdec58-1883-4ddb-9aa0-b6eca2498ed4-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:41:44 crc kubenswrapper[4789]: I1122 08:41:44.072212 4789 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07bdec58-1883-4ddb-9aa0-b6eca2498ed4-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:41:44 crc kubenswrapper[4789]: I1122 08:41:44.072235 4789 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/07bdec58-1883-4ddb-9aa0-b6eca2498ed4-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 22 08:41:44 crc kubenswrapper[4789]: I1122 08:41:44.072244 4789 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07bdec58-1883-4ddb-9aa0-b6eca2498ed4-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:41:44 crc kubenswrapper[4789]: I1122 08:41:44.072253 4789 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07bdec58-1883-4ddb-9aa0-b6eca2498ed4-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:41:44 crc kubenswrapper[4789]: I1122 08:41:44.072261 4789 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/07bdec58-1883-4ddb-9aa0-b6eca2498ed4-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 22 08:41:44 crc kubenswrapper[4789]: I1122 08:41:44.072269 4789 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07bdec58-1883-4ddb-9aa0-b6eca2498ed4-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:41:44 crc kubenswrapper[4789]: I1122 08:41:44.072278 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z4wkg\" (UniqueName: \"kubernetes.io/projected/07bdec58-1883-4ddb-9aa0-b6eca2498ed4-kube-api-access-z4wkg\") on node \"crc\" DevicePath \"\"" Nov 22 08:41:44 crc kubenswrapper[4789]: I1122 08:41:44.072291 4789 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/07bdec58-1883-4ddb-9aa0-b6eca2498ed4-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 22 08:41:44 crc kubenswrapper[4789]: I1122 08:41:44.072302 4789 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07bdec58-1883-4ddb-9aa0-b6eca2498ed4-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:41:44 crc kubenswrapper[4789]: I1122 08:41:44.455063 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nvsts" event={"ID":"07bdec58-1883-4ddb-9aa0-b6eca2498ed4","Type":"ContainerDied","Data":"1e5a7962ee7ad8d55e57fc2fc4eff25c53c86b1c04f7b0411cd6788490905c05"} Nov 22 08:41:44 crc kubenswrapper[4789]: I1122 08:41:44.455099 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nvsts" Nov 22 08:41:44 crc kubenswrapper[4789]: I1122 08:41:44.455124 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1e5a7962ee7ad8d55e57fc2fc4eff25c53c86b1c04f7b0411cd6788490905c05" Nov 22 08:41:44 crc kubenswrapper[4789]: I1122 08:41:44.575862 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-k4r2f"] Nov 22 08:41:44 crc kubenswrapper[4789]: E1122 08:41:44.576245 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07bdec58-1883-4ddb-9aa0-b6eca2498ed4" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Nov 22 08:41:44 crc kubenswrapper[4789]: I1122 08:41:44.576262 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="07bdec58-1883-4ddb-9aa0-b6eca2498ed4" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Nov 22 08:41:44 crc kubenswrapper[4789]: I1122 08:41:44.576431 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="07bdec58-1883-4ddb-9aa0-b6eca2498ed4" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Nov 22 08:41:44 crc kubenswrapper[4789]: I1122 08:41:44.577050 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-k4r2f" Nov 22 08:41:44 crc kubenswrapper[4789]: I1122 08:41:44.579316 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 22 08:41:44 crc kubenswrapper[4789]: I1122 08:41:44.579367 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 22 08:41:44 crc kubenswrapper[4789]: I1122 08:41:44.580190 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 22 08:41:44 crc kubenswrapper[4789]: I1122 08:41:44.580218 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 22 08:41:44 crc kubenswrapper[4789]: I1122 08:41:44.580347 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-glmr7" Nov 22 08:41:44 crc kubenswrapper[4789]: I1122 08:41:44.596643 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-k4r2f"] Nov 22 08:41:44 crc kubenswrapper[4789]: I1122 08:41:44.683055 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f8d83872-bd0a-4fad-ae6d-3bfd1c175caa-ceph\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-k4r2f\" (UID: \"f8d83872-bd0a-4fad-ae6d-3bfd1c175caa\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-k4r2f" Nov 22 08:41:44 crc kubenswrapper[4789]: I1122 08:41:44.683161 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f8d83872-bd0a-4fad-ae6d-3bfd1c175caa-inventory\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-k4r2f\" (UID: \"f8d83872-bd0a-4fad-ae6d-3bfd1c175caa\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-k4r2f" Nov 22 08:41:44 crc kubenswrapper[4789]: I1122 08:41:44.683288 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d28zp\" (UniqueName: \"kubernetes.io/projected/f8d83872-bd0a-4fad-ae6d-3bfd1c175caa-kube-api-access-d28zp\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-k4r2f\" (UID: \"f8d83872-bd0a-4fad-ae6d-3bfd1c175caa\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-k4r2f" Nov 22 08:41:44 crc kubenswrapper[4789]: I1122 08:41:44.683414 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f8d83872-bd0a-4fad-ae6d-3bfd1c175caa-ssh-key\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-k4r2f\" (UID: \"f8d83872-bd0a-4fad-ae6d-3bfd1c175caa\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-k4r2f" Nov 22 08:41:44 crc kubenswrapper[4789]: I1122 08:41:44.785312 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d28zp\" (UniqueName: \"kubernetes.io/projected/f8d83872-bd0a-4fad-ae6d-3bfd1c175caa-kube-api-access-d28zp\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-k4r2f\" (UID: \"f8d83872-bd0a-4fad-ae6d-3bfd1c175caa\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-k4r2f" Nov 22 08:41:44 crc kubenswrapper[4789]: I1122 08:41:44.785450 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f8d83872-bd0a-4fad-ae6d-3bfd1c175caa-ssh-key\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-k4r2f\" (UID: \"f8d83872-bd0a-4fad-ae6d-3bfd1c175caa\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-k4r2f" Nov 22 08:41:44 crc kubenswrapper[4789]: I1122 08:41:44.785528 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f8d83872-bd0a-4fad-ae6d-3bfd1c175caa-ceph\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-k4r2f\" (UID: \"f8d83872-bd0a-4fad-ae6d-3bfd1c175caa\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-k4r2f" Nov 22 08:41:44 crc kubenswrapper[4789]: I1122 08:41:44.785627 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f8d83872-bd0a-4fad-ae6d-3bfd1c175caa-inventory\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-k4r2f\" (UID: \"f8d83872-bd0a-4fad-ae6d-3bfd1c175caa\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-k4r2f" Nov 22 08:41:44 crc kubenswrapper[4789]: I1122 08:41:44.789398 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f8d83872-bd0a-4fad-ae6d-3bfd1c175caa-ssh-key\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-k4r2f\" (UID: \"f8d83872-bd0a-4fad-ae6d-3bfd1c175caa\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-k4r2f" Nov 22 08:41:44 crc kubenswrapper[4789]: I1122 08:41:44.790638 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f8d83872-bd0a-4fad-ae6d-3bfd1c175caa-ceph\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-k4r2f\" (UID: \"f8d83872-bd0a-4fad-ae6d-3bfd1c175caa\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-k4r2f" Nov 22 08:41:44 crc kubenswrapper[4789]: I1122 08:41:44.791415 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f8d83872-bd0a-4fad-ae6d-3bfd1c175caa-inventory\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-k4r2f\" (UID: \"f8d83872-bd0a-4fad-ae6d-3bfd1c175caa\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-k4r2f" Nov 22 08:41:44 crc kubenswrapper[4789]: I1122 08:41:44.804685 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d28zp\" (UniqueName: \"kubernetes.io/projected/f8d83872-bd0a-4fad-ae6d-3bfd1c175caa-kube-api-access-d28zp\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-k4r2f\" (UID: \"f8d83872-bd0a-4fad-ae6d-3bfd1c175caa\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-k4r2f" Nov 22 08:41:44 crc kubenswrapper[4789]: I1122 08:41:44.895444 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-k4r2f" Nov 22 08:41:45 crc kubenswrapper[4789]: I1122 08:41:45.413702 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-k4r2f"] Nov 22 08:41:45 crc kubenswrapper[4789]: I1122 08:41:45.417881 4789 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 22 08:41:45 crc kubenswrapper[4789]: I1122 08:41:45.466014 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-k4r2f" event={"ID":"f8d83872-bd0a-4fad-ae6d-3bfd1c175caa","Type":"ContainerStarted","Data":"bad3a78fff611e0a67da5c1b0e36d461b188b691d1965095b0bbfbd35fe2c9bc"} Nov 22 08:41:47 crc kubenswrapper[4789]: I1122 08:41:47.486199 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-k4r2f" event={"ID":"f8d83872-bd0a-4fad-ae6d-3bfd1c175caa","Type":"ContainerStarted","Data":"ff3f86183b1ccd1504f8f902797c242d756e665c1377ccb5bfd881b0d7f17d36"} Nov 22 08:41:47 crc kubenswrapper[4789]: I1122 08:41:47.504824 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-nncxd"] Nov 22 08:41:47 crc kubenswrapper[4789]: I1122 08:41:47.518426 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nncxd" Nov 22 08:41:47 crc kubenswrapper[4789]: I1122 08:41:47.532118 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nncxd"] Nov 22 08:41:47 crc kubenswrapper[4789]: I1122 08:41:47.543864 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-k4r2f" podStartSLOduration=2.789946575 podStartE2EDuration="3.543841222s" podCreationTimestamp="2025-11-22 08:41:44 +0000 UTC" firstStartedPulling="2025-11-22 08:41:45.417605962 +0000 UTC m=+2719.652006235" lastFinishedPulling="2025-11-22 08:41:46.171500609 +0000 UTC m=+2720.405900882" observedRunningTime="2025-11-22 08:41:47.514466456 +0000 UTC m=+2721.748866729" watchObservedRunningTime="2025-11-22 08:41:47.543841222 +0000 UTC m=+2721.778241495" Nov 22 08:41:47 crc kubenswrapper[4789]: I1122 08:41:47.642962 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d8a1533f-6b43-4ecb-b9b4-702841b049b5-utilities\") pod \"community-operators-nncxd\" (UID: \"d8a1533f-6b43-4ecb-b9b4-702841b049b5\") " pod="openshift-marketplace/community-operators-nncxd" Nov 22 08:41:47 crc kubenswrapper[4789]: I1122 08:41:47.643836 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4x4db\" (UniqueName: \"kubernetes.io/projected/d8a1533f-6b43-4ecb-b9b4-702841b049b5-kube-api-access-4x4db\") pod \"community-operators-nncxd\" (UID: \"d8a1533f-6b43-4ecb-b9b4-702841b049b5\") " pod="openshift-marketplace/community-operators-nncxd" Nov 22 08:41:47 crc kubenswrapper[4789]: I1122 08:41:47.644131 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d8a1533f-6b43-4ecb-b9b4-702841b049b5-catalog-content\") pod \"community-operators-nncxd\" (UID: \"d8a1533f-6b43-4ecb-b9b4-702841b049b5\") " pod="openshift-marketplace/community-operators-nncxd" Nov 22 08:41:47 crc kubenswrapper[4789]: I1122 08:41:47.745407 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4x4db\" (UniqueName: \"kubernetes.io/projected/d8a1533f-6b43-4ecb-b9b4-702841b049b5-kube-api-access-4x4db\") pod \"community-operators-nncxd\" (UID: \"d8a1533f-6b43-4ecb-b9b4-702841b049b5\") " pod="openshift-marketplace/community-operators-nncxd" Nov 22 08:41:47 crc kubenswrapper[4789]: I1122 08:41:47.745521 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d8a1533f-6b43-4ecb-b9b4-702841b049b5-catalog-content\") pod \"community-operators-nncxd\" (UID: \"d8a1533f-6b43-4ecb-b9b4-702841b049b5\") " pod="openshift-marketplace/community-operators-nncxd" Nov 22 08:41:47 crc kubenswrapper[4789]: I1122 08:41:47.745607 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d8a1533f-6b43-4ecb-b9b4-702841b049b5-utilities\") pod \"community-operators-nncxd\" (UID: \"d8a1533f-6b43-4ecb-b9b4-702841b049b5\") " pod="openshift-marketplace/community-operators-nncxd" Nov 22 08:41:47 crc kubenswrapper[4789]: I1122 08:41:47.746018 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d8a1533f-6b43-4ecb-b9b4-702841b049b5-catalog-content\") pod \"community-operators-nncxd\" (UID: \"d8a1533f-6b43-4ecb-b9b4-702841b049b5\") " pod="openshift-marketplace/community-operators-nncxd" Nov 22 08:41:47 crc kubenswrapper[4789]: I1122 08:41:47.746125 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d8a1533f-6b43-4ecb-b9b4-702841b049b5-utilities\") pod \"community-operators-nncxd\" (UID: \"d8a1533f-6b43-4ecb-b9b4-702841b049b5\") " pod="openshift-marketplace/community-operators-nncxd" Nov 22 08:41:47 crc kubenswrapper[4789]: I1122 08:41:47.776632 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4x4db\" (UniqueName: \"kubernetes.io/projected/d8a1533f-6b43-4ecb-b9b4-702841b049b5-kube-api-access-4x4db\") pod \"community-operators-nncxd\" (UID: \"d8a1533f-6b43-4ecb-b9b4-702841b049b5\") " pod="openshift-marketplace/community-operators-nncxd" Nov 22 08:41:47 crc kubenswrapper[4789]: I1122 08:41:47.844381 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nncxd" Nov 22 08:41:48 crc kubenswrapper[4789]: I1122 08:41:48.303833 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nncxd"] Nov 22 08:41:48 crc kubenswrapper[4789]: I1122 08:41:48.496883 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nncxd" event={"ID":"d8a1533f-6b43-4ecb-b9b4-702841b049b5","Type":"ContainerStarted","Data":"f3fbcd3629404e976f09e77da2de258462f3a2a5e5ce455a15d4bd2695323539"} Nov 22 08:41:48 crc kubenswrapper[4789]: I1122 08:41:48.496940 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nncxd" event={"ID":"d8a1533f-6b43-4ecb-b9b4-702841b049b5","Type":"ContainerStarted","Data":"1665c2619ff95ad6f88dfca6b744b876b2e63e55b4a002423323543f679ee646"} Nov 22 08:41:49 crc kubenswrapper[4789]: I1122 08:41:49.505018 4789 generic.go:334] "Generic (PLEG): container finished" podID="d8a1533f-6b43-4ecb-b9b4-702841b049b5" containerID="f3fbcd3629404e976f09e77da2de258462f3a2a5e5ce455a15d4bd2695323539" exitCode=0 Nov 22 08:41:49 crc kubenswrapper[4789]: I1122 08:41:49.505061 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nncxd" event={"ID":"d8a1533f-6b43-4ecb-b9b4-702841b049b5","Type":"ContainerDied","Data":"f3fbcd3629404e976f09e77da2de258462f3a2a5e5ce455a15d4bd2695323539"} Nov 22 08:41:51 crc kubenswrapper[4789]: I1122 08:41:51.529218 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nncxd" event={"ID":"d8a1533f-6b43-4ecb-b9b4-702841b049b5","Type":"ContainerStarted","Data":"7b9db75b1392663b0f41f18662e717f5ed881f51a9af9ab09caefd3d84a821ba"} Nov 22 08:41:52 crc kubenswrapper[4789]: I1122 08:41:52.542203 4789 generic.go:334] "Generic (PLEG): container finished" podID="d8a1533f-6b43-4ecb-b9b4-702841b049b5" containerID="7b9db75b1392663b0f41f18662e717f5ed881f51a9af9ab09caefd3d84a821ba" exitCode=0 Nov 22 08:41:52 crc kubenswrapper[4789]: I1122 08:41:52.542271 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nncxd" event={"ID":"d8a1533f-6b43-4ecb-b9b4-702841b049b5","Type":"ContainerDied","Data":"7b9db75b1392663b0f41f18662e717f5ed881f51a9af9ab09caefd3d84a821ba"} Nov 22 08:41:52 crc kubenswrapper[4789]: I1122 08:41:52.546611 4789 generic.go:334] "Generic (PLEG): container finished" podID="f8d83872-bd0a-4fad-ae6d-3bfd1c175caa" containerID="ff3f86183b1ccd1504f8f902797c242d756e665c1377ccb5bfd881b0d7f17d36" exitCode=0 Nov 22 08:41:52 crc kubenswrapper[4789]: I1122 08:41:52.546650 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-k4r2f" event={"ID":"f8d83872-bd0a-4fad-ae6d-3bfd1c175caa","Type":"ContainerDied","Data":"ff3f86183b1ccd1504f8f902797c242d756e665c1377ccb5bfd881b0d7f17d36"} Nov 22 08:41:53 crc kubenswrapper[4789]: I1122 08:41:53.559921 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nncxd" event={"ID":"d8a1533f-6b43-4ecb-b9b4-702841b049b5","Type":"ContainerStarted","Data":"6f2db7cbfeea46b223a1e800eb8f3c082b6cf399c41c4bcd2d5365a5f247f648"} Nov 22 08:41:53 crc kubenswrapper[4789]: I1122 08:41:53.583409 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-nncxd" podStartSLOduration=3.09045741 podStartE2EDuration="6.583382949s" podCreationTimestamp="2025-11-22 08:41:47 +0000 UTC" firstStartedPulling="2025-11-22 08:41:49.507141416 +0000 UTC m=+2723.741541679" lastFinishedPulling="2025-11-22 08:41:53.000066925 +0000 UTC m=+2727.234467218" observedRunningTime="2025-11-22 08:41:53.578922577 +0000 UTC m=+2727.813322850" watchObservedRunningTime="2025-11-22 08:41:53.583382949 +0000 UTC m=+2727.817783222" Nov 22 08:41:53 crc kubenswrapper[4789]: I1122 08:41:53.931031 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-k4r2f" Nov 22 08:41:54 crc kubenswrapper[4789]: I1122 08:41:54.062161 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d28zp\" (UniqueName: \"kubernetes.io/projected/f8d83872-bd0a-4fad-ae6d-3bfd1c175caa-kube-api-access-d28zp\") pod \"f8d83872-bd0a-4fad-ae6d-3bfd1c175caa\" (UID: \"f8d83872-bd0a-4fad-ae6d-3bfd1c175caa\") " Nov 22 08:41:54 crc kubenswrapper[4789]: I1122 08:41:54.062326 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f8d83872-bd0a-4fad-ae6d-3bfd1c175caa-ceph\") pod \"f8d83872-bd0a-4fad-ae6d-3bfd1c175caa\" (UID: \"f8d83872-bd0a-4fad-ae6d-3bfd1c175caa\") " Nov 22 08:41:54 crc kubenswrapper[4789]: I1122 08:41:54.062355 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f8d83872-bd0a-4fad-ae6d-3bfd1c175caa-inventory\") pod \"f8d83872-bd0a-4fad-ae6d-3bfd1c175caa\" (UID: \"f8d83872-bd0a-4fad-ae6d-3bfd1c175caa\") " Nov 22 08:41:54 crc kubenswrapper[4789]: I1122 08:41:54.062537 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f8d83872-bd0a-4fad-ae6d-3bfd1c175caa-ssh-key\") pod \"f8d83872-bd0a-4fad-ae6d-3bfd1c175caa\" (UID: \"f8d83872-bd0a-4fad-ae6d-3bfd1c175caa\") " Nov 22 08:41:54 crc kubenswrapper[4789]: I1122 08:41:54.068470 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8d83872-bd0a-4fad-ae6d-3bfd1c175caa-ceph" (OuterVolumeSpecName: "ceph") pod "f8d83872-bd0a-4fad-ae6d-3bfd1c175caa" (UID: "f8d83872-bd0a-4fad-ae6d-3bfd1c175caa"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:41:54 crc kubenswrapper[4789]: I1122 08:41:54.073488 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f8d83872-bd0a-4fad-ae6d-3bfd1c175caa-kube-api-access-d28zp" (OuterVolumeSpecName: "kube-api-access-d28zp") pod "f8d83872-bd0a-4fad-ae6d-3bfd1c175caa" (UID: "f8d83872-bd0a-4fad-ae6d-3bfd1c175caa"). InnerVolumeSpecName "kube-api-access-d28zp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:41:54 crc kubenswrapper[4789]: E1122 08:41:54.086818 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f8d83872-bd0a-4fad-ae6d-3bfd1c175caa-inventory podName:f8d83872-bd0a-4fad-ae6d-3bfd1c175caa nodeName:}" failed. No retries permitted until 2025-11-22 08:41:54.586773165 +0000 UTC m=+2728.821173438 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "inventory" (UniqueName: "kubernetes.io/secret/f8d83872-bd0a-4fad-ae6d-3bfd1c175caa-inventory") pod "f8d83872-bd0a-4fad-ae6d-3bfd1c175caa" (UID: "f8d83872-bd0a-4fad-ae6d-3bfd1c175caa") : error deleting /var/lib/kubelet/pods/f8d83872-bd0a-4fad-ae6d-3bfd1c175caa/volume-subpaths: remove /var/lib/kubelet/pods/f8d83872-bd0a-4fad-ae6d-3bfd1c175caa/volume-subpaths: no such file or directory Nov 22 08:41:54 crc kubenswrapper[4789]: I1122 08:41:54.089570 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8d83872-bd0a-4fad-ae6d-3bfd1c175caa-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f8d83872-bd0a-4fad-ae6d-3bfd1c175caa" (UID: "f8d83872-bd0a-4fad-ae6d-3bfd1c175caa"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:41:54 crc kubenswrapper[4789]: I1122 08:41:54.166121 4789 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f8d83872-bd0a-4fad-ae6d-3bfd1c175caa-ceph\") on node \"crc\" DevicePath \"\"" Nov 22 08:41:54 crc kubenswrapper[4789]: I1122 08:41:54.166482 4789 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f8d83872-bd0a-4fad-ae6d-3bfd1c175caa-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 22 08:41:54 crc kubenswrapper[4789]: I1122 08:41:54.166500 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d28zp\" (UniqueName: \"kubernetes.io/projected/f8d83872-bd0a-4fad-ae6d-3bfd1c175caa-kube-api-access-d28zp\") on node \"crc\" DevicePath \"\"" Nov 22 08:41:54 crc kubenswrapper[4789]: I1122 08:41:54.571186 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-k4r2f" event={"ID":"f8d83872-bd0a-4fad-ae6d-3bfd1c175caa","Type":"ContainerDied","Data":"bad3a78fff611e0a67da5c1b0e36d461b188b691d1965095b0bbfbd35fe2c9bc"} Nov 22 08:41:54 crc kubenswrapper[4789]: I1122 08:41:54.572130 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bad3a78fff611e0a67da5c1b0e36d461b188b691d1965095b0bbfbd35fe2c9bc" Nov 22 08:41:54 crc kubenswrapper[4789]: I1122 08:41:54.571226 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-k4r2f" Nov 22 08:41:54 crc kubenswrapper[4789]: I1122 08:41:54.665943 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-tqgkq"] Nov 22 08:41:54 crc kubenswrapper[4789]: E1122 08:41:54.666364 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8d83872-bd0a-4fad-ae6d-3bfd1c175caa" containerName="ceph-client-edpm-deployment-openstack-edpm-ipam" Nov 22 08:41:54 crc kubenswrapper[4789]: I1122 08:41:54.666385 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8d83872-bd0a-4fad-ae6d-3bfd1c175caa" containerName="ceph-client-edpm-deployment-openstack-edpm-ipam" Nov 22 08:41:54 crc kubenswrapper[4789]: I1122 08:41:54.666594 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8d83872-bd0a-4fad-ae6d-3bfd1c175caa" containerName="ceph-client-edpm-deployment-openstack-edpm-ipam" Nov 22 08:41:54 crc kubenswrapper[4789]: I1122 08:41:54.667339 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-tqgkq" Nov 22 08:41:54 crc kubenswrapper[4789]: I1122 08:41:54.669660 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Nov 22 08:41:54 crc kubenswrapper[4789]: I1122 08:41:54.685088 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-tqgkq"] Nov 22 08:41:54 crc kubenswrapper[4789]: I1122 08:41:54.689449 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f8d83872-bd0a-4fad-ae6d-3bfd1c175caa-inventory\") pod \"f8d83872-bd0a-4fad-ae6d-3bfd1c175caa\" (UID: \"f8d83872-bd0a-4fad-ae6d-3bfd1c175caa\") " Nov 22 08:41:54 crc kubenswrapper[4789]: I1122 08:41:54.707487 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8d83872-bd0a-4fad-ae6d-3bfd1c175caa-inventory" (OuterVolumeSpecName: "inventory") pod "f8d83872-bd0a-4fad-ae6d-3bfd1c175caa" (UID: "f8d83872-bd0a-4fad-ae6d-3bfd1c175caa"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:41:54 crc kubenswrapper[4789]: I1122 08:41:54.792863 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/afd5b574-89f9-4b13-a3ee-b5e4c39095ff-ceph\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-tqgkq\" (UID: \"afd5b574-89f9-4b13-a3ee-b5e4c39095ff\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-tqgkq" Nov 22 08:41:54 crc kubenswrapper[4789]: I1122 08:41:54.792950 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/afd5b574-89f9-4b13-a3ee-b5e4c39095ff-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-tqgkq\" (UID: \"afd5b574-89f9-4b13-a3ee-b5e4c39095ff\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-tqgkq" Nov 22 08:41:54 crc kubenswrapper[4789]: I1122 08:41:54.793032 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qm9jz\" (UniqueName: \"kubernetes.io/projected/afd5b574-89f9-4b13-a3ee-b5e4c39095ff-kube-api-access-qm9jz\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-tqgkq\" (UID: \"afd5b574-89f9-4b13-a3ee-b5e4c39095ff\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-tqgkq" Nov 22 08:41:54 crc kubenswrapper[4789]: I1122 08:41:54.793093 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/afd5b574-89f9-4b13-a3ee-b5e4c39095ff-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-tqgkq\" (UID: \"afd5b574-89f9-4b13-a3ee-b5e4c39095ff\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-tqgkq" Nov 22 08:41:54 crc kubenswrapper[4789]: I1122 08:41:54.793113 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/afd5b574-89f9-4b13-a3ee-b5e4c39095ff-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-tqgkq\" (UID: \"afd5b574-89f9-4b13-a3ee-b5e4c39095ff\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-tqgkq" Nov 22 08:41:54 crc kubenswrapper[4789]: I1122 08:41:54.793167 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/afd5b574-89f9-4b13-a3ee-b5e4c39095ff-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-tqgkq\" (UID: \"afd5b574-89f9-4b13-a3ee-b5e4c39095ff\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-tqgkq" Nov 22 08:41:54 crc kubenswrapper[4789]: I1122 08:41:54.793249 4789 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f8d83872-bd0a-4fad-ae6d-3bfd1c175caa-inventory\") on node \"crc\" DevicePath \"\"" Nov 22 08:41:54 crc kubenswrapper[4789]: I1122 08:41:54.894609 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/afd5b574-89f9-4b13-a3ee-b5e4c39095ff-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-tqgkq\" (UID: \"afd5b574-89f9-4b13-a3ee-b5e4c39095ff\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-tqgkq" Nov 22 08:41:54 crc kubenswrapper[4789]: I1122 08:41:54.894649 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/afd5b574-89f9-4b13-a3ee-b5e4c39095ff-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-tqgkq\" (UID: \"afd5b574-89f9-4b13-a3ee-b5e4c39095ff\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-tqgkq" Nov 22 08:41:54 crc kubenswrapper[4789]: I1122 08:41:54.894687 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/afd5b574-89f9-4b13-a3ee-b5e4c39095ff-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-tqgkq\" (UID: \"afd5b574-89f9-4b13-a3ee-b5e4c39095ff\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-tqgkq" Nov 22 08:41:54 crc kubenswrapper[4789]: I1122 08:41:54.894782 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/afd5b574-89f9-4b13-a3ee-b5e4c39095ff-ceph\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-tqgkq\" (UID: \"afd5b574-89f9-4b13-a3ee-b5e4c39095ff\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-tqgkq" Nov 22 08:41:54 crc kubenswrapper[4789]: I1122 08:41:54.894815 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/afd5b574-89f9-4b13-a3ee-b5e4c39095ff-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-tqgkq\" (UID: \"afd5b574-89f9-4b13-a3ee-b5e4c39095ff\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-tqgkq" Nov 22 08:41:54 crc kubenswrapper[4789]: I1122 08:41:54.894846 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qm9jz\" (UniqueName: \"kubernetes.io/projected/afd5b574-89f9-4b13-a3ee-b5e4c39095ff-kube-api-access-qm9jz\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-tqgkq\" (UID: \"afd5b574-89f9-4b13-a3ee-b5e4c39095ff\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-tqgkq" Nov 22 08:41:54 crc kubenswrapper[4789]: I1122 08:41:54.895554 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/afd5b574-89f9-4b13-a3ee-b5e4c39095ff-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-tqgkq\" (UID: \"afd5b574-89f9-4b13-a3ee-b5e4c39095ff\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-tqgkq" Nov 22 08:41:54 crc kubenswrapper[4789]: I1122 08:41:54.899500 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/afd5b574-89f9-4b13-a3ee-b5e4c39095ff-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-tqgkq\" (UID: \"afd5b574-89f9-4b13-a3ee-b5e4c39095ff\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-tqgkq" Nov 22 08:41:54 crc kubenswrapper[4789]: I1122 08:41:54.900378 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/afd5b574-89f9-4b13-a3ee-b5e4c39095ff-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-tqgkq\" (UID: \"afd5b574-89f9-4b13-a3ee-b5e4c39095ff\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-tqgkq" Nov 22 08:41:54 crc kubenswrapper[4789]: I1122 08:41:54.901564 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/afd5b574-89f9-4b13-a3ee-b5e4c39095ff-ceph\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-tqgkq\" (UID: \"afd5b574-89f9-4b13-a3ee-b5e4c39095ff\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-tqgkq" Nov 22 08:41:54 crc kubenswrapper[4789]: I1122 08:41:54.901771 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/afd5b574-89f9-4b13-a3ee-b5e4c39095ff-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-tqgkq\" (UID: \"afd5b574-89f9-4b13-a3ee-b5e4c39095ff\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-tqgkq" Nov 22 08:41:54 crc kubenswrapper[4789]: I1122 08:41:54.912050 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qm9jz\" (UniqueName: \"kubernetes.io/projected/afd5b574-89f9-4b13-a3ee-b5e4c39095ff-kube-api-access-qm9jz\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-tqgkq\" (UID: \"afd5b574-89f9-4b13-a3ee-b5e4c39095ff\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-tqgkq" Nov 22 08:41:54 crc kubenswrapper[4789]: I1122 08:41:54.995170 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-tqgkq" Nov 22 08:41:55 crc kubenswrapper[4789]: I1122 08:41:55.487683 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-tqgkq"] Nov 22 08:41:55 crc kubenswrapper[4789]: W1122 08:41:55.491497 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podafd5b574_89f9_4b13_a3ee_b5e4c39095ff.slice/crio-d0e1df0a7d7a252c254565ef2007458522bce2cf9f348eefd6b7919f3e70d3aa WatchSource:0}: Error finding container d0e1df0a7d7a252c254565ef2007458522bce2cf9f348eefd6b7919f3e70d3aa: Status 404 returned error can't find the container with id d0e1df0a7d7a252c254565ef2007458522bce2cf9f348eefd6b7919f3e70d3aa Nov 22 08:41:55 crc kubenswrapper[4789]: I1122 08:41:55.580849 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-tqgkq" event={"ID":"afd5b574-89f9-4b13-a3ee-b5e4c39095ff","Type":"ContainerStarted","Data":"d0e1df0a7d7a252c254565ef2007458522bce2cf9f348eefd6b7919f3e70d3aa"} Nov 22 08:41:56 crc kubenswrapper[4789]: I1122 08:41:56.590820 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-tqgkq" event={"ID":"afd5b574-89f9-4b13-a3ee-b5e4c39095ff","Type":"ContainerStarted","Data":"48ab9540830de88f3403f93978df93af82fd5058c07dc2a20fd3193c417fc881"} Nov 22 08:41:56 crc kubenswrapper[4789]: I1122 08:41:56.615113 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-tqgkq" podStartSLOduration=2.155569879 podStartE2EDuration="2.615093226s" podCreationTimestamp="2025-11-22 08:41:54 +0000 UTC" firstStartedPulling="2025-11-22 08:41:55.493714066 +0000 UTC m=+2729.728114339" lastFinishedPulling="2025-11-22 08:41:55.953237393 +0000 UTC m=+2730.187637686" observedRunningTime="2025-11-22 08:41:56.610308636 +0000 UTC m=+2730.844708919" watchObservedRunningTime="2025-11-22 08:41:56.615093226 +0000 UTC m=+2730.849493509" Nov 22 08:41:57 crc kubenswrapper[4789]: I1122 08:41:57.844892 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-nncxd" Nov 22 08:41:57 crc kubenswrapper[4789]: I1122 08:41:57.845041 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-nncxd" Nov 22 08:41:57 crc kubenswrapper[4789]: I1122 08:41:57.894634 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-nncxd" Nov 22 08:41:58 crc kubenswrapper[4789]: I1122 08:41:58.650699 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-nncxd" Nov 22 08:41:58 crc kubenswrapper[4789]: I1122 08:41:58.693987 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nncxd"] Nov 22 08:42:00 crc kubenswrapper[4789]: I1122 08:42:00.621304 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-nncxd" podUID="d8a1533f-6b43-4ecb-b9b4-702841b049b5" containerName="registry-server" containerID="cri-o://6f2db7cbfeea46b223a1e800eb8f3c082b6cf399c41c4bcd2d5365a5f247f648" gracePeriod=2 Nov 22 08:42:01 crc kubenswrapper[4789]: I1122 08:42:01.046251 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nncxd" Nov 22 08:42:01 crc kubenswrapper[4789]: I1122 08:42:01.201603 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d8a1533f-6b43-4ecb-b9b4-702841b049b5-catalog-content\") pod \"d8a1533f-6b43-4ecb-b9b4-702841b049b5\" (UID: \"d8a1533f-6b43-4ecb-b9b4-702841b049b5\") " Nov 22 08:42:01 crc kubenswrapper[4789]: I1122 08:42:01.202081 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4x4db\" (UniqueName: \"kubernetes.io/projected/d8a1533f-6b43-4ecb-b9b4-702841b049b5-kube-api-access-4x4db\") pod \"d8a1533f-6b43-4ecb-b9b4-702841b049b5\" (UID: \"d8a1533f-6b43-4ecb-b9b4-702841b049b5\") " Nov 22 08:42:01 crc kubenswrapper[4789]: I1122 08:42:01.202277 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d8a1533f-6b43-4ecb-b9b4-702841b049b5-utilities\") pod \"d8a1533f-6b43-4ecb-b9b4-702841b049b5\" (UID: \"d8a1533f-6b43-4ecb-b9b4-702841b049b5\") " Nov 22 08:42:01 crc kubenswrapper[4789]: I1122 08:42:01.203017 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d8a1533f-6b43-4ecb-b9b4-702841b049b5-utilities" (OuterVolumeSpecName: "utilities") pod "d8a1533f-6b43-4ecb-b9b4-702841b049b5" (UID: "d8a1533f-6b43-4ecb-b9b4-702841b049b5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:42:01 crc kubenswrapper[4789]: I1122 08:42:01.207376 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d8a1533f-6b43-4ecb-b9b4-702841b049b5-kube-api-access-4x4db" (OuterVolumeSpecName: "kube-api-access-4x4db") pod "d8a1533f-6b43-4ecb-b9b4-702841b049b5" (UID: "d8a1533f-6b43-4ecb-b9b4-702841b049b5"). InnerVolumeSpecName "kube-api-access-4x4db". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:42:01 crc kubenswrapper[4789]: I1122 08:42:01.254707 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d8a1533f-6b43-4ecb-b9b4-702841b049b5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d8a1533f-6b43-4ecb-b9b4-702841b049b5" (UID: "d8a1533f-6b43-4ecb-b9b4-702841b049b5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:42:01 crc kubenswrapper[4789]: I1122 08:42:01.304971 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d8a1533f-6b43-4ecb-b9b4-702841b049b5-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 08:42:01 crc kubenswrapper[4789]: I1122 08:42:01.305263 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d8a1533f-6b43-4ecb-b9b4-702841b049b5-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 08:42:01 crc kubenswrapper[4789]: I1122 08:42:01.305569 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4x4db\" (UniqueName: \"kubernetes.io/projected/d8a1533f-6b43-4ecb-b9b4-702841b049b5-kube-api-access-4x4db\") on node \"crc\" DevicePath \"\"" Nov 22 08:42:01 crc kubenswrapper[4789]: I1122 08:42:01.633083 4789 generic.go:334] "Generic (PLEG): container finished" podID="d8a1533f-6b43-4ecb-b9b4-702841b049b5" containerID="6f2db7cbfeea46b223a1e800eb8f3c082b6cf399c41c4bcd2d5365a5f247f648" exitCode=0 Nov 22 08:42:01 crc kubenswrapper[4789]: I1122 08:42:01.633143 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nncxd" event={"ID":"d8a1533f-6b43-4ecb-b9b4-702841b049b5","Type":"ContainerDied","Data":"6f2db7cbfeea46b223a1e800eb8f3c082b6cf399c41c4bcd2d5365a5f247f648"} Nov 22 08:42:01 crc kubenswrapper[4789]: I1122 08:42:01.633188 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nncxd" event={"ID":"d8a1533f-6b43-4ecb-b9b4-702841b049b5","Type":"ContainerDied","Data":"1665c2619ff95ad6f88dfca6b744b876b2e63e55b4a002423323543f679ee646"} Nov 22 08:42:01 crc kubenswrapper[4789]: I1122 08:42:01.633216 4789 scope.go:117] "RemoveContainer" containerID="6f2db7cbfeea46b223a1e800eb8f3c082b6cf399c41c4bcd2d5365a5f247f648" Nov 22 08:42:01 crc kubenswrapper[4789]: I1122 08:42:01.633397 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nncxd" Nov 22 08:42:01 crc kubenswrapper[4789]: I1122 08:42:01.655722 4789 scope.go:117] "RemoveContainer" containerID="7b9db75b1392663b0f41f18662e717f5ed881f51a9af9ab09caefd3d84a821ba" Nov 22 08:42:01 crc kubenswrapper[4789]: I1122 08:42:01.663411 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nncxd"] Nov 22 08:42:01 crc kubenswrapper[4789]: I1122 08:42:01.670285 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-nncxd"] Nov 22 08:42:01 crc kubenswrapper[4789]: I1122 08:42:01.682492 4789 scope.go:117] "RemoveContainer" containerID="f3fbcd3629404e976f09e77da2de258462f3a2a5e5ce455a15d4bd2695323539" Nov 22 08:42:01 crc kubenswrapper[4789]: I1122 08:42:01.714496 4789 scope.go:117] "RemoveContainer" containerID="6f2db7cbfeea46b223a1e800eb8f3c082b6cf399c41c4bcd2d5365a5f247f648" Nov 22 08:42:01 crc kubenswrapper[4789]: E1122 08:42:01.714962 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6f2db7cbfeea46b223a1e800eb8f3c082b6cf399c41c4bcd2d5365a5f247f648\": container with ID starting with 6f2db7cbfeea46b223a1e800eb8f3c082b6cf399c41c4bcd2d5365a5f247f648 not found: ID does not exist" containerID="6f2db7cbfeea46b223a1e800eb8f3c082b6cf399c41c4bcd2d5365a5f247f648" Nov 22 08:42:01 crc kubenswrapper[4789]: I1122 08:42:01.715013 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f2db7cbfeea46b223a1e800eb8f3c082b6cf399c41c4bcd2d5365a5f247f648"} err="failed to get container status \"6f2db7cbfeea46b223a1e800eb8f3c082b6cf399c41c4bcd2d5365a5f247f648\": rpc error: code = NotFound desc = could not find container \"6f2db7cbfeea46b223a1e800eb8f3c082b6cf399c41c4bcd2d5365a5f247f648\": container with ID starting with 6f2db7cbfeea46b223a1e800eb8f3c082b6cf399c41c4bcd2d5365a5f247f648 not found: ID does not exist" Nov 22 08:42:01 crc kubenswrapper[4789]: I1122 08:42:01.715046 4789 scope.go:117] "RemoveContainer" containerID="7b9db75b1392663b0f41f18662e717f5ed881f51a9af9ab09caefd3d84a821ba" Nov 22 08:42:01 crc kubenswrapper[4789]: E1122 08:42:01.715563 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7b9db75b1392663b0f41f18662e717f5ed881f51a9af9ab09caefd3d84a821ba\": container with ID starting with 7b9db75b1392663b0f41f18662e717f5ed881f51a9af9ab09caefd3d84a821ba not found: ID does not exist" containerID="7b9db75b1392663b0f41f18662e717f5ed881f51a9af9ab09caefd3d84a821ba" Nov 22 08:42:01 crc kubenswrapper[4789]: I1122 08:42:01.715602 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b9db75b1392663b0f41f18662e717f5ed881f51a9af9ab09caefd3d84a821ba"} err="failed to get container status \"7b9db75b1392663b0f41f18662e717f5ed881f51a9af9ab09caefd3d84a821ba\": rpc error: code = NotFound desc = could not find container \"7b9db75b1392663b0f41f18662e717f5ed881f51a9af9ab09caefd3d84a821ba\": container with ID starting with 7b9db75b1392663b0f41f18662e717f5ed881f51a9af9ab09caefd3d84a821ba not found: ID does not exist" Nov 22 08:42:01 crc kubenswrapper[4789]: I1122 08:42:01.715632 4789 scope.go:117] "RemoveContainer" containerID="f3fbcd3629404e976f09e77da2de258462f3a2a5e5ce455a15d4bd2695323539" Nov 22 08:42:01 crc kubenswrapper[4789]: E1122 08:42:01.716010 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f3fbcd3629404e976f09e77da2de258462f3a2a5e5ce455a15d4bd2695323539\": container with ID starting with f3fbcd3629404e976f09e77da2de258462f3a2a5e5ce455a15d4bd2695323539 not found: ID does not exist" containerID="f3fbcd3629404e976f09e77da2de258462f3a2a5e5ce455a15d4bd2695323539" Nov 22 08:42:01 crc kubenswrapper[4789]: I1122 08:42:01.716157 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f3fbcd3629404e976f09e77da2de258462f3a2a5e5ce455a15d4bd2695323539"} err="failed to get container status \"f3fbcd3629404e976f09e77da2de258462f3a2a5e5ce455a15d4bd2695323539\": rpc error: code = NotFound desc = could not find container \"f3fbcd3629404e976f09e77da2de258462f3a2a5e5ce455a15d4bd2695323539\": container with ID starting with f3fbcd3629404e976f09e77da2de258462f3a2a5e5ce455a15d4bd2695323539 not found: ID does not exist" Nov 22 08:42:01 crc kubenswrapper[4789]: I1122 08:42:01.975175 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d8a1533f-6b43-4ecb-b9b4-702841b049b5" path="/var/lib/kubelet/pods/d8a1533f-6b43-4ecb-b9b4-702841b049b5/volumes" Nov 22 08:42:05 crc kubenswrapper[4789]: I1122 08:42:05.372128 4789 patch_prober.go:28] interesting pod/machine-config-daemon-pxvrm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 08:42:05 crc kubenswrapper[4789]: I1122 08:42:05.373005 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 08:42:15 crc kubenswrapper[4789]: I1122 08:42:15.144868 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-tdnsr"] Nov 22 08:42:15 crc kubenswrapper[4789]: E1122 08:42:15.145737 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8a1533f-6b43-4ecb-b9b4-702841b049b5" containerName="registry-server" Nov 22 08:42:15 crc kubenswrapper[4789]: I1122 08:42:15.145784 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8a1533f-6b43-4ecb-b9b4-702841b049b5" containerName="registry-server" Nov 22 08:42:15 crc kubenswrapper[4789]: E1122 08:42:15.145823 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8a1533f-6b43-4ecb-b9b4-702841b049b5" containerName="extract-utilities" Nov 22 08:42:15 crc kubenswrapper[4789]: I1122 08:42:15.145830 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8a1533f-6b43-4ecb-b9b4-702841b049b5" containerName="extract-utilities" Nov 22 08:42:15 crc kubenswrapper[4789]: E1122 08:42:15.145839 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8a1533f-6b43-4ecb-b9b4-702841b049b5" containerName="extract-content" Nov 22 08:42:15 crc kubenswrapper[4789]: I1122 08:42:15.145845 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8a1533f-6b43-4ecb-b9b4-702841b049b5" containerName="extract-content" Nov 22 08:42:15 crc kubenswrapper[4789]: I1122 08:42:15.146014 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8a1533f-6b43-4ecb-b9b4-702841b049b5" containerName="registry-server" Nov 22 08:42:15 crc kubenswrapper[4789]: I1122 08:42:15.147304 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tdnsr" Nov 22 08:42:15 crc kubenswrapper[4789]: I1122 08:42:15.156338 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tdnsr"] Nov 22 08:42:15 crc kubenswrapper[4789]: I1122 08:42:15.264361 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2fb15975-e0fb-4712-9339-9eefdc468c2f-catalog-content\") pod \"redhat-marketplace-tdnsr\" (UID: \"2fb15975-e0fb-4712-9339-9eefdc468c2f\") " pod="openshift-marketplace/redhat-marketplace-tdnsr" Nov 22 08:42:15 crc kubenswrapper[4789]: I1122 08:42:15.264692 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2fb15975-e0fb-4712-9339-9eefdc468c2f-utilities\") pod \"redhat-marketplace-tdnsr\" (UID: \"2fb15975-e0fb-4712-9339-9eefdc468c2f\") " pod="openshift-marketplace/redhat-marketplace-tdnsr" Nov 22 08:42:15 crc kubenswrapper[4789]: I1122 08:42:15.264930 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bkrp2\" (UniqueName: \"kubernetes.io/projected/2fb15975-e0fb-4712-9339-9eefdc468c2f-kube-api-access-bkrp2\") pod \"redhat-marketplace-tdnsr\" (UID: \"2fb15975-e0fb-4712-9339-9eefdc468c2f\") " pod="openshift-marketplace/redhat-marketplace-tdnsr" Nov 22 08:42:15 crc kubenswrapper[4789]: I1122 08:42:15.367206 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2fb15975-e0fb-4712-9339-9eefdc468c2f-utilities\") pod \"redhat-marketplace-tdnsr\" (UID: \"2fb15975-e0fb-4712-9339-9eefdc468c2f\") " pod="openshift-marketplace/redhat-marketplace-tdnsr" Nov 22 08:42:15 crc kubenswrapper[4789]: I1122 08:42:15.367290 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bkrp2\" (UniqueName: \"kubernetes.io/projected/2fb15975-e0fb-4712-9339-9eefdc468c2f-kube-api-access-bkrp2\") pod \"redhat-marketplace-tdnsr\" (UID: \"2fb15975-e0fb-4712-9339-9eefdc468c2f\") " pod="openshift-marketplace/redhat-marketplace-tdnsr" Nov 22 08:42:15 crc kubenswrapper[4789]: I1122 08:42:15.367553 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2fb15975-e0fb-4712-9339-9eefdc468c2f-catalog-content\") pod \"redhat-marketplace-tdnsr\" (UID: \"2fb15975-e0fb-4712-9339-9eefdc468c2f\") " pod="openshift-marketplace/redhat-marketplace-tdnsr" Nov 22 08:42:15 crc kubenswrapper[4789]: I1122 08:42:15.368125 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2fb15975-e0fb-4712-9339-9eefdc468c2f-utilities\") pod \"redhat-marketplace-tdnsr\" (UID: \"2fb15975-e0fb-4712-9339-9eefdc468c2f\") " pod="openshift-marketplace/redhat-marketplace-tdnsr" Nov 22 08:42:15 crc kubenswrapper[4789]: I1122 08:42:15.368193 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2fb15975-e0fb-4712-9339-9eefdc468c2f-catalog-content\") pod \"redhat-marketplace-tdnsr\" (UID: \"2fb15975-e0fb-4712-9339-9eefdc468c2f\") " pod="openshift-marketplace/redhat-marketplace-tdnsr" Nov 22 08:42:15 crc kubenswrapper[4789]: I1122 08:42:15.387944 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bkrp2\" (UniqueName: \"kubernetes.io/projected/2fb15975-e0fb-4712-9339-9eefdc468c2f-kube-api-access-bkrp2\") pod \"redhat-marketplace-tdnsr\" (UID: \"2fb15975-e0fb-4712-9339-9eefdc468c2f\") " pod="openshift-marketplace/redhat-marketplace-tdnsr" Nov 22 08:42:15 crc kubenswrapper[4789]: I1122 08:42:15.472327 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tdnsr" Nov 22 08:42:15 crc kubenswrapper[4789]: I1122 08:42:15.893018 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tdnsr"] Nov 22 08:42:15 crc kubenswrapper[4789]: W1122 08:42:15.898630 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2fb15975_e0fb_4712_9339_9eefdc468c2f.slice/crio-f070d381c72c72252c321cbf738cc77d79a3acb4c5285e3abed299979349a5c8 WatchSource:0}: Error finding container f070d381c72c72252c321cbf738cc77d79a3acb4c5285e3abed299979349a5c8: Status 404 returned error can't find the container with id f070d381c72c72252c321cbf738cc77d79a3acb4c5285e3abed299979349a5c8 Nov 22 08:42:16 crc kubenswrapper[4789]: I1122 08:42:16.758529 4789 generic.go:334] "Generic (PLEG): container finished" podID="2fb15975-e0fb-4712-9339-9eefdc468c2f" containerID="ff2aa211f565da801208480f8aca502f11ee6facbea52ce40ca0ff9a2e4f9548" exitCode=0 Nov 22 08:42:16 crc kubenswrapper[4789]: I1122 08:42:16.758643 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tdnsr" event={"ID":"2fb15975-e0fb-4712-9339-9eefdc468c2f","Type":"ContainerDied","Data":"ff2aa211f565da801208480f8aca502f11ee6facbea52ce40ca0ff9a2e4f9548"} Nov 22 08:42:16 crc kubenswrapper[4789]: I1122 08:42:16.758894 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tdnsr" event={"ID":"2fb15975-e0fb-4712-9339-9eefdc468c2f","Type":"ContainerStarted","Data":"f070d381c72c72252c321cbf738cc77d79a3acb4c5285e3abed299979349a5c8"} Nov 22 08:42:17 crc kubenswrapper[4789]: I1122 08:42:17.770166 4789 generic.go:334] "Generic (PLEG): container finished" podID="2fb15975-e0fb-4712-9339-9eefdc468c2f" containerID="18b4f6085e9dd383e88850a7d3a3e3411a40f1f5381f4afa21022051ef97f808" exitCode=0 Nov 22 08:42:17 crc kubenswrapper[4789]: I1122 08:42:17.770378 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tdnsr" event={"ID":"2fb15975-e0fb-4712-9339-9eefdc468c2f","Type":"ContainerDied","Data":"18b4f6085e9dd383e88850a7d3a3e3411a40f1f5381f4afa21022051ef97f808"} Nov 22 08:42:18 crc kubenswrapper[4789]: I1122 08:42:18.780543 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tdnsr" event={"ID":"2fb15975-e0fb-4712-9339-9eefdc468c2f","Type":"ContainerStarted","Data":"10a19262c442c536471df2e60dfcde14a1e374ae61fc83124a63945ae0ed02d6"} Nov 22 08:42:18 crc kubenswrapper[4789]: I1122 08:42:18.811609 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-tdnsr" podStartSLOduration=2.372520403 podStartE2EDuration="3.811592266s" podCreationTimestamp="2025-11-22 08:42:15 +0000 UTC" firstStartedPulling="2025-11-22 08:42:16.761055786 +0000 UTC m=+2750.995456059" lastFinishedPulling="2025-11-22 08:42:18.200127629 +0000 UTC m=+2752.434527922" observedRunningTime="2025-11-22 08:42:18.806432685 +0000 UTC m=+2753.040832958" watchObservedRunningTime="2025-11-22 08:42:18.811592266 +0000 UTC m=+2753.045992549" Nov 22 08:42:25 crc kubenswrapper[4789]: I1122 08:42:25.473114 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-tdnsr" Nov 22 08:42:25 crc kubenswrapper[4789]: I1122 08:42:25.473548 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-tdnsr" Nov 22 08:42:25 crc kubenswrapper[4789]: I1122 08:42:25.522230 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-tdnsr" Nov 22 08:42:25 crc kubenswrapper[4789]: I1122 08:42:25.887729 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-tdnsr" Nov 22 08:42:25 crc kubenswrapper[4789]: I1122 08:42:25.936518 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-tdnsr"] Nov 22 08:42:27 crc kubenswrapper[4789]: I1122 08:42:27.860739 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-tdnsr" podUID="2fb15975-e0fb-4712-9339-9eefdc468c2f" containerName="registry-server" containerID="cri-o://10a19262c442c536471df2e60dfcde14a1e374ae61fc83124a63945ae0ed02d6" gracePeriod=2 Nov 22 08:42:28 crc kubenswrapper[4789]: I1122 08:42:28.257274 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tdnsr" Nov 22 08:42:28 crc kubenswrapper[4789]: I1122 08:42:28.299157 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2fb15975-e0fb-4712-9339-9eefdc468c2f-catalog-content\") pod \"2fb15975-e0fb-4712-9339-9eefdc468c2f\" (UID: \"2fb15975-e0fb-4712-9339-9eefdc468c2f\") " Nov 22 08:42:28 crc kubenswrapper[4789]: I1122 08:42:28.299313 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2fb15975-e0fb-4712-9339-9eefdc468c2f-utilities\") pod \"2fb15975-e0fb-4712-9339-9eefdc468c2f\" (UID: \"2fb15975-e0fb-4712-9339-9eefdc468c2f\") " Nov 22 08:42:28 crc kubenswrapper[4789]: I1122 08:42:28.299339 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bkrp2\" (UniqueName: \"kubernetes.io/projected/2fb15975-e0fb-4712-9339-9eefdc468c2f-kube-api-access-bkrp2\") pod \"2fb15975-e0fb-4712-9339-9eefdc468c2f\" (UID: \"2fb15975-e0fb-4712-9339-9eefdc468c2f\") " Nov 22 08:42:28 crc kubenswrapper[4789]: I1122 08:42:28.300428 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2fb15975-e0fb-4712-9339-9eefdc468c2f-utilities" (OuterVolumeSpecName: "utilities") pod "2fb15975-e0fb-4712-9339-9eefdc468c2f" (UID: "2fb15975-e0fb-4712-9339-9eefdc468c2f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:42:28 crc kubenswrapper[4789]: I1122 08:42:28.305491 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2fb15975-e0fb-4712-9339-9eefdc468c2f-kube-api-access-bkrp2" (OuterVolumeSpecName: "kube-api-access-bkrp2") pod "2fb15975-e0fb-4712-9339-9eefdc468c2f" (UID: "2fb15975-e0fb-4712-9339-9eefdc468c2f"). InnerVolumeSpecName "kube-api-access-bkrp2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:42:28 crc kubenswrapper[4789]: I1122 08:42:28.324586 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2fb15975-e0fb-4712-9339-9eefdc468c2f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2fb15975-e0fb-4712-9339-9eefdc468c2f" (UID: "2fb15975-e0fb-4712-9339-9eefdc468c2f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:42:28 crc kubenswrapper[4789]: I1122 08:42:28.401454 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2fb15975-e0fb-4712-9339-9eefdc468c2f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 08:42:28 crc kubenswrapper[4789]: I1122 08:42:28.401524 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2fb15975-e0fb-4712-9339-9eefdc468c2f-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 08:42:28 crc kubenswrapper[4789]: I1122 08:42:28.401534 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bkrp2\" (UniqueName: \"kubernetes.io/projected/2fb15975-e0fb-4712-9339-9eefdc468c2f-kube-api-access-bkrp2\") on node \"crc\" DevicePath \"\"" Nov 22 08:42:28 crc kubenswrapper[4789]: I1122 08:42:28.883736 4789 generic.go:334] "Generic (PLEG): container finished" podID="2fb15975-e0fb-4712-9339-9eefdc468c2f" containerID="10a19262c442c536471df2e60dfcde14a1e374ae61fc83124a63945ae0ed02d6" exitCode=0 Nov 22 08:42:28 crc kubenswrapper[4789]: I1122 08:42:28.883806 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tdnsr" event={"ID":"2fb15975-e0fb-4712-9339-9eefdc468c2f","Type":"ContainerDied","Data":"10a19262c442c536471df2e60dfcde14a1e374ae61fc83124a63945ae0ed02d6"} Nov 22 08:42:28 crc kubenswrapper[4789]: I1122 08:42:28.883840 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tdnsr" event={"ID":"2fb15975-e0fb-4712-9339-9eefdc468c2f","Type":"ContainerDied","Data":"f070d381c72c72252c321cbf738cc77d79a3acb4c5285e3abed299979349a5c8"} Nov 22 08:42:28 crc kubenswrapper[4789]: I1122 08:42:28.883873 4789 scope.go:117] "RemoveContainer" containerID="10a19262c442c536471df2e60dfcde14a1e374ae61fc83124a63945ae0ed02d6" Nov 22 08:42:28 crc kubenswrapper[4789]: I1122 08:42:28.883903 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tdnsr" Nov 22 08:42:28 crc kubenswrapper[4789]: I1122 08:42:28.916524 4789 scope.go:117] "RemoveContainer" containerID="18b4f6085e9dd383e88850a7d3a3e3411a40f1f5381f4afa21022051ef97f808" Nov 22 08:42:28 crc kubenswrapper[4789]: I1122 08:42:28.921098 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-tdnsr"] Nov 22 08:42:28 crc kubenswrapper[4789]: I1122 08:42:28.927854 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-tdnsr"] Nov 22 08:42:28 crc kubenswrapper[4789]: I1122 08:42:28.957821 4789 scope.go:117] "RemoveContainer" containerID="ff2aa211f565da801208480f8aca502f11ee6facbea52ce40ca0ff9a2e4f9548" Nov 22 08:42:28 crc kubenswrapper[4789]: I1122 08:42:28.979548 4789 scope.go:117] "RemoveContainer" containerID="10a19262c442c536471df2e60dfcde14a1e374ae61fc83124a63945ae0ed02d6" Nov 22 08:42:28 crc kubenswrapper[4789]: E1122 08:42:28.980084 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"10a19262c442c536471df2e60dfcde14a1e374ae61fc83124a63945ae0ed02d6\": container with ID starting with 10a19262c442c536471df2e60dfcde14a1e374ae61fc83124a63945ae0ed02d6 not found: ID does not exist" containerID="10a19262c442c536471df2e60dfcde14a1e374ae61fc83124a63945ae0ed02d6" Nov 22 08:42:28 crc kubenswrapper[4789]: I1122 08:42:28.980128 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"10a19262c442c536471df2e60dfcde14a1e374ae61fc83124a63945ae0ed02d6"} err="failed to get container status \"10a19262c442c536471df2e60dfcde14a1e374ae61fc83124a63945ae0ed02d6\": rpc error: code = NotFound desc = could not find container \"10a19262c442c536471df2e60dfcde14a1e374ae61fc83124a63945ae0ed02d6\": container with ID starting with 10a19262c442c536471df2e60dfcde14a1e374ae61fc83124a63945ae0ed02d6 not found: ID does not exist" Nov 22 08:42:28 crc kubenswrapper[4789]: I1122 08:42:28.980160 4789 scope.go:117] "RemoveContainer" containerID="18b4f6085e9dd383e88850a7d3a3e3411a40f1f5381f4afa21022051ef97f808" Nov 22 08:42:28 crc kubenswrapper[4789]: E1122 08:42:28.980638 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"18b4f6085e9dd383e88850a7d3a3e3411a40f1f5381f4afa21022051ef97f808\": container with ID starting with 18b4f6085e9dd383e88850a7d3a3e3411a40f1f5381f4afa21022051ef97f808 not found: ID does not exist" containerID="18b4f6085e9dd383e88850a7d3a3e3411a40f1f5381f4afa21022051ef97f808" Nov 22 08:42:28 crc kubenswrapper[4789]: I1122 08:42:28.980665 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"18b4f6085e9dd383e88850a7d3a3e3411a40f1f5381f4afa21022051ef97f808"} err="failed to get container status \"18b4f6085e9dd383e88850a7d3a3e3411a40f1f5381f4afa21022051ef97f808\": rpc error: code = NotFound desc = could not find container \"18b4f6085e9dd383e88850a7d3a3e3411a40f1f5381f4afa21022051ef97f808\": container with ID starting with 18b4f6085e9dd383e88850a7d3a3e3411a40f1f5381f4afa21022051ef97f808 not found: ID does not exist" Nov 22 08:42:28 crc kubenswrapper[4789]: I1122 08:42:28.980683 4789 scope.go:117] "RemoveContainer" containerID="ff2aa211f565da801208480f8aca502f11ee6facbea52ce40ca0ff9a2e4f9548" Nov 22 08:42:28 crc kubenswrapper[4789]: E1122 08:42:28.980994 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff2aa211f565da801208480f8aca502f11ee6facbea52ce40ca0ff9a2e4f9548\": container with ID starting with ff2aa211f565da801208480f8aca502f11ee6facbea52ce40ca0ff9a2e4f9548 not found: ID does not exist" containerID="ff2aa211f565da801208480f8aca502f11ee6facbea52ce40ca0ff9a2e4f9548" Nov 22 08:42:28 crc kubenswrapper[4789]: I1122 08:42:28.981020 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff2aa211f565da801208480f8aca502f11ee6facbea52ce40ca0ff9a2e4f9548"} err="failed to get container status \"ff2aa211f565da801208480f8aca502f11ee6facbea52ce40ca0ff9a2e4f9548\": rpc error: code = NotFound desc = could not find container \"ff2aa211f565da801208480f8aca502f11ee6facbea52ce40ca0ff9a2e4f9548\": container with ID starting with ff2aa211f565da801208480f8aca502f11ee6facbea52ce40ca0ff9a2e4f9548 not found: ID does not exist" Nov 22 08:42:29 crc kubenswrapper[4789]: I1122 08:42:29.976036 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2fb15975-e0fb-4712-9339-9eefdc468c2f" path="/var/lib/kubelet/pods/2fb15975-e0fb-4712-9339-9eefdc468c2f/volumes" Nov 22 08:42:35 crc kubenswrapper[4789]: I1122 08:42:35.372920 4789 patch_prober.go:28] interesting pod/machine-config-daemon-pxvrm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 08:42:35 crc kubenswrapper[4789]: I1122 08:42:35.373499 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 08:42:35 crc kubenswrapper[4789]: I1122 08:42:35.373551 4789 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" Nov 22 08:42:35 crc kubenswrapper[4789]: I1122 08:42:35.374434 4789 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"95ee0f2b0a62b96c30cc7400004c7e7b084559a4811441f88de89dadf3d054d1"} pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 22 08:42:35 crc kubenswrapper[4789]: I1122 08:42:35.374504 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" containerID="cri-o://95ee0f2b0a62b96c30cc7400004c7e7b084559a4811441f88de89dadf3d054d1" gracePeriod=600 Nov 22 08:42:35 crc kubenswrapper[4789]: I1122 08:42:35.954472 4789 generic.go:334] "Generic (PLEG): container finished" podID="58aa8071-7ecc-4692-8789-2db04ec70647" containerID="95ee0f2b0a62b96c30cc7400004c7e7b084559a4811441f88de89dadf3d054d1" exitCode=0 Nov 22 08:42:35 crc kubenswrapper[4789]: I1122 08:42:35.954574 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" event={"ID":"58aa8071-7ecc-4692-8789-2db04ec70647","Type":"ContainerDied","Data":"95ee0f2b0a62b96c30cc7400004c7e7b084559a4811441f88de89dadf3d054d1"} Nov 22 08:42:35 crc kubenswrapper[4789]: I1122 08:42:35.955368 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" event={"ID":"58aa8071-7ecc-4692-8789-2db04ec70647","Type":"ContainerStarted","Data":"e6fcfa0e8fd700efa0433d598c23ce626ca7a007b2b45b45ee839e17923bad02"} Nov 22 08:42:35 crc kubenswrapper[4789]: I1122 08:42:35.955459 4789 scope.go:117] "RemoveContainer" containerID="c7898b5c08854d5df6033d34564729f053b077e4fcadc5a1ea77706f1981be3b" Nov 22 08:43:08 crc kubenswrapper[4789]: I1122 08:43:08.242644 4789 generic.go:334] "Generic (PLEG): container finished" podID="afd5b574-89f9-4b13-a3ee-b5e4c39095ff" containerID="48ab9540830de88f3403f93978df93af82fd5058c07dc2a20fd3193c417fc881" exitCode=0 Nov 22 08:43:08 crc kubenswrapper[4789]: I1122 08:43:08.242770 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-tqgkq" event={"ID":"afd5b574-89f9-4b13-a3ee-b5e4c39095ff","Type":"ContainerDied","Data":"48ab9540830de88f3403f93978df93af82fd5058c07dc2a20fd3193c417fc881"} Nov 22 08:43:09 crc kubenswrapper[4789]: I1122 08:43:09.645224 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-tqgkq" Nov 22 08:43:09 crc kubenswrapper[4789]: I1122 08:43:09.771288 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qm9jz\" (UniqueName: \"kubernetes.io/projected/afd5b574-89f9-4b13-a3ee-b5e4c39095ff-kube-api-access-qm9jz\") pod \"afd5b574-89f9-4b13-a3ee-b5e4c39095ff\" (UID: \"afd5b574-89f9-4b13-a3ee-b5e4c39095ff\") " Nov 22 08:43:09 crc kubenswrapper[4789]: I1122 08:43:09.771808 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/afd5b574-89f9-4b13-a3ee-b5e4c39095ff-ssh-key\") pod \"afd5b574-89f9-4b13-a3ee-b5e4c39095ff\" (UID: \"afd5b574-89f9-4b13-a3ee-b5e4c39095ff\") " Nov 22 08:43:09 crc kubenswrapper[4789]: I1122 08:43:09.771895 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/afd5b574-89f9-4b13-a3ee-b5e4c39095ff-inventory\") pod \"afd5b574-89f9-4b13-a3ee-b5e4c39095ff\" (UID: \"afd5b574-89f9-4b13-a3ee-b5e4c39095ff\") " Nov 22 08:43:09 crc kubenswrapper[4789]: I1122 08:43:09.772044 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/afd5b574-89f9-4b13-a3ee-b5e4c39095ff-ceph\") pod \"afd5b574-89f9-4b13-a3ee-b5e4c39095ff\" (UID: \"afd5b574-89f9-4b13-a3ee-b5e4c39095ff\") " Nov 22 08:43:09 crc kubenswrapper[4789]: I1122 08:43:09.772089 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/afd5b574-89f9-4b13-a3ee-b5e4c39095ff-ovn-combined-ca-bundle\") pod \"afd5b574-89f9-4b13-a3ee-b5e4c39095ff\" (UID: \"afd5b574-89f9-4b13-a3ee-b5e4c39095ff\") " Nov 22 08:43:09 crc kubenswrapper[4789]: I1122 08:43:09.772132 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/afd5b574-89f9-4b13-a3ee-b5e4c39095ff-ovncontroller-config-0\") pod \"afd5b574-89f9-4b13-a3ee-b5e4c39095ff\" (UID: \"afd5b574-89f9-4b13-a3ee-b5e4c39095ff\") " Nov 22 08:43:09 crc kubenswrapper[4789]: I1122 08:43:09.778658 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/afd5b574-89f9-4b13-a3ee-b5e4c39095ff-ceph" (OuterVolumeSpecName: "ceph") pod "afd5b574-89f9-4b13-a3ee-b5e4c39095ff" (UID: "afd5b574-89f9-4b13-a3ee-b5e4c39095ff"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:43:09 crc kubenswrapper[4789]: I1122 08:43:09.778876 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/afd5b574-89f9-4b13-a3ee-b5e4c39095ff-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "afd5b574-89f9-4b13-a3ee-b5e4c39095ff" (UID: "afd5b574-89f9-4b13-a3ee-b5e4c39095ff"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:43:09 crc kubenswrapper[4789]: I1122 08:43:09.779004 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/afd5b574-89f9-4b13-a3ee-b5e4c39095ff-kube-api-access-qm9jz" (OuterVolumeSpecName: "kube-api-access-qm9jz") pod "afd5b574-89f9-4b13-a3ee-b5e4c39095ff" (UID: "afd5b574-89f9-4b13-a3ee-b5e4c39095ff"). InnerVolumeSpecName "kube-api-access-qm9jz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:43:09 crc kubenswrapper[4789]: I1122 08:43:09.798560 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/afd5b574-89f9-4b13-a3ee-b5e4c39095ff-inventory" (OuterVolumeSpecName: "inventory") pod "afd5b574-89f9-4b13-a3ee-b5e4c39095ff" (UID: "afd5b574-89f9-4b13-a3ee-b5e4c39095ff"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:43:09 crc kubenswrapper[4789]: I1122 08:43:09.801030 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/afd5b574-89f9-4b13-a3ee-b5e4c39095ff-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "afd5b574-89f9-4b13-a3ee-b5e4c39095ff" (UID: "afd5b574-89f9-4b13-a3ee-b5e4c39095ff"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:43:09 crc kubenswrapper[4789]: I1122 08:43:09.805667 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/afd5b574-89f9-4b13-a3ee-b5e4c39095ff-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "afd5b574-89f9-4b13-a3ee-b5e4c39095ff" (UID: "afd5b574-89f9-4b13-a3ee-b5e4c39095ff"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:43:09 crc kubenswrapper[4789]: I1122 08:43:09.874292 4789 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/afd5b574-89f9-4b13-a3ee-b5e4c39095ff-inventory\") on node \"crc\" DevicePath \"\"" Nov 22 08:43:09 crc kubenswrapper[4789]: I1122 08:43:09.874343 4789 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/afd5b574-89f9-4b13-a3ee-b5e4c39095ff-ceph\") on node \"crc\" DevicePath \"\"" Nov 22 08:43:09 crc kubenswrapper[4789]: I1122 08:43:09.874356 4789 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/afd5b574-89f9-4b13-a3ee-b5e4c39095ff-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:43:09 crc kubenswrapper[4789]: I1122 08:43:09.874369 4789 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/afd5b574-89f9-4b13-a3ee-b5e4c39095ff-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Nov 22 08:43:09 crc kubenswrapper[4789]: I1122 08:43:09.874382 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qm9jz\" (UniqueName: \"kubernetes.io/projected/afd5b574-89f9-4b13-a3ee-b5e4c39095ff-kube-api-access-qm9jz\") on node \"crc\" DevicePath \"\"" Nov 22 08:43:09 crc kubenswrapper[4789]: I1122 08:43:09.874391 4789 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/afd5b574-89f9-4b13-a3ee-b5e4c39095ff-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 22 08:43:10 crc kubenswrapper[4789]: I1122 08:43:10.260018 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-tqgkq" event={"ID":"afd5b574-89f9-4b13-a3ee-b5e4c39095ff","Type":"ContainerDied","Data":"d0e1df0a7d7a252c254565ef2007458522bce2cf9f348eefd6b7919f3e70d3aa"} Nov 22 08:43:10 crc kubenswrapper[4789]: I1122 08:43:10.260057 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-tqgkq" Nov 22 08:43:10 crc kubenswrapper[4789]: I1122 08:43:10.260059 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d0e1df0a7d7a252c254565ef2007458522bce2cf9f348eefd6b7919f3e70d3aa" Nov 22 08:43:10 crc kubenswrapper[4789]: I1122 08:43:10.335828 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-qn45p"] Nov 22 08:43:10 crc kubenswrapper[4789]: E1122 08:43:10.336502 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afd5b574-89f9-4b13-a3ee-b5e4c39095ff" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Nov 22 08:43:10 crc kubenswrapper[4789]: I1122 08:43:10.336588 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="afd5b574-89f9-4b13-a3ee-b5e4c39095ff" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Nov 22 08:43:10 crc kubenswrapper[4789]: E1122 08:43:10.336654 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fb15975-e0fb-4712-9339-9eefdc468c2f" containerName="extract-utilities" Nov 22 08:43:10 crc kubenswrapper[4789]: I1122 08:43:10.336775 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fb15975-e0fb-4712-9339-9eefdc468c2f" containerName="extract-utilities" Nov 22 08:43:10 crc kubenswrapper[4789]: E1122 08:43:10.336877 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fb15975-e0fb-4712-9339-9eefdc468c2f" containerName="extract-content" Nov 22 08:43:10 crc kubenswrapper[4789]: I1122 08:43:10.336947 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fb15975-e0fb-4712-9339-9eefdc468c2f" containerName="extract-content" Nov 22 08:43:10 crc kubenswrapper[4789]: E1122 08:43:10.337049 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fb15975-e0fb-4712-9339-9eefdc468c2f" containerName="registry-server" Nov 22 08:43:10 crc kubenswrapper[4789]: I1122 08:43:10.337125 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fb15975-e0fb-4712-9339-9eefdc468c2f" containerName="registry-server" Nov 22 08:43:10 crc kubenswrapper[4789]: I1122 08:43:10.337377 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="2fb15975-e0fb-4712-9339-9eefdc468c2f" containerName="registry-server" Nov 22 08:43:10 crc kubenswrapper[4789]: I1122 08:43:10.337445 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="afd5b574-89f9-4b13-a3ee-b5e4c39095ff" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Nov 22 08:43:10 crc kubenswrapper[4789]: I1122 08:43:10.338122 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-qn45p" Nov 22 08:43:10 crc kubenswrapper[4789]: I1122 08:43:10.345018 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Nov 22 08:43:10 crc kubenswrapper[4789]: I1122 08:43:10.345038 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 22 08:43:10 crc kubenswrapper[4789]: I1122 08:43:10.345092 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 22 08:43:10 crc kubenswrapper[4789]: I1122 08:43:10.345158 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 22 08:43:10 crc kubenswrapper[4789]: I1122 08:43:10.345041 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 22 08:43:10 crc kubenswrapper[4789]: I1122 08:43:10.345240 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Nov 22 08:43:10 crc kubenswrapper[4789]: I1122 08:43:10.345535 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-glmr7" Nov 22 08:43:10 crc kubenswrapper[4789]: I1122 08:43:10.350101 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-qn45p"] Nov 22 08:43:10 crc kubenswrapper[4789]: I1122 08:43:10.382391 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/65952d06-ccf5-425f-bb4a-ad127a6844df-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-qn45p\" (UID: \"65952d06-ccf5-425f-bb4a-ad127a6844df\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-qn45p" Nov 22 08:43:10 crc kubenswrapper[4789]: I1122 08:43:10.382443 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/65952d06-ccf5-425f-bb4a-ad127a6844df-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-qn45p\" (UID: \"65952d06-ccf5-425f-bb4a-ad127a6844df\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-qn45p" Nov 22 08:43:10 crc kubenswrapper[4789]: I1122 08:43:10.382521 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65952d06-ccf5-425f-bb4a-ad127a6844df-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-qn45p\" (UID: \"65952d06-ccf5-425f-bb4a-ad127a6844df\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-qn45p" Nov 22 08:43:10 crc kubenswrapper[4789]: I1122 08:43:10.382548 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/65952d06-ccf5-425f-bb4a-ad127a6844df-ceph\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-qn45p\" (UID: \"65952d06-ccf5-425f-bb4a-ad127a6844df\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-qn45p" Nov 22 08:43:10 crc kubenswrapper[4789]: I1122 08:43:10.382595 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c7hhr\" (UniqueName: \"kubernetes.io/projected/65952d06-ccf5-425f-bb4a-ad127a6844df-kube-api-access-c7hhr\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-qn45p\" (UID: \"65952d06-ccf5-425f-bb4a-ad127a6844df\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-qn45p" Nov 22 08:43:10 crc kubenswrapper[4789]: I1122 08:43:10.382694 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/65952d06-ccf5-425f-bb4a-ad127a6844df-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-qn45p\" (UID: \"65952d06-ccf5-425f-bb4a-ad127a6844df\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-qn45p" Nov 22 08:43:10 crc kubenswrapper[4789]: I1122 08:43:10.382739 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/65952d06-ccf5-425f-bb4a-ad127a6844df-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-qn45p\" (UID: \"65952d06-ccf5-425f-bb4a-ad127a6844df\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-qn45p" Nov 22 08:43:10 crc kubenswrapper[4789]: I1122 08:43:10.484450 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/65952d06-ccf5-425f-bb4a-ad127a6844df-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-qn45p\" (UID: \"65952d06-ccf5-425f-bb4a-ad127a6844df\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-qn45p" Nov 22 08:43:10 crc kubenswrapper[4789]: I1122 08:43:10.484541 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65952d06-ccf5-425f-bb4a-ad127a6844df-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-qn45p\" (UID: \"65952d06-ccf5-425f-bb4a-ad127a6844df\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-qn45p" Nov 22 08:43:10 crc kubenswrapper[4789]: I1122 08:43:10.484571 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/65952d06-ccf5-425f-bb4a-ad127a6844df-ceph\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-qn45p\" (UID: \"65952d06-ccf5-425f-bb4a-ad127a6844df\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-qn45p" Nov 22 08:43:10 crc kubenswrapper[4789]: I1122 08:43:10.484619 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c7hhr\" (UniqueName: \"kubernetes.io/projected/65952d06-ccf5-425f-bb4a-ad127a6844df-kube-api-access-c7hhr\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-qn45p\" (UID: \"65952d06-ccf5-425f-bb4a-ad127a6844df\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-qn45p" Nov 22 08:43:10 crc kubenswrapper[4789]: I1122 08:43:10.484661 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/65952d06-ccf5-425f-bb4a-ad127a6844df-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-qn45p\" (UID: \"65952d06-ccf5-425f-bb4a-ad127a6844df\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-qn45p" Nov 22 08:43:10 crc kubenswrapper[4789]: I1122 08:43:10.484702 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/65952d06-ccf5-425f-bb4a-ad127a6844df-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-qn45p\" (UID: \"65952d06-ccf5-425f-bb4a-ad127a6844df\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-qn45p" Nov 22 08:43:10 crc kubenswrapper[4789]: I1122 08:43:10.484775 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/65952d06-ccf5-425f-bb4a-ad127a6844df-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-qn45p\" (UID: \"65952d06-ccf5-425f-bb4a-ad127a6844df\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-qn45p" Nov 22 08:43:10 crc kubenswrapper[4789]: I1122 08:43:10.490355 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/65952d06-ccf5-425f-bb4a-ad127a6844df-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-qn45p\" (UID: \"65952d06-ccf5-425f-bb4a-ad127a6844df\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-qn45p" Nov 22 08:43:10 crc kubenswrapper[4789]: I1122 08:43:10.490634 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/65952d06-ccf5-425f-bb4a-ad127a6844df-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-qn45p\" (UID: \"65952d06-ccf5-425f-bb4a-ad127a6844df\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-qn45p" Nov 22 08:43:10 crc kubenswrapper[4789]: I1122 08:43:10.490703 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/65952d06-ccf5-425f-bb4a-ad127a6844df-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-qn45p\" (UID: \"65952d06-ccf5-425f-bb4a-ad127a6844df\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-qn45p" Nov 22 08:43:10 crc kubenswrapper[4789]: I1122 08:43:10.491441 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/65952d06-ccf5-425f-bb4a-ad127a6844df-ceph\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-qn45p\" (UID: \"65952d06-ccf5-425f-bb4a-ad127a6844df\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-qn45p" Nov 22 08:43:10 crc kubenswrapper[4789]: I1122 08:43:10.492347 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/65952d06-ccf5-425f-bb4a-ad127a6844df-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-qn45p\" (UID: \"65952d06-ccf5-425f-bb4a-ad127a6844df\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-qn45p" Nov 22 08:43:10 crc kubenswrapper[4789]: I1122 08:43:10.492747 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65952d06-ccf5-425f-bb4a-ad127a6844df-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-qn45p\" (UID: \"65952d06-ccf5-425f-bb4a-ad127a6844df\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-qn45p" Nov 22 08:43:10 crc kubenswrapper[4789]: I1122 08:43:10.504898 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c7hhr\" (UniqueName: \"kubernetes.io/projected/65952d06-ccf5-425f-bb4a-ad127a6844df-kube-api-access-c7hhr\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-qn45p\" (UID: \"65952d06-ccf5-425f-bb4a-ad127a6844df\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-qn45p" Nov 22 08:43:10 crc kubenswrapper[4789]: I1122 08:43:10.663170 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-qn45p" Nov 22 08:43:11 crc kubenswrapper[4789]: I1122 08:43:11.145464 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-qn45p"] Nov 22 08:43:11 crc kubenswrapper[4789]: I1122 08:43:11.270352 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-qn45p" event={"ID":"65952d06-ccf5-425f-bb4a-ad127a6844df","Type":"ContainerStarted","Data":"61c7830a9b4f472ee4b737b98e2c0a680fbeb0c75fdd55cf2f14fefc36c47a84"} Nov 22 08:43:12 crc kubenswrapper[4789]: I1122 08:43:12.279649 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-qn45p" event={"ID":"65952d06-ccf5-425f-bb4a-ad127a6844df","Type":"ContainerStarted","Data":"8871490a223e15a7aa4dab635155863f4632c7bf1d7f5c034d95cb2cba202dbc"} Nov 22 08:43:12 crc kubenswrapper[4789]: I1122 08:43:12.309544 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-qn45p" podStartSLOduration=1.861292462 podStartE2EDuration="2.309517573s" podCreationTimestamp="2025-11-22 08:43:10 +0000 UTC" firstStartedPulling="2025-11-22 08:43:11.153746051 +0000 UTC m=+2805.388146324" lastFinishedPulling="2025-11-22 08:43:11.601971162 +0000 UTC m=+2805.836371435" observedRunningTime="2025-11-22 08:43:12.297184619 +0000 UTC m=+2806.531584902" watchObservedRunningTime="2025-11-22 08:43:12.309517573 +0000 UTC m=+2806.543917886" Nov 22 08:44:27 crc kubenswrapper[4789]: I1122 08:44:27.929766 4789 generic.go:334] "Generic (PLEG): container finished" podID="65952d06-ccf5-425f-bb4a-ad127a6844df" containerID="8871490a223e15a7aa4dab635155863f4632c7bf1d7f5c034d95cb2cba202dbc" exitCode=0 Nov 22 08:44:27 crc kubenswrapper[4789]: I1122 08:44:27.929784 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-qn45p" event={"ID":"65952d06-ccf5-425f-bb4a-ad127a6844df","Type":"ContainerDied","Data":"8871490a223e15a7aa4dab635155863f4632c7bf1d7f5c034d95cb2cba202dbc"} Nov 22 08:44:29 crc kubenswrapper[4789]: I1122 08:44:29.139384 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-7tj9w"] Nov 22 08:44:29 crc kubenswrapper[4789]: I1122 08:44:29.145148 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7tj9w" Nov 22 08:44:29 crc kubenswrapper[4789]: I1122 08:44:29.153220 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7tj9w"] Nov 22 08:44:29 crc kubenswrapper[4789]: I1122 08:44:29.175712 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6xlzq\" (UniqueName: \"kubernetes.io/projected/3d54f943-2f4d-424a-8bf7-b523da5eb0f6-kube-api-access-6xlzq\") pod \"redhat-operators-7tj9w\" (UID: \"3d54f943-2f4d-424a-8bf7-b523da5eb0f6\") " pod="openshift-marketplace/redhat-operators-7tj9w" Nov 22 08:44:29 crc kubenswrapper[4789]: I1122 08:44:29.175819 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d54f943-2f4d-424a-8bf7-b523da5eb0f6-utilities\") pod \"redhat-operators-7tj9w\" (UID: \"3d54f943-2f4d-424a-8bf7-b523da5eb0f6\") " pod="openshift-marketplace/redhat-operators-7tj9w" Nov 22 08:44:29 crc kubenswrapper[4789]: I1122 08:44:29.176292 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d54f943-2f4d-424a-8bf7-b523da5eb0f6-catalog-content\") pod \"redhat-operators-7tj9w\" (UID: \"3d54f943-2f4d-424a-8bf7-b523da5eb0f6\") " pod="openshift-marketplace/redhat-operators-7tj9w" Nov 22 08:44:29 crc kubenswrapper[4789]: I1122 08:44:29.279613 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d54f943-2f4d-424a-8bf7-b523da5eb0f6-catalog-content\") pod \"redhat-operators-7tj9w\" (UID: \"3d54f943-2f4d-424a-8bf7-b523da5eb0f6\") " pod="openshift-marketplace/redhat-operators-7tj9w" Nov 22 08:44:29 crc kubenswrapper[4789]: I1122 08:44:29.279716 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6xlzq\" (UniqueName: \"kubernetes.io/projected/3d54f943-2f4d-424a-8bf7-b523da5eb0f6-kube-api-access-6xlzq\") pod \"redhat-operators-7tj9w\" (UID: \"3d54f943-2f4d-424a-8bf7-b523da5eb0f6\") " pod="openshift-marketplace/redhat-operators-7tj9w" Nov 22 08:44:29 crc kubenswrapper[4789]: I1122 08:44:29.279793 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d54f943-2f4d-424a-8bf7-b523da5eb0f6-utilities\") pod \"redhat-operators-7tj9w\" (UID: \"3d54f943-2f4d-424a-8bf7-b523da5eb0f6\") " pod="openshift-marketplace/redhat-operators-7tj9w" Nov 22 08:44:29 crc kubenswrapper[4789]: I1122 08:44:29.280317 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d54f943-2f4d-424a-8bf7-b523da5eb0f6-catalog-content\") pod \"redhat-operators-7tj9w\" (UID: \"3d54f943-2f4d-424a-8bf7-b523da5eb0f6\") " pod="openshift-marketplace/redhat-operators-7tj9w" Nov 22 08:44:29 crc kubenswrapper[4789]: I1122 08:44:29.280354 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d54f943-2f4d-424a-8bf7-b523da5eb0f6-utilities\") pod \"redhat-operators-7tj9w\" (UID: \"3d54f943-2f4d-424a-8bf7-b523da5eb0f6\") " pod="openshift-marketplace/redhat-operators-7tj9w" Nov 22 08:44:29 crc kubenswrapper[4789]: I1122 08:44:29.302411 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6xlzq\" (UniqueName: \"kubernetes.io/projected/3d54f943-2f4d-424a-8bf7-b523da5eb0f6-kube-api-access-6xlzq\") pod \"redhat-operators-7tj9w\" (UID: \"3d54f943-2f4d-424a-8bf7-b523da5eb0f6\") " pod="openshift-marketplace/redhat-operators-7tj9w" Nov 22 08:44:29 crc kubenswrapper[4789]: I1122 08:44:29.350318 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-qn45p" Nov 22 08:44:29 crc kubenswrapper[4789]: I1122 08:44:29.380670 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65952d06-ccf5-425f-bb4a-ad127a6844df-neutron-metadata-combined-ca-bundle\") pod \"65952d06-ccf5-425f-bb4a-ad127a6844df\" (UID: \"65952d06-ccf5-425f-bb4a-ad127a6844df\") " Nov 22 08:44:29 crc kubenswrapper[4789]: I1122 08:44:29.380742 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/65952d06-ccf5-425f-bb4a-ad127a6844df-inventory\") pod \"65952d06-ccf5-425f-bb4a-ad127a6844df\" (UID: \"65952d06-ccf5-425f-bb4a-ad127a6844df\") " Nov 22 08:44:29 crc kubenswrapper[4789]: I1122 08:44:29.380853 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/65952d06-ccf5-425f-bb4a-ad127a6844df-neutron-ovn-metadata-agent-neutron-config-0\") pod \"65952d06-ccf5-425f-bb4a-ad127a6844df\" (UID: \"65952d06-ccf5-425f-bb4a-ad127a6844df\") " Nov 22 08:44:29 crc kubenswrapper[4789]: I1122 08:44:29.380895 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/65952d06-ccf5-425f-bb4a-ad127a6844df-ssh-key\") pod \"65952d06-ccf5-425f-bb4a-ad127a6844df\" (UID: \"65952d06-ccf5-425f-bb4a-ad127a6844df\") " Nov 22 08:44:29 crc kubenswrapper[4789]: I1122 08:44:29.381003 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c7hhr\" (UniqueName: \"kubernetes.io/projected/65952d06-ccf5-425f-bb4a-ad127a6844df-kube-api-access-c7hhr\") pod \"65952d06-ccf5-425f-bb4a-ad127a6844df\" (UID: \"65952d06-ccf5-425f-bb4a-ad127a6844df\") " Nov 22 08:44:29 crc kubenswrapper[4789]: I1122 08:44:29.381037 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/65952d06-ccf5-425f-bb4a-ad127a6844df-nova-metadata-neutron-config-0\") pod \"65952d06-ccf5-425f-bb4a-ad127a6844df\" (UID: \"65952d06-ccf5-425f-bb4a-ad127a6844df\") " Nov 22 08:44:29 crc kubenswrapper[4789]: I1122 08:44:29.381084 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/65952d06-ccf5-425f-bb4a-ad127a6844df-ceph\") pod \"65952d06-ccf5-425f-bb4a-ad127a6844df\" (UID: \"65952d06-ccf5-425f-bb4a-ad127a6844df\") " Nov 22 08:44:29 crc kubenswrapper[4789]: I1122 08:44:29.387569 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65952d06-ccf5-425f-bb4a-ad127a6844df-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "65952d06-ccf5-425f-bb4a-ad127a6844df" (UID: "65952d06-ccf5-425f-bb4a-ad127a6844df"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:44:29 crc kubenswrapper[4789]: I1122 08:44:29.405533 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65952d06-ccf5-425f-bb4a-ad127a6844df-ceph" (OuterVolumeSpecName: "ceph") pod "65952d06-ccf5-425f-bb4a-ad127a6844df" (UID: "65952d06-ccf5-425f-bb4a-ad127a6844df"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:44:29 crc kubenswrapper[4789]: I1122 08:44:29.406345 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/65952d06-ccf5-425f-bb4a-ad127a6844df-kube-api-access-c7hhr" (OuterVolumeSpecName: "kube-api-access-c7hhr") pod "65952d06-ccf5-425f-bb4a-ad127a6844df" (UID: "65952d06-ccf5-425f-bb4a-ad127a6844df"). InnerVolumeSpecName "kube-api-access-c7hhr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:44:29 crc kubenswrapper[4789]: I1122 08:44:29.415814 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65952d06-ccf5-425f-bb4a-ad127a6844df-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "65952d06-ccf5-425f-bb4a-ad127a6844df" (UID: "65952d06-ccf5-425f-bb4a-ad127a6844df"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:44:29 crc kubenswrapper[4789]: I1122 08:44:29.433227 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65952d06-ccf5-425f-bb4a-ad127a6844df-inventory" (OuterVolumeSpecName: "inventory") pod "65952d06-ccf5-425f-bb4a-ad127a6844df" (UID: "65952d06-ccf5-425f-bb4a-ad127a6844df"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:44:29 crc kubenswrapper[4789]: I1122 08:44:29.438282 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65952d06-ccf5-425f-bb4a-ad127a6844df-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "65952d06-ccf5-425f-bb4a-ad127a6844df" (UID: "65952d06-ccf5-425f-bb4a-ad127a6844df"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:44:29 crc kubenswrapper[4789]: I1122 08:44:29.450894 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65952d06-ccf5-425f-bb4a-ad127a6844df-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "65952d06-ccf5-425f-bb4a-ad127a6844df" (UID: "65952d06-ccf5-425f-bb4a-ad127a6844df"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:44:29 crc kubenswrapper[4789]: I1122 08:44:29.486126 4789 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65952d06-ccf5-425f-bb4a-ad127a6844df-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:44:29 crc kubenswrapper[4789]: I1122 08:44:29.486176 4789 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/65952d06-ccf5-425f-bb4a-ad127a6844df-inventory\") on node \"crc\" DevicePath \"\"" Nov 22 08:44:29 crc kubenswrapper[4789]: I1122 08:44:29.486190 4789 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/65952d06-ccf5-425f-bb4a-ad127a6844df-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Nov 22 08:44:29 crc kubenswrapper[4789]: I1122 08:44:29.486205 4789 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/65952d06-ccf5-425f-bb4a-ad127a6844df-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 22 08:44:29 crc kubenswrapper[4789]: I1122 08:44:29.486217 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c7hhr\" (UniqueName: \"kubernetes.io/projected/65952d06-ccf5-425f-bb4a-ad127a6844df-kube-api-access-c7hhr\") on node \"crc\" DevicePath \"\"" Nov 22 08:44:29 crc kubenswrapper[4789]: I1122 08:44:29.486231 4789 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/65952d06-ccf5-425f-bb4a-ad127a6844df-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Nov 22 08:44:29 crc kubenswrapper[4789]: I1122 08:44:29.486247 4789 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/65952d06-ccf5-425f-bb4a-ad127a6844df-ceph\") on node \"crc\" DevicePath \"\"" Nov 22 08:44:29 crc kubenswrapper[4789]: I1122 08:44:29.492053 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7tj9w" Nov 22 08:44:29 crc kubenswrapper[4789]: I1122 08:44:29.942786 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7tj9w"] Nov 22 08:44:29 crc kubenswrapper[4789]: I1122 08:44:29.952678 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-qn45p" event={"ID":"65952d06-ccf5-425f-bb4a-ad127a6844df","Type":"ContainerDied","Data":"61c7830a9b4f472ee4b737b98e2c0a680fbeb0c75fdd55cf2f14fefc36c47a84"} Nov 22 08:44:29 crc kubenswrapper[4789]: I1122 08:44:29.952715 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="61c7830a9b4f472ee4b737b98e2c0a680fbeb0c75fdd55cf2f14fefc36c47a84" Nov 22 08:44:29 crc kubenswrapper[4789]: I1122 08:44:29.952779 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-qn45p" Nov 22 08:44:30 crc kubenswrapper[4789]: I1122 08:44:30.055635 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-vbx9r"] Nov 22 08:44:30 crc kubenswrapper[4789]: E1122 08:44:30.056004 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65952d06-ccf5-425f-bb4a-ad127a6844df" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Nov 22 08:44:30 crc kubenswrapper[4789]: I1122 08:44:30.056021 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="65952d06-ccf5-425f-bb4a-ad127a6844df" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Nov 22 08:44:30 crc kubenswrapper[4789]: I1122 08:44:30.056389 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="65952d06-ccf5-425f-bb4a-ad127a6844df" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Nov 22 08:44:30 crc kubenswrapper[4789]: I1122 08:44:30.056984 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-vbx9r" Nov 22 08:44:30 crc kubenswrapper[4789]: I1122 08:44:30.062337 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 22 08:44:30 crc kubenswrapper[4789]: I1122 08:44:30.062409 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 22 08:44:30 crc kubenswrapper[4789]: I1122 08:44:30.062432 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 22 08:44:30 crc kubenswrapper[4789]: I1122 08:44:30.062518 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 22 08:44:30 crc kubenswrapper[4789]: I1122 08:44:30.062417 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-glmr7" Nov 22 08:44:30 crc kubenswrapper[4789]: I1122 08:44:30.062641 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Nov 22 08:44:30 crc kubenswrapper[4789]: I1122 08:44:30.073907 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-vbx9r"] Nov 22 08:44:30 crc kubenswrapper[4789]: I1122 08:44:30.098865 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/fd548b7c-689e-4221-ba0f-5c57f8b12927-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-vbx9r\" (UID: \"fd548b7c-689e-4221-ba0f-5c57f8b12927\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-vbx9r" Nov 22 08:44:30 crc kubenswrapper[4789]: I1122 08:44:30.098994 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/fd548b7c-689e-4221-ba0f-5c57f8b12927-ceph\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-vbx9r\" (UID: \"fd548b7c-689e-4221-ba0f-5c57f8b12927\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-vbx9r" Nov 22 08:44:30 crc kubenswrapper[4789]: I1122 08:44:30.099020 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fd548b7c-689e-4221-ba0f-5c57f8b12927-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-vbx9r\" (UID: \"fd548b7c-689e-4221-ba0f-5c57f8b12927\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-vbx9r" Nov 22 08:44:30 crc kubenswrapper[4789]: I1122 08:44:30.099062 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd548b7c-689e-4221-ba0f-5c57f8b12927-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-vbx9r\" (UID: \"fd548b7c-689e-4221-ba0f-5c57f8b12927\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-vbx9r" Nov 22 08:44:30 crc kubenswrapper[4789]: I1122 08:44:30.099093 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fd548b7c-689e-4221-ba0f-5c57f8b12927-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-vbx9r\" (UID: \"fd548b7c-689e-4221-ba0f-5c57f8b12927\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-vbx9r" Nov 22 08:44:30 crc kubenswrapper[4789]: I1122 08:44:30.099133 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9kr2s\" (UniqueName: \"kubernetes.io/projected/fd548b7c-689e-4221-ba0f-5c57f8b12927-kube-api-access-9kr2s\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-vbx9r\" (UID: \"fd548b7c-689e-4221-ba0f-5c57f8b12927\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-vbx9r" Nov 22 08:44:30 crc kubenswrapper[4789]: I1122 08:44:30.200713 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/fd548b7c-689e-4221-ba0f-5c57f8b12927-ceph\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-vbx9r\" (UID: \"fd548b7c-689e-4221-ba0f-5c57f8b12927\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-vbx9r" Nov 22 08:44:30 crc kubenswrapper[4789]: I1122 08:44:30.200773 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fd548b7c-689e-4221-ba0f-5c57f8b12927-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-vbx9r\" (UID: \"fd548b7c-689e-4221-ba0f-5c57f8b12927\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-vbx9r" Nov 22 08:44:30 crc kubenswrapper[4789]: I1122 08:44:30.200800 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd548b7c-689e-4221-ba0f-5c57f8b12927-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-vbx9r\" (UID: \"fd548b7c-689e-4221-ba0f-5c57f8b12927\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-vbx9r" Nov 22 08:44:30 crc kubenswrapper[4789]: I1122 08:44:30.200838 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fd548b7c-689e-4221-ba0f-5c57f8b12927-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-vbx9r\" (UID: \"fd548b7c-689e-4221-ba0f-5c57f8b12927\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-vbx9r" Nov 22 08:44:30 crc kubenswrapper[4789]: I1122 08:44:30.200861 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9kr2s\" (UniqueName: \"kubernetes.io/projected/fd548b7c-689e-4221-ba0f-5c57f8b12927-kube-api-access-9kr2s\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-vbx9r\" (UID: \"fd548b7c-689e-4221-ba0f-5c57f8b12927\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-vbx9r" Nov 22 08:44:30 crc kubenswrapper[4789]: I1122 08:44:30.200930 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/fd548b7c-689e-4221-ba0f-5c57f8b12927-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-vbx9r\" (UID: \"fd548b7c-689e-4221-ba0f-5c57f8b12927\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-vbx9r" Nov 22 08:44:30 crc kubenswrapper[4789]: I1122 08:44:30.206321 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fd548b7c-689e-4221-ba0f-5c57f8b12927-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-vbx9r\" (UID: \"fd548b7c-689e-4221-ba0f-5c57f8b12927\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-vbx9r" Nov 22 08:44:30 crc kubenswrapper[4789]: I1122 08:44:30.206362 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fd548b7c-689e-4221-ba0f-5c57f8b12927-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-vbx9r\" (UID: \"fd548b7c-689e-4221-ba0f-5c57f8b12927\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-vbx9r" Nov 22 08:44:30 crc kubenswrapper[4789]: I1122 08:44:30.206576 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/fd548b7c-689e-4221-ba0f-5c57f8b12927-ceph\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-vbx9r\" (UID: \"fd548b7c-689e-4221-ba0f-5c57f8b12927\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-vbx9r" Nov 22 08:44:30 crc kubenswrapper[4789]: I1122 08:44:30.208578 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/fd548b7c-689e-4221-ba0f-5c57f8b12927-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-vbx9r\" (UID: \"fd548b7c-689e-4221-ba0f-5c57f8b12927\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-vbx9r" Nov 22 08:44:30 crc kubenswrapper[4789]: I1122 08:44:30.209443 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd548b7c-689e-4221-ba0f-5c57f8b12927-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-vbx9r\" (UID: \"fd548b7c-689e-4221-ba0f-5c57f8b12927\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-vbx9r" Nov 22 08:44:30 crc kubenswrapper[4789]: I1122 08:44:30.222462 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9kr2s\" (UniqueName: \"kubernetes.io/projected/fd548b7c-689e-4221-ba0f-5c57f8b12927-kube-api-access-9kr2s\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-vbx9r\" (UID: \"fd548b7c-689e-4221-ba0f-5c57f8b12927\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-vbx9r" Nov 22 08:44:30 crc kubenswrapper[4789]: I1122 08:44:30.386936 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-vbx9r" Nov 22 08:44:30 crc kubenswrapper[4789]: I1122 08:44:30.872433 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-vbx9r"] Nov 22 08:44:30 crc kubenswrapper[4789]: W1122 08:44:30.876420 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfd548b7c_689e_4221_ba0f_5c57f8b12927.slice/crio-e2d9f1090dd6bdf3fcd62ebe50c7481df6f0f51073332a9d8cdb77c0ba4f910f WatchSource:0}: Error finding container e2d9f1090dd6bdf3fcd62ebe50c7481df6f0f51073332a9d8cdb77c0ba4f910f: Status 404 returned error can't find the container with id e2d9f1090dd6bdf3fcd62ebe50c7481df6f0f51073332a9d8cdb77c0ba4f910f Nov 22 08:44:30 crc kubenswrapper[4789]: I1122 08:44:30.964117 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-vbx9r" event={"ID":"fd548b7c-689e-4221-ba0f-5c57f8b12927","Type":"ContainerStarted","Data":"e2d9f1090dd6bdf3fcd62ebe50c7481df6f0f51073332a9d8cdb77c0ba4f910f"} Nov 22 08:44:30 crc kubenswrapper[4789]: I1122 08:44:30.966613 4789 generic.go:334] "Generic (PLEG): container finished" podID="3d54f943-2f4d-424a-8bf7-b523da5eb0f6" containerID="7fabad6c3e939170ff873d0de0d1d9da4910d9f91f5b12dbf73ec897c4afd371" exitCode=0 Nov 22 08:44:30 crc kubenswrapper[4789]: I1122 08:44:30.966657 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7tj9w" event={"ID":"3d54f943-2f4d-424a-8bf7-b523da5eb0f6","Type":"ContainerDied","Data":"7fabad6c3e939170ff873d0de0d1d9da4910d9f91f5b12dbf73ec897c4afd371"} Nov 22 08:44:30 crc kubenswrapper[4789]: I1122 08:44:30.966681 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7tj9w" event={"ID":"3d54f943-2f4d-424a-8bf7-b523da5eb0f6","Type":"ContainerStarted","Data":"4b41edbb4dd9591bf83c2d161b176f02cf1a5878c45ce754af6530fc320aa5df"} Nov 22 08:44:33 crc kubenswrapper[4789]: I1122 08:44:33.991953 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-vbx9r" event={"ID":"fd548b7c-689e-4221-ba0f-5c57f8b12927","Type":"ContainerStarted","Data":"2bd54faf2ecc5a86c3094b03a85a5f3a1e4eb8eab9c65075f3f49928138c007a"} Nov 22 08:44:34 crc kubenswrapper[4789]: I1122 08:44:34.019614 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-vbx9r" podStartSLOduration=2.222263889 podStartE2EDuration="4.019589823s" podCreationTimestamp="2025-11-22 08:44:30 +0000 UTC" firstStartedPulling="2025-11-22 08:44:30.879279061 +0000 UTC m=+2885.113679334" lastFinishedPulling="2025-11-22 08:44:32.676604995 +0000 UTC m=+2886.911005268" observedRunningTime="2025-11-22 08:44:34.009996002 +0000 UTC m=+2888.244396275" watchObservedRunningTime="2025-11-22 08:44:34.019589823 +0000 UTC m=+2888.253990106" Nov 22 08:44:35 crc kubenswrapper[4789]: I1122 08:44:35.372849 4789 patch_prober.go:28] interesting pod/machine-config-daemon-pxvrm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 08:44:35 crc kubenswrapper[4789]: I1122 08:44:35.372943 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 08:44:38 crc kubenswrapper[4789]: I1122 08:44:38.034384 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7tj9w" event={"ID":"3d54f943-2f4d-424a-8bf7-b523da5eb0f6","Type":"ContainerStarted","Data":"c77393e196d7dd0b4b2b381ccb8a23f1599e464705496e70d3f3257ebe9dffbc"} Nov 22 08:44:39 crc kubenswrapper[4789]: I1122 08:44:39.049525 4789 generic.go:334] "Generic (PLEG): container finished" podID="3d54f943-2f4d-424a-8bf7-b523da5eb0f6" containerID="c77393e196d7dd0b4b2b381ccb8a23f1599e464705496e70d3f3257ebe9dffbc" exitCode=0 Nov 22 08:44:39 crc kubenswrapper[4789]: I1122 08:44:39.049596 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7tj9w" event={"ID":"3d54f943-2f4d-424a-8bf7-b523da5eb0f6","Type":"ContainerDied","Data":"c77393e196d7dd0b4b2b381ccb8a23f1599e464705496e70d3f3257ebe9dffbc"} Nov 22 08:44:42 crc kubenswrapper[4789]: I1122 08:44:42.084562 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7tj9w" event={"ID":"3d54f943-2f4d-424a-8bf7-b523da5eb0f6","Type":"ContainerStarted","Data":"c771d3ebe66215b3a95b2b222f7f35e71d7ce149c5477f9f14281be53746ce74"} Nov 22 08:44:42 crc kubenswrapper[4789]: I1122 08:44:42.112481 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-7tj9w" podStartSLOduration=4.570509048 podStartE2EDuration="13.112454454s" podCreationTimestamp="2025-11-22 08:44:29 +0000 UTC" firstStartedPulling="2025-11-22 08:44:32.480777867 +0000 UTC m=+2886.715178140" lastFinishedPulling="2025-11-22 08:44:41.022723283 +0000 UTC m=+2895.257123546" observedRunningTime="2025-11-22 08:44:42.102673859 +0000 UTC m=+2896.337074132" watchObservedRunningTime="2025-11-22 08:44:42.112454454 +0000 UTC m=+2896.346854727" Nov 22 08:44:49 crc kubenswrapper[4789]: I1122 08:44:49.492846 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-7tj9w" Nov 22 08:44:49 crc kubenswrapper[4789]: I1122 08:44:49.493696 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-7tj9w" Nov 22 08:44:49 crc kubenswrapper[4789]: I1122 08:44:49.549964 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-7tj9w" Nov 22 08:44:50 crc kubenswrapper[4789]: I1122 08:44:50.197197 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-7tj9w" Nov 22 08:44:50 crc kubenswrapper[4789]: I1122 08:44:50.252474 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7tj9w"] Nov 22 08:44:52 crc kubenswrapper[4789]: I1122 08:44:52.166628 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-7tj9w" podUID="3d54f943-2f4d-424a-8bf7-b523da5eb0f6" containerName="registry-server" containerID="cri-o://c771d3ebe66215b3a95b2b222f7f35e71d7ce149c5477f9f14281be53746ce74" gracePeriod=2 Nov 22 08:44:53 crc kubenswrapper[4789]: I1122 08:44:53.177339 4789 generic.go:334] "Generic (PLEG): container finished" podID="3d54f943-2f4d-424a-8bf7-b523da5eb0f6" containerID="c771d3ebe66215b3a95b2b222f7f35e71d7ce149c5477f9f14281be53746ce74" exitCode=0 Nov 22 08:44:53 crc kubenswrapper[4789]: I1122 08:44:53.177431 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7tj9w" event={"ID":"3d54f943-2f4d-424a-8bf7-b523da5eb0f6","Type":"ContainerDied","Data":"c771d3ebe66215b3a95b2b222f7f35e71d7ce149c5477f9f14281be53746ce74"} Nov 22 08:44:53 crc kubenswrapper[4789]: I1122 08:44:53.271759 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7tj9w" Nov 22 08:44:53 crc kubenswrapper[4789]: I1122 08:44:53.340930 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d54f943-2f4d-424a-8bf7-b523da5eb0f6-utilities\") pod \"3d54f943-2f4d-424a-8bf7-b523da5eb0f6\" (UID: \"3d54f943-2f4d-424a-8bf7-b523da5eb0f6\") " Nov 22 08:44:53 crc kubenswrapper[4789]: I1122 08:44:53.340993 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d54f943-2f4d-424a-8bf7-b523da5eb0f6-catalog-content\") pod \"3d54f943-2f4d-424a-8bf7-b523da5eb0f6\" (UID: \"3d54f943-2f4d-424a-8bf7-b523da5eb0f6\") " Nov 22 08:44:53 crc kubenswrapper[4789]: I1122 08:44:53.341152 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6xlzq\" (UniqueName: \"kubernetes.io/projected/3d54f943-2f4d-424a-8bf7-b523da5eb0f6-kube-api-access-6xlzq\") pod \"3d54f943-2f4d-424a-8bf7-b523da5eb0f6\" (UID: \"3d54f943-2f4d-424a-8bf7-b523da5eb0f6\") " Nov 22 08:44:53 crc kubenswrapper[4789]: I1122 08:44:53.342126 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3d54f943-2f4d-424a-8bf7-b523da5eb0f6-utilities" (OuterVolumeSpecName: "utilities") pod "3d54f943-2f4d-424a-8bf7-b523da5eb0f6" (UID: "3d54f943-2f4d-424a-8bf7-b523da5eb0f6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:44:53 crc kubenswrapper[4789]: I1122 08:44:53.347295 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d54f943-2f4d-424a-8bf7-b523da5eb0f6-kube-api-access-6xlzq" (OuterVolumeSpecName: "kube-api-access-6xlzq") pod "3d54f943-2f4d-424a-8bf7-b523da5eb0f6" (UID: "3d54f943-2f4d-424a-8bf7-b523da5eb0f6"). InnerVolumeSpecName "kube-api-access-6xlzq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:44:53 crc kubenswrapper[4789]: I1122 08:44:53.442643 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d54f943-2f4d-424a-8bf7-b523da5eb0f6-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 08:44:53 crc kubenswrapper[4789]: I1122 08:44:53.442672 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6xlzq\" (UniqueName: \"kubernetes.io/projected/3d54f943-2f4d-424a-8bf7-b523da5eb0f6-kube-api-access-6xlzq\") on node \"crc\" DevicePath \"\"" Nov 22 08:44:53 crc kubenswrapper[4789]: I1122 08:44:53.449284 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3d54f943-2f4d-424a-8bf7-b523da5eb0f6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3d54f943-2f4d-424a-8bf7-b523da5eb0f6" (UID: "3d54f943-2f4d-424a-8bf7-b523da5eb0f6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:44:53 crc kubenswrapper[4789]: I1122 08:44:53.543615 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d54f943-2f4d-424a-8bf7-b523da5eb0f6-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 08:44:54 crc kubenswrapper[4789]: I1122 08:44:54.189599 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7tj9w" event={"ID":"3d54f943-2f4d-424a-8bf7-b523da5eb0f6","Type":"ContainerDied","Data":"4b41edbb4dd9591bf83c2d161b176f02cf1a5878c45ce754af6530fc320aa5df"} Nov 22 08:44:54 crc kubenswrapper[4789]: I1122 08:44:54.189680 4789 scope.go:117] "RemoveContainer" containerID="c771d3ebe66215b3a95b2b222f7f35e71d7ce149c5477f9f14281be53746ce74" Nov 22 08:44:54 crc kubenswrapper[4789]: I1122 08:44:54.189683 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7tj9w" Nov 22 08:44:54 crc kubenswrapper[4789]: I1122 08:44:54.218057 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7tj9w"] Nov 22 08:44:54 crc kubenswrapper[4789]: I1122 08:44:54.220783 4789 scope.go:117] "RemoveContainer" containerID="c77393e196d7dd0b4b2b381ccb8a23f1599e464705496e70d3f3257ebe9dffbc" Nov 22 08:44:54 crc kubenswrapper[4789]: I1122 08:44:54.231572 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-7tj9w"] Nov 22 08:44:54 crc kubenswrapper[4789]: I1122 08:44:54.241276 4789 scope.go:117] "RemoveContainer" containerID="7fabad6c3e939170ff873d0de0d1d9da4910d9f91f5b12dbf73ec897c4afd371" Nov 22 08:44:55 crc kubenswrapper[4789]: I1122 08:44:55.975502 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3d54f943-2f4d-424a-8bf7-b523da5eb0f6" path="/var/lib/kubelet/pods/3d54f943-2f4d-424a-8bf7-b523da5eb0f6/volumes" Nov 22 08:45:00 crc kubenswrapper[4789]: I1122 08:45:00.174679 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396685-jt9jd"] Nov 22 08:45:00 crc kubenswrapper[4789]: E1122 08:45:00.176510 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d54f943-2f4d-424a-8bf7-b523da5eb0f6" containerName="extract-utilities" Nov 22 08:45:00 crc kubenswrapper[4789]: I1122 08:45:00.176561 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d54f943-2f4d-424a-8bf7-b523da5eb0f6" containerName="extract-utilities" Nov 22 08:45:00 crc kubenswrapper[4789]: E1122 08:45:00.176602 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d54f943-2f4d-424a-8bf7-b523da5eb0f6" containerName="extract-content" Nov 22 08:45:00 crc kubenswrapper[4789]: I1122 08:45:00.176614 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d54f943-2f4d-424a-8bf7-b523da5eb0f6" containerName="extract-content" Nov 22 08:45:00 crc kubenswrapper[4789]: E1122 08:45:00.176649 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d54f943-2f4d-424a-8bf7-b523da5eb0f6" containerName="registry-server" Nov 22 08:45:00 crc kubenswrapper[4789]: I1122 08:45:00.176659 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d54f943-2f4d-424a-8bf7-b523da5eb0f6" containerName="registry-server" Nov 22 08:45:00 crc kubenswrapper[4789]: I1122 08:45:00.176884 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d54f943-2f4d-424a-8bf7-b523da5eb0f6" containerName="registry-server" Nov 22 08:45:00 crc kubenswrapper[4789]: I1122 08:45:00.177659 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396685-jt9jd" Nov 22 08:45:00 crc kubenswrapper[4789]: I1122 08:45:00.180164 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 22 08:45:00 crc kubenswrapper[4789]: I1122 08:45:00.180839 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 22 08:45:00 crc kubenswrapper[4789]: I1122 08:45:00.184653 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396685-jt9jd"] Nov 22 08:45:00 crc kubenswrapper[4789]: I1122 08:45:00.278985 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5sbxb\" (UniqueName: \"kubernetes.io/projected/625a9eb2-dfa9-4b93-b6b9-17402117c6a2-kube-api-access-5sbxb\") pod \"collect-profiles-29396685-jt9jd\" (UID: \"625a9eb2-dfa9-4b93-b6b9-17402117c6a2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396685-jt9jd" Nov 22 08:45:00 crc kubenswrapper[4789]: I1122 08:45:00.279057 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/625a9eb2-dfa9-4b93-b6b9-17402117c6a2-config-volume\") pod \"collect-profiles-29396685-jt9jd\" (UID: \"625a9eb2-dfa9-4b93-b6b9-17402117c6a2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396685-jt9jd" Nov 22 08:45:00 crc kubenswrapper[4789]: I1122 08:45:00.279348 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/625a9eb2-dfa9-4b93-b6b9-17402117c6a2-secret-volume\") pod \"collect-profiles-29396685-jt9jd\" (UID: \"625a9eb2-dfa9-4b93-b6b9-17402117c6a2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396685-jt9jd" Nov 22 08:45:00 crc kubenswrapper[4789]: I1122 08:45:00.380901 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/625a9eb2-dfa9-4b93-b6b9-17402117c6a2-secret-volume\") pod \"collect-profiles-29396685-jt9jd\" (UID: \"625a9eb2-dfa9-4b93-b6b9-17402117c6a2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396685-jt9jd" Nov 22 08:45:00 crc kubenswrapper[4789]: I1122 08:45:00.380989 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5sbxb\" (UniqueName: \"kubernetes.io/projected/625a9eb2-dfa9-4b93-b6b9-17402117c6a2-kube-api-access-5sbxb\") pod \"collect-profiles-29396685-jt9jd\" (UID: \"625a9eb2-dfa9-4b93-b6b9-17402117c6a2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396685-jt9jd" Nov 22 08:45:00 crc kubenswrapper[4789]: I1122 08:45:00.381023 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/625a9eb2-dfa9-4b93-b6b9-17402117c6a2-config-volume\") pod \"collect-profiles-29396685-jt9jd\" (UID: \"625a9eb2-dfa9-4b93-b6b9-17402117c6a2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396685-jt9jd" Nov 22 08:45:00 crc kubenswrapper[4789]: I1122 08:45:00.382040 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/625a9eb2-dfa9-4b93-b6b9-17402117c6a2-config-volume\") pod \"collect-profiles-29396685-jt9jd\" (UID: \"625a9eb2-dfa9-4b93-b6b9-17402117c6a2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396685-jt9jd" Nov 22 08:45:00 crc kubenswrapper[4789]: I1122 08:45:00.388211 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/625a9eb2-dfa9-4b93-b6b9-17402117c6a2-secret-volume\") pod \"collect-profiles-29396685-jt9jd\" (UID: \"625a9eb2-dfa9-4b93-b6b9-17402117c6a2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396685-jt9jd" Nov 22 08:45:00 crc kubenswrapper[4789]: I1122 08:45:00.399293 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5sbxb\" (UniqueName: \"kubernetes.io/projected/625a9eb2-dfa9-4b93-b6b9-17402117c6a2-kube-api-access-5sbxb\") pod \"collect-profiles-29396685-jt9jd\" (UID: \"625a9eb2-dfa9-4b93-b6b9-17402117c6a2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396685-jt9jd" Nov 22 08:45:00 crc kubenswrapper[4789]: I1122 08:45:00.514199 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396685-jt9jd" Nov 22 08:45:00 crc kubenswrapper[4789]: I1122 08:45:00.769557 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-9pbxm"] Nov 22 08:45:00 crc kubenswrapper[4789]: I1122 08:45:00.772721 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9pbxm" Nov 22 08:45:00 crc kubenswrapper[4789]: I1122 08:45:00.777169 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9pbxm"] Nov 22 08:45:00 crc kubenswrapper[4789]: I1122 08:45:00.891367 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9cd271df-d265-459e-8a1a-30473065e990-utilities\") pod \"certified-operators-9pbxm\" (UID: \"9cd271df-d265-459e-8a1a-30473065e990\") " pod="openshift-marketplace/certified-operators-9pbxm" Nov 22 08:45:00 crc kubenswrapper[4789]: I1122 08:45:00.891491 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zx64t\" (UniqueName: \"kubernetes.io/projected/9cd271df-d265-459e-8a1a-30473065e990-kube-api-access-zx64t\") pod \"certified-operators-9pbxm\" (UID: \"9cd271df-d265-459e-8a1a-30473065e990\") " pod="openshift-marketplace/certified-operators-9pbxm" Nov 22 08:45:00 crc kubenswrapper[4789]: I1122 08:45:00.891629 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9cd271df-d265-459e-8a1a-30473065e990-catalog-content\") pod \"certified-operators-9pbxm\" (UID: \"9cd271df-d265-459e-8a1a-30473065e990\") " pod="openshift-marketplace/certified-operators-9pbxm" Nov 22 08:45:00 crc kubenswrapper[4789]: I1122 08:45:00.973352 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396685-jt9jd"] Nov 22 08:45:00 crc kubenswrapper[4789]: I1122 08:45:00.993338 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zx64t\" (UniqueName: \"kubernetes.io/projected/9cd271df-d265-459e-8a1a-30473065e990-kube-api-access-zx64t\") pod \"certified-operators-9pbxm\" (UID: \"9cd271df-d265-459e-8a1a-30473065e990\") " pod="openshift-marketplace/certified-operators-9pbxm" Nov 22 08:45:00 crc kubenswrapper[4789]: I1122 08:45:00.993465 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9cd271df-d265-459e-8a1a-30473065e990-catalog-content\") pod \"certified-operators-9pbxm\" (UID: \"9cd271df-d265-459e-8a1a-30473065e990\") " pod="openshift-marketplace/certified-operators-9pbxm" Nov 22 08:45:00 crc kubenswrapper[4789]: I1122 08:45:00.993651 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9cd271df-d265-459e-8a1a-30473065e990-utilities\") pod \"certified-operators-9pbxm\" (UID: \"9cd271df-d265-459e-8a1a-30473065e990\") " pod="openshift-marketplace/certified-operators-9pbxm" Nov 22 08:45:00 crc kubenswrapper[4789]: I1122 08:45:00.993893 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9cd271df-d265-459e-8a1a-30473065e990-catalog-content\") pod \"certified-operators-9pbxm\" (UID: \"9cd271df-d265-459e-8a1a-30473065e990\") " pod="openshift-marketplace/certified-operators-9pbxm" Nov 22 08:45:00 crc kubenswrapper[4789]: I1122 08:45:00.994157 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9cd271df-d265-459e-8a1a-30473065e990-utilities\") pod \"certified-operators-9pbxm\" (UID: \"9cd271df-d265-459e-8a1a-30473065e990\") " pod="openshift-marketplace/certified-operators-9pbxm" Nov 22 08:45:01 crc kubenswrapper[4789]: I1122 08:45:01.015942 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zx64t\" (UniqueName: \"kubernetes.io/projected/9cd271df-d265-459e-8a1a-30473065e990-kube-api-access-zx64t\") pod \"certified-operators-9pbxm\" (UID: \"9cd271df-d265-459e-8a1a-30473065e990\") " pod="openshift-marketplace/certified-operators-9pbxm" Nov 22 08:45:01 crc kubenswrapper[4789]: I1122 08:45:01.104627 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9pbxm" Nov 22 08:45:01 crc kubenswrapper[4789]: I1122 08:45:01.255793 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29396685-jt9jd" event={"ID":"625a9eb2-dfa9-4b93-b6b9-17402117c6a2","Type":"ContainerStarted","Data":"b83f6dc97ff74bd8b8279ba66bb48c826eea080a4febe93a0c405776737ae8dc"} Nov 22 08:45:01 crc kubenswrapper[4789]: I1122 08:45:01.377261 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9pbxm"] Nov 22 08:45:02 crc kubenswrapper[4789]: I1122 08:45:02.266803 4789 generic.go:334] "Generic (PLEG): container finished" podID="9cd271df-d265-459e-8a1a-30473065e990" containerID="a94ad56161e63849291d7177bc82b70c44b09ac28ee5ce42183d337539f398c2" exitCode=0 Nov 22 08:45:02 crc kubenswrapper[4789]: I1122 08:45:02.266919 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9pbxm" event={"ID":"9cd271df-d265-459e-8a1a-30473065e990","Type":"ContainerDied","Data":"a94ad56161e63849291d7177bc82b70c44b09ac28ee5ce42183d337539f398c2"} Nov 22 08:45:02 crc kubenswrapper[4789]: I1122 08:45:02.267240 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9pbxm" event={"ID":"9cd271df-d265-459e-8a1a-30473065e990","Type":"ContainerStarted","Data":"0327edd12e09fa5e4e59df995429a75c6eafaf6354e9768572eb03f348733b95"} Nov 22 08:45:02 crc kubenswrapper[4789]: I1122 08:45:02.270688 4789 generic.go:334] "Generic (PLEG): container finished" podID="625a9eb2-dfa9-4b93-b6b9-17402117c6a2" containerID="6ce056936c6c9b5253e11a773ae44c4f000ac7e44793b4a172a4c739489b03ac" exitCode=0 Nov 22 08:45:02 crc kubenswrapper[4789]: I1122 08:45:02.270739 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29396685-jt9jd" event={"ID":"625a9eb2-dfa9-4b93-b6b9-17402117c6a2","Type":"ContainerDied","Data":"6ce056936c6c9b5253e11a773ae44c4f000ac7e44793b4a172a4c739489b03ac"} Nov 22 08:45:03 crc kubenswrapper[4789]: I1122 08:45:03.595209 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396685-jt9jd" Nov 22 08:45:03 crc kubenswrapper[4789]: I1122 08:45:03.649375 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5sbxb\" (UniqueName: \"kubernetes.io/projected/625a9eb2-dfa9-4b93-b6b9-17402117c6a2-kube-api-access-5sbxb\") pod \"625a9eb2-dfa9-4b93-b6b9-17402117c6a2\" (UID: \"625a9eb2-dfa9-4b93-b6b9-17402117c6a2\") " Nov 22 08:45:03 crc kubenswrapper[4789]: I1122 08:45:03.649492 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/625a9eb2-dfa9-4b93-b6b9-17402117c6a2-secret-volume\") pod \"625a9eb2-dfa9-4b93-b6b9-17402117c6a2\" (UID: \"625a9eb2-dfa9-4b93-b6b9-17402117c6a2\") " Nov 22 08:45:03 crc kubenswrapper[4789]: I1122 08:45:03.649655 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/625a9eb2-dfa9-4b93-b6b9-17402117c6a2-config-volume\") pod \"625a9eb2-dfa9-4b93-b6b9-17402117c6a2\" (UID: \"625a9eb2-dfa9-4b93-b6b9-17402117c6a2\") " Nov 22 08:45:03 crc kubenswrapper[4789]: I1122 08:45:03.650429 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/625a9eb2-dfa9-4b93-b6b9-17402117c6a2-config-volume" (OuterVolumeSpecName: "config-volume") pod "625a9eb2-dfa9-4b93-b6b9-17402117c6a2" (UID: "625a9eb2-dfa9-4b93-b6b9-17402117c6a2"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:45:03 crc kubenswrapper[4789]: I1122 08:45:03.654889 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/625a9eb2-dfa9-4b93-b6b9-17402117c6a2-kube-api-access-5sbxb" (OuterVolumeSpecName: "kube-api-access-5sbxb") pod "625a9eb2-dfa9-4b93-b6b9-17402117c6a2" (UID: "625a9eb2-dfa9-4b93-b6b9-17402117c6a2"). InnerVolumeSpecName "kube-api-access-5sbxb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:45:03 crc kubenswrapper[4789]: I1122 08:45:03.655103 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/625a9eb2-dfa9-4b93-b6b9-17402117c6a2-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "625a9eb2-dfa9-4b93-b6b9-17402117c6a2" (UID: "625a9eb2-dfa9-4b93-b6b9-17402117c6a2"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:45:03 crc kubenswrapper[4789]: I1122 08:45:03.751372 4789 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/625a9eb2-dfa9-4b93-b6b9-17402117c6a2-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 22 08:45:03 crc kubenswrapper[4789]: I1122 08:45:03.751413 4789 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/625a9eb2-dfa9-4b93-b6b9-17402117c6a2-config-volume\") on node \"crc\" DevicePath \"\"" Nov 22 08:45:03 crc kubenswrapper[4789]: I1122 08:45:03.751423 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5sbxb\" (UniqueName: \"kubernetes.io/projected/625a9eb2-dfa9-4b93-b6b9-17402117c6a2-kube-api-access-5sbxb\") on node \"crc\" DevicePath \"\"" Nov 22 08:45:04 crc kubenswrapper[4789]: I1122 08:45:04.291474 4789 generic.go:334] "Generic (PLEG): container finished" podID="9cd271df-d265-459e-8a1a-30473065e990" containerID="8601dd3b9575970c53d2245610eed5ff6a569dad4876eb3df1975ca6cd81f12d" exitCode=0 Nov 22 08:45:04 crc kubenswrapper[4789]: I1122 08:45:04.291582 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9pbxm" event={"ID":"9cd271df-d265-459e-8a1a-30473065e990","Type":"ContainerDied","Data":"8601dd3b9575970c53d2245610eed5ff6a569dad4876eb3df1975ca6cd81f12d"} Nov 22 08:45:04 crc kubenswrapper[4789]: I1122 08:45:04.297048 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29396685-jt9jd" event={"ID":"625a9eb2-dfa9-4b93-b6b9-17402117c6a2","Type":"ContainerDied","Data":"b83f6dc97ff74bd8b8279ba66bb48c826eea080a4febe93a0c405776737ae8dc"} Nov 22 08:45:04 crc kubenswrapper[4789]: I1122 08:45:04.297118 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b83f6dc97ff74bd8b8279ba66bb48c826eea080a4febe93a0c405776737ae8dc" Nov 22 08:45:04 crc kubenswrapper[4789]: I1122 08:45:04.297182 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396685-jt9jd" Nov 22 08:45:04 crc kubenswrapper[4789]: I1122 08:45:04.666906 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396640-mhslx"] Nov 22 08:45:04 crc kubenswrapper[4789]: I1122 08:45:04.677240 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396640-mhslx"] Nov 22 08:45:05 crc kubenswrapper[4789]: I1122 08:45:05.372406 4789 patch_prober.go:28] interesting pod/machine-config-daemon-pxvrm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 08:45:05 crc kubenswrapper[4789]: I1122 08:45:05.372891 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 08:45:05 crc kubenswrapper[4789]: I1122 08:45:05.979958 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="032913e2-a38e-4c15-b11b-5e47cb60c458" path="/var/lib/kubelet/pods/032913e2-a38e-4c15-b11b-5e47cb60c458/volumes" Nov 22 08:45:06 crc kubenswrapper[4789]: I1122 08:45:06.317335 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9pbxm" event={"ID":"9cd271df-d265-459e-8a1a-30473065e990","Type":"ContainerStarted","Data":"155561bbe1ca5478bc2e2e2e75bb325be77d0fb30797c68ec87d8bf785063d34"} Nov 22 08:45:06 crc kubenswrapper[4789]: I1122 08:45:06.337377 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-9pbxm" podStartSLOduration=3.648166221 podStartE2EDuration="6.337355341s" podCreationTimestamp="2025-11-22 08:45:00 +0000 UTC" firstStartedPulling="2025-11-22 08:45:02.269457498 +0000 UTC m=+2916.503857771" lastFinishedPulling="2025-11-22 08:45:04.958646558 +0000 UTC m=+2919.193046891" observedRunningTime="2025-11-22 08:45:06.333555198 +0000 UTC m=+2920.567955481" watchObservedRunningTime="2025-11-22 08:45:06.337355341 +0000 UTC m=+2920.571755614" Nov 22 08:45:11 crc kubenswrapper[4789]: I1122 08:45:11.105293 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-9pbxm" Nov 22 08:45:11 crc kubenswrapper[4789]: I1122 08:45:11.106051 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-9pbxm" Nov 22 08:45:11 crc kubenswrapper[4789]: I1122 08:45:11.157026 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-9pbxm" Nov 22 08:45:11 crc kubenswrapper[4789]: I1122 08:45:11.405237 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-9pbxm" Nov 22 08:45:11 crc kubenswrapper[4789]: I1122 08:45:11.467018 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9pbxm"] Nov 22 08:45:13 crc kubenswrapper[4789]: I1122 08:45:13.383390 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-9pbxm" podUID="9cd271df-d265-459e-8a1a-30473065e990" containerName="registry-server" containerID="cri-o://155561bbe1ca5478bc2e2e2e75bb325be77d0fb30797c68ec87d8bf785063d34" gracePeriod=2 Nov 22 08:45:13 crc kubenswrapper[4789]: I1122 08:45:13.906483 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9pbxm" Nov 22 08:45:13 crc kubenswrapper[4789]: I1122 08:45:13.937432 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zx64t\" (UniqueName: \"kubernetes.io/projected/9cd271df-d265-459e-8a1a-30473065e990-kube-api-access-zx64t\") pod \"9cd271df-d265-459e-8a1a-30473065e990\" (UID: \"9cd271df-d265-459e-8a1a-30473065e990\") " Nov 22 08:45:13 crc kubenswrapper[4789]: I1122 08:45:13.938138 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9cd271df-d265-459e-8a1a-30473065e990-utilities\") pod \"9cd271df-d265-459e-8a1a-30473065e990\" (UID: \"9cd271df-d265-459e-8a1a-30473065e990\") " Nov 22 08:45:13 crc kubenswrapper[4789]: I1122 08:45:13.938364 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9cd271df-d265-459e-8a1a-30473065e990-catalog-content\") pod \"9cd271df-d265-459e-8a1a-30473065e990\" (UID: \"9cd271df-d265-459e-8a1a-30473065e990\") " Nov 22 08:45:13 crc kubenswrapper[4789]: I1122 08:45:13.939082 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9cd271df-d265-459e-8a1a-30473065e990-utilities" (OuterVolumeSpecName: "utilities") pod "9cd271df-d265-459e-8a1a-30473065e990" (UID: "9cd271df-d265-459e-8a1a-30473065e990"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:45:13 crc kubenswrapper[4789]: I1122 08:45:13.947160 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9cd271df-d265-459e-8a1a-30473065e990-kube-api-access-zx64t" (OuterVolumeSpecName: "kube-api-access-zx64t") pod "9cd271df-d265-459e-8a1a-30473065e990" (UID: "9cd271df-d265-459e-8a1a-30473065e990"). InnerVolumeSpecName "kube-api-access-zx64t". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:45:14 crc kubenswrapper[4789]: I1122 08:45:14.003504 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9cd271df-d265-459e-8a1a-30473065e990-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9cd271df-d265-459e-8a1a-30473065e990" (UID: "9cd271df-d265-459e-8a1a-30473065e990"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:45:14 crc kubenswrapper[4789]: I1122 08:45:14.040662 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9cd271df-d265-459e-8a1a-30473065e990-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 08:45:14 crc kubenswrapper[4789]: I1122 08:45:14.040696 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9cd271df-d265-459e-8a1a-30473065e990-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 08:45:14 crc kubenswrapper[4789]: I1122 08:45:14.040706 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zx64t\" (UniqueName: \"kubernetes.io/projected/9cd271df-d265-459e-8a1a-30473065e990-kube-api-access-zx64t\") on node \"crc\" DevicePath \"\"" Nov 22 08:45:14 crc kubenswrapper[4789]: I1122 08:45:14.393864 4789 generic.go:334] "Generic (PLEG): container finished" podID="9cd271df-d265-459e-8a1a-30473065e990" containerID="155561bbe1ca5478bc2e2e2e75bb325be77d0fb30797c68ec87d8bf785063d34" exitCode=0 Nov 22 08:45:14 crc kubenswrapper[4789]: I1122 08:45:14.393927 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9pbxm" event={"ID":"9cd271df-d265-459e-8a1a-30473065e990","Type":"ContainerDied","Data":"155561bbe1ca5478bc2e2e2e75bb325be77d0fb30797c68ec87d8bf785063d34"} Nov 22 08:45:14 crc kubenswrapper[4789]: I1122 08:45:14.393948 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9pbxm" Nov 22 08:45:14 crc kubenswrapper[4789]: I1122 08:45:14.393980 4789 scope.go:117] "RemoveContainer" containerID="155561bbe1ca5478bc2e2e2e75bb325be77d0fb30797c68ec87d8bf785063d34" Nov 22 08:45:14 crc kubenswrapper[4789]: I1122 08:45:14.393964 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9pbxm" event={"ID":"9cd271df-d265-459e-8a1a-30473065e990","Type":"ContainerDied","Data":"0327edd12e09fa5e4e59df995429a75c6eafaf6354e9768572eb03f348733b95"} Nov 22 08:45:14 crc kubenswrapper[4789]: I1122 08:45:14.418023 4789 scope.go:117] "RemoveContainer" containerID="8601dd3b9575970c53d2245610eed5ff6a569dad4876eb3df1975ca6cd81f12d" Nov 22 08:45:14 crc kubenswrapper[4789]: I1122 08:45:14.442721 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9pbxm"] Nov 22 08:45:14 crc kubenswrapper[4789]: I1122 08:45:14.452411 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-9pbxm"] Nov 22 08:45:14 crc kubenswrapper[4789]: I1122 08:45:14.458975 4789 scope.go:117] "RemoveContainer" containerID="a94ad56161e63849291d7177bc82b70c44b09ac28ee5ce42183d337539f398c2" Nov 22 08:45:14 crc kubenswrapper[4789]: I1122 08:45:14.485878 4789 scope.go:117] "RemoveContainer" containerID="155561bbe1ca5478bc2e2e2e75bb325be77d0fb30797c68ec87d8bf785063d34" Nov 22 08:45:14 crc kubenswrapper[4789]: E1122 08:45:14.486877 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"155561bbe1ca5478bc2e2e2e75bb325be77d0fb30797c68ec87d8bf785063d34\": container with ID starting with 155561bbe1ca5478bc2e2e2e75bb325be77d0fb30797c68ec87d8bf785063d34 not found: ID does not exist" containerID="155561bbe1ca5478bc2e2e2e75bb325be77d0fb30797c68ec87d8bf785063d34" Nov 22 08:45:14 crc kubenswrapper[4789]: I1122 08:45:14.486923 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"155561bbe1ca5478bc2e2e2e75bb325be77d0fb30797c68ec87d8bf785063d34"} err="failed to get container status \"155561bbe1ca5478bc2e2e2e75bb325be77d0fb30797c68ec87d8bf785063d34\": rpc error: code = NotFound desc = could not find container \"155561bbe1ca5478bc2e2e2e75bb325be77d0fb30797c68ec87d8bf785063d34\": container with ID starting with 155561bbe1ca5478bc2e2e2e75bb325be77d0fb30797c68ec87d8bf785063d34 not found: ID does not exist" Nov 22 08:45:14 crc kubenswrapper[4789]: I1122 08:45:14.487022 4789 scope.go:117] "RemoveContainer" containerID="8601dd3b9575970c53d2245610eed5ff6a569dad4876eb3df1975ca6cd81f12d" Nov 22 08:45:14 crc kubenswrapper[4789]: E1122 08:45:14.487490 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8601dd3b9575970c53d2245610eed5ff6a569dad4876eb3df1975ca6cd81f12d\": container with ID starting with 8601dd3b9575970c53d2245610eed5ff6a569dad4876eb3df1975ca6cd81f12d not found: ID does not exist" containerID="8601dd3b9575970c53d2245610eed5ff6a569dad4876eb3df1975ca6cd81f12d" Nov 22 08:45:14 crc kubenswrapper[4789]: I1122 08:45:14.487523 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8601dd3b9575970c53d2245610eed5ff6a569dad4876eb3df1975ca6cd81f12d"} err="failed to get container status \"8601dd3b9575970c53d2245610eed5ff6a569dad4876eb3df1975ca6cd81f12d\": rpc error: code = NotFound desc = could not find container \"8601dd3b9575970c53d2245610eed5ff6a569dad4876eb3df1975ca6cd81f12d\": container with ID starting with 8601dd3b9575970c53d2245610eed5ff6a569dad4876eb3df1975ca6cd81f12d not found: ID does not exist" Nov 22 08:45:14 crc kubenswrapper[4789]: I1122 08:45:14.487545 4789 scope.go:117] "RemoveContainer" containerID="a94ad56161e63849291d7177bc82b70c44b09ac28ee5ce42183d337539f398c2" Nov 22 08:45:14 crc kubenswrapper[4789]: E1122 08:45:14.487821 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a94ad56161e63849291d7177bc82b70c44b09ac28ee5ce42183d337539f398c2\": container with ID starting with a94ad56161e63849291d7177bc82b70c44b09ac28ee5ce42183d337539f398c2 not found: ID does not exist" containerID="a94ad56161e63849291d7177bc82b70c44b09ac28ee5ce42183d337539f398c2" Nov 22 08:45:14 crc kubenswrapper[4789]: I1122 08:45:14.487848 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a94ad56161e63849291d7177bc82b70c44b09ac28ee5ce42183d337539f398c2"} err="failed to get container status \"a94ad56161e63849291d7177bc82b70c44b09ac28ee5ce42183d337539f398c2\": rpc error: code = NotFound desc = could not find container \"a94ad56161e63849291d7177bc82b70c44b09ac28ee5ce42183d337539f398c2\": container with ID starting with a94ad56161e63849291d7177bc82b70c44b09ac28ee5ce42183d337539f398c2 not found: ID does not exist" Nov 22 08:45:15 crc kubenswrapper[4789]: I1122 08:45:15.975013 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9cd271df-d265-459e-8a1a-30473065e990" path="/var/lib/kubelet/pods/9cd271df-d265-459e-8a1a-30473065e990/volumes" Nov 22 08:45:27 crc kubenswrapper[4789]: I1122 08:45:27.552199 4789 scope.go:117] "RemoveContainer" containerID="71f0fe50a32dfe9aeda16a4708c8a0cf7f62844adc58aa09c8a8444050c78770" Nov 22 08:45:35 crc kubenswrapper[4789]: I1122 08:45:35.371841 4789 patch_prober.go:28] interesting pod/machine-config-daemon-pxvrm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 08:45:35 crc kubenswrapper[4789]: I1122 08:45:35.372434 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 08:45:35 crc kubenswrapper[4789]: I1122 08:45:35.372534 4789 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" Nov 22 08:45:35 crc kubenswrapper[4789]: I1122 08:45:35.373548 4789 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e6fcfa0e8fd700efa0433d598c23ce626ca7a007b2b45b45ee839e17923bad02"} pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 22 08:45:35 crc kubenswrapper[4789]: I1122 08:45:35.373643 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" containerID="cri-o://e6fcfa0e8fd700efa0433d598c23ce626ca7a007b2b45b45ee839e17923bad02" gracePeriod=600 Nov 22 08:45:36 crc kubenswrapper[4789]: E1122 08:45:36.196497 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 08:45:36 crc kubenswrapper[4789]: I1122 08:45:36.596457 4789 generic.go:334] "Generic (PLEG): container finished" podID="58aa8071-7ecc-4692-8789-2db04ec70647" containerID="e6fcfa0e8fd700efa0433d598c23ce626ca7a007b2b45b45ee839e17923bad02" exitCode=0 Nov 22 08:45:36 crc kubenswrapper[4789]: I1122 08:45:36.596538 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" event={"ID":"58aa8071-7ecc-4692-8789-2db04ec70647","Type":"ContainerDied","Data":"e6fcfa0e8fd700efa0433d598c23ce626ca7a007b2b45b45ee839e17923bad02"} Nov 22 08:45:36 crc kubenswrapper[4789]: I1122 08:45:36.596587 4789 scope.go:117] "RemoveContainer" containerID="95ee0f2b0a62b96c30cc7400004c7e7b084559a4811441f88de89dadf3d054d1" Nov 22 08:45:36 crc kubenswrapper[4789]: I1122 08:45:36.597445 4789 scope.go:117] "RemoveContainer" containerID="e6fcfa0e8fd700efa0433d598c23ce626ca7a007b2b45b45ee839e17923bad02" Nov 22 08:45:36 crc kubenswrapper[4789]: E1122 08:45:36.597813 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 08:45:47 crc kubenswrapper[4789]: I1122 08:45:47.971388 4789 scope.go:117] "RemoveContainer" containerID="e6fcfa0e8fd700efa0433d598c23ce626ca7a007b2b45b45ee839e17923bad02" Nov 22 08:45:47 crc kubenswrapper[4789]: E1122 08:45:47.972623 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 08:45:58 crc kubenswrapper[4789]: I1122 08:45:58.965459 4789 scope.go:117] "RemoveContainer" containerID="e6fcfa0e8fd700efa0433d598c23ce626ca7a007b2b45b45ee839e17923bad02" Nov 22 08:45:58 crc kubenswrapper[4789]: E1122 08:45:58.966588 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 08:46:09 crc kubenswrapper[4789]: I1122 08:46:09.965306 4789 scope.go:117] "RemoveContainer" containerID="e6fcfa0e8fd700efa0433d598c23ce626ca7a007b2b45b45ee839e17923bad02" Nov 22 08:46:10 crc kubenswrapper[4789]: E1122 08:46:09.966106 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 08:46:23 crc kubenswrapper[4789]: I1122 08:46:23.965157 4789 scope.go:117] "RemoveContainer" containerID="e6fcfa0e8fd700efa0433d598c23ce626ca7a007b2b45b45ee839e17923bad02" Nov 22 08:46:23 crc kubenswrapper[4789]: E1122 08:46:23.966206 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 08:46:35 crc kubenswrapper[4789]: I1122 08:46:35.966108 4789 scope.go:117] "RemoveContainer" containerID="e6fcfa0e8fd700efa0433d598c23ce626ca7a007b2b45b45ee839e17923bad02" Nov 22 08:46:35 crc kubenswrapper[4789]: E1122 08:46:35.967085 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 08:46:49 crc kubenswrapper[4789]: I1122 08:46:49.966214 4789 scope.go:117] "RemoveContainer" containerID="e6fcfa0e8fd700efa0433d598c23ce626ca7a007b2b45b45ee839e17923bad02" Nov 22 08:46:49 crc kubenswrapper[4789]: E1122 08:46:49.967021 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 08:47:03 crc kubenswrapper[4789]: I1122 08:47:03.965768 4789 scope.go:117] "RemoveContainer" containerID="e6fcfa0e8fd700efa0433d598c23ce626ca7a007b2b45b45ee839e17923bad02" Nov 22 08:47:03 crc kubenswrapper[4789]: E1122 08:47:03.966587 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 08:47:18 crc kubenswrapper[4789]: I1122 08:47:18.964958 4789 scope.go:117] "RemoveContainer" containerID="e6fcfa0e8fd700efa0433d598c23ce626ca7a007b2b45b45ee839e17923bad02" Nov 22 08:47:18 crc kubenswrapper[4789]: E1122 08:47:18.966007 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 08:47:32 crc kubenswrapper[4789]: I1122 08:47:32.965561 4789 scope.go:117] "RemoveContainer" containerID="e6fcfa0e8fd700efa0433d598c23ce626ca7a007b2b45b45ee839e17923bad02" Nov 22 08:47:32 crc kubenswrapper[4789]: E1122 08:47:32.966465 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 08:47:47 crc kubenswrapper[4789]: I1122 08:47:47.974238 4789 scope.go:117] "RemoveContainer" containerID="e6fcfa0e8fd700efa0433d598c23ce626ca7a007b2b45b45ee839e17923bad02" Nov 22 08:47:47 crc kubenswrapper[4789]: E1122 08:47:47.975063 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 08:48:01 crc kubenswrapper[4789]: I1122 08:48:01.965428 4789 scope.go:117] "RemoveContainer" containerID="e6fcfa0e8fd700efa0433d598c23ce626ca7a007b2b45b45ee839e17923bad02" Nov 22 08:48:01 crc kubenswrapper[4789]: E1122 08:48:01.966274 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 08:48:14 crc kubenswrapper[4789]: I1122 08:48:14.965056 4789 scope.go:117] "RemoveContainer" containerID="e6fcfa0e8fd700efa0433d598c23ce626ca7a007b2b45b45ee839e17923bad02" Nov 22 08:48:14 crc kubenswrapper[4789]: E1122 08:48:14.967479 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 08:48:27 crc kubenswrapper[4789]: I1122 08:48:27.971381 4789 scope.go:117] "RemoveContainer" containerID="e6fcfa0e8fd700efa0433d598c23ce626ca7a007b2b45b45ee839e17923bad02" Nov 22 08:48:27 crc kubenswrapper[4789]: E1122 08:48:27.972229 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 08:48:39 crc kubenswrapper[4789]: I1122 08:48:39.965933 4789 scope.go:117] "RemoveContainer" containerID="e6fcfa0e8fd700efa0433d598c23ce626ca7a007b2b45b45ee839e17923bad02" Nov 22 08:48:39 crc kubenswrapper[4789]: E1122 08:48:39.966821 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 08:48:50 crc kubenswrapper[4789]: I1122 08:48:50.965360 4789 scope.go:117] "RemoveContainer" containerID="e6fcfa0e8fd700efa0433d598c23ce626ca7a007b2b45b45ee839e17923bad02" Nov 22 08:48:50 crc kubenswrapper[4789]: E1122 08:48:50.966201 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 08:49:03 crc kubenswrapper[4789]: I1122 08:49:03.965665 4789 scope.go:117] "RemoveContainer" containerID="e6fcfa0e8fd700efa0433d598c23ce626ca7a007b2b45b45ee839e17923bad02" Nov 22 08:49:03 crc kubenswrapper[4789]: E1122 08:49:03.966464 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 08:49:14 crc kubenswrapper[4789]: I1122 08:49:14.965714 4789 scope.go:117] "RemoveContainer" containerID="e6fcfa0e8fd700efa0433d598c23ce626ca7a007b2b45b45ee839e17923bad02" Nov 22 08:49:14 crc kubenswrapper[4789]: E1122 08:49:14.967057 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 08:49:28 crc kubenswrapper[4789]: I1122 08:49:28.965837 4789 scope.go:117] "RemoveContainer" containerID="e6fcfa0e8fd700efa0433d598c23ce626ca7a007b2b45b45ee839e17923bad02" Nov 22 08:49:28 crc kubenswrapper[4789]: E1122 08:49:28.966712 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 08:49:33 crc kubenswrapper[4789]: I1122 08:49:33.828788 4789 generic.go:334] "Generic (PLEG): container finished" podID="fd548b7c-689e-4221-ba0f-5c57f8b12927" containerID="2bd54faf2ecc5a86c3094b03a85a5f3a1e4eb8eab9c65075f3f49928138c007a" exitCode=0 Nov 22 08:49:33 crc kubenswrapper[4789]: I1122 08:49:33.828904 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-vbx9r" event={"ID":"fd548b7c-689e-4221-ba0f-5c57f8b12927","Type":"ContainerDied","Data":"2bd54faf2ecc5a86c3094b03a85a5f3a1e4eb8eab9c65075f3f49928138c007a"} Nov 22 08:49:35 crc kubenswrapper[4789]: I1122 08:49:35.224957 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-vbx9r" Nov 22 08:49:35 crc kubenswrapper[4789]: I1122 08:49:35.379931 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fd548b7c-689e-4221-ba0f-5c57f8b12927-inventory\") pod \"fd548b7c-689e-4221-ba0f-5c57f8b12927\" (UID: \"fd548b7c-689e-4221-ba0f-5c57f8b12927\") " Nov 22 08:49:35 crc kubenswrapper[4789]: I1122 08:49:35.380045 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9kr2s\" (UniqueName: \"kubernetes.io/projected/fd548b7c-689e-4221-ba0f-5c57f8b12927-kube-api-access-9kr2s\") pod \"fd548b7c-689e-4221-ba0f-5c57f8b12927\" (UID: \"fd548b7c-689e-4221-ba0f-5c57f8b12927\") " Nov 22 08:49:35 crc kubenswrapper[4789]: I1122 08:49:35.380130 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/fd548b7c-689e-4221-ba0f-5c57f8b12927-libvirt-secret-0\") pod \"fd548b7c-689e-4221-ba0f-5c57f8b12927\" (UID: \"fd548b7c-689e-4221-ba0f-5c57f8b12927\") " Nov 22 08:49:35 crc kubenswrapper[4789]: I1122 08:49:35.380213 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fd548b7c-689e-4221-ba0f-5c57f8b12927-ssh-key\") pod \"fd548b7c-689e-4221-ba0f-5c57f8b12927\" (UID: \"fd548b7c-689e-4221-ba0f-5c57f8b12927\") " Nov 22 08:49:35 crc kubenswrapper[4789]: I1122 08:49:35.380238 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd548b7c-689e-4221-ba0f-5c57f8b12927-libvirt-combined-ca-bundle\") pod \"fd548b7c-689e-4221-ba0f-5c57f8b12927\" (UID: \"fd548b7c-689e-4221-ba0f-5c57f8b12927\") " Nov 22 08:49:35 crc kubenswrapper[4789]: I1122 08:49:35.380313 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/fd548b7c-689e-4221-ba0f-5c57f8b12927-ceph\") pod \"fd548b7c-689e-4221-ba0f-5c57f8b12927\" (UID: \"fd548b7c-689e-4221-ba0f-5c57f8b12927\") " Nov 22 08:49:35 crc kubenswrapper[4789]: I1122 08:49:35.387227 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd548b7c-689e-4221-ba0f-5c57f8b12927-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "fd548b7c-689e-4221-ba0f-5c57f8b12927" (UID: "fd548b7c-689e-4221-ba0f-5c57f8b12927"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:49:35 crc kubenswrapper[4789]: I1122 08:49:35.388003 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd548b7c-689e-4221-ba0f-5c57f8b12927-kube-api-access-9kr2s" (OuterVolumeSpecName: "kube-api-access-9kr2s") pod "fd548b7c-689e-4221-ba0f-5c57f8b12927" (UID: "fd548b7c-689e-4221-ba0f-5c57f8b12927"). InnerVolumeSpecName "kube-api-access-9kr2s". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:49:35 crc kubenswrapper[4789]: I1122 08:49:35.388739 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd548b7c-689e-4221-ba0f-5c57f8b12927-ceph" (OuterVolumeSpecName: "ceph") pod "fd548b7c-689e-4221-ba0f-5c57f8b12927" (UID: "fd548b7c-689e-4221-ba0f-5c57f8b12927"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:49:35 crc kubenswrapper[4789]: I1122 08:49:35.409971 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd548b7c-689e-4221-ba0f-5c57f8b12927-inventory" (OuterVolumeSpecName: "inventory") pod "fd548b7c-689e-4221-ba0f-5c57f8b12927" (UID: "fd548b7c-689e-4221-ba0f-5c57f8b12927"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:49:35 crc kubenswrapper[4789]: I1122 08:49:35.411009 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd548b7c-689e-4221-ba0f-5c57f8b12927-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "fd548b7c-689e-4221-ba0f-5c57f8b12927" (UID: "fd548b7c-689e-4221-ba0f-5c57f8b12927"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:49:35 crc kubenswrapper[4789]: I1122 08:49:35.413516 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd548b7c-689e-4221-ba0f-5c57f8b12927-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "fd548b7c-689e-4221-ba0f-5c57f8b12927" (UID: "fd548b7c-689e-4221-ba0f-5c57f8b12927"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:49:35 crc kubenswrapper[4789]: I1122 08:49:35.482221 4789 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/fd548b7c-689e-4221-ba0f-5c57f8b12927-ceph\") on node \"crc\" DevicePath \"\"" Nov 22 08:49:35 crc kubenswrapper[4789]: I1122 08:49:35.482266 4789 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fd548b7c-689e-4221-ba0f-5c57f8b12927-inventory\") on node \"crc\" DevicePath \"\"" Nov 22 08:49:35 crc kubenswrapper[4789]: I1122 08:49:35.482280 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9kr2s\" (UniqueName: \"kubernetes.io/projected/fd548b7c-689e-4221-ba0f-5c57f8b12927-kube-api-access-9kr2s\") on node \"crc\" DevicePath \"\"" Nov 22 08:49:35 crc kubenswrapper[4789]: I1122 08:49:35.482294 4789 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/fd548b7c-689e-4221-ba0f-5c57f8b12927-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Nov 22 08:49:35 crc kubenswrapper[4789]: I1122 08:49:35.482306 4789 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fd548b7c-689e-4221-ba0f-5c57f8b12927-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 22 08:49:35 crc kubenswrapper[4789]: I1122 08:49:35.482317 4789 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd548b7c-689e-4221-ba0f-5c57f8b12927-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:49:35 crc kubenswrapper[4789]: I1122 08:49:35.846915 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-vbx9r" Nov 22 08:49:35 crc kubenswrapper[4789]: I1122 08:49:35.846931 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-vbx9r" event={"ID":"fd548b7c-689e-4221-ba0f-5c57f8b12927","Type":"ContainerDied","Data":"e2d9f1090dd6bdf3fcd62ebe50c7481df6f0f51073332a9d8cdb77c0ba4f910f"} Nov 22 08:49:35 crc kubenswrapper[4789]: I1122 08:49:35.846990 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e2d9f1090dd6bdf3fcd62ebe50c7481df6f0f51073332a9d8cdb77c0ba4f910f" Nov 22 08:49:35 crc kubenswrapper[4789]: I1122 08:49:35.934932 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b47tx"] Nov 22 08:49:35 crc kubenswrapper[4789]: E1122 08:49:35.935514 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cd271df-d265-459e-8a1a-30473065e990" containerName="extract-utilities" Nov 22 08:49:35 crc kubenswrapper[4789]: I1122 08:49:35.935627 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cd271df-d265-459e-8a1a-30473065e990" containerName="extract-utilities" Nov 22 08:49:35 crc kubenswrapper[4789]: E1122 08:49:35.935697 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cd271df-d265-459e-8a1a-30473065e990" containerName="registry-server" Nov 22 08:49:35 crc kubenswrapper[4789]: I1122 08:49:35.935778 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cd271df-d265-459e-8a1a-30473065e990" containerName="registry-server" Nov 22 08:49:35 crc kubenswrapper[4789]: E1122 08:49:35.935843 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd548b7c-689e-4221-ba0f-5c57f8b12927" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Nov 22 08:49:35 crc kubenswrapper[4789]: I1122 08:49:35.935895 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd548b7c-689e-4221-ba0f-5c57f8b12927" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Nov 22 08:49:35 crc kubenswrapper[4789]: E1122 08:49:35.935957 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cd271df-d265-459e-8a1a-30473065e990" containerName="extract-content" Nov 22 08:49:35 crc kubenswrapper[4789]: I1122 08:49:35.936014 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cd271df-d265-459e-8a1a-30473065e990" containerName="extract-content" Nov 22 08:49:35 crc kubenswrapper[4789]: E1122 08:49:35.936092 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="625a9eb2-dfa9-4b93-b6b9-17402117c6a2" containerName="collect-profiles" Nov 22 08:49:35 crc kubenswrapper[4789]: I1122 08:49:35.936143 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="625a9eb2-dfa9-4b93-b6b9-17402117c6a2" containerName="collect-profiles" Nov 22 08:49:35 crc kubenswrapper[4789]: I1122 08:49:35.936355 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="9cd271df-d265-459e-8a1a-30473065e990" containerName="registry-server" Nov 22 08:49:35 crc kubenswrapper[4789]: I1122 08:49:35.936420 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd548b7c-689e-4221-ba0f-5c57f8b12927" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Nov 22 08:49:35 crc kubenswrapper[4789]: I1122 08:49:35.936484 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="625a9eb2-dfa9-4b93-b6b9-17402117c6a2" containerName="collect-profiles" Nov 22 08:49:35 crc kubenswrapper[4789]: I1122 08:49:35.937189 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b47tx" Nov 22 08:49:35 crc kubenswrapper[4789]: I1122 08:49:35.940166 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 22 08:49:35 crc kubenswrapper[4789]: I1122 08:49:35.940681 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 22 08:49:35 crc kubenswrapper[4789]: I1122 08:49:35.940924 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 22 08:49:35 crc kubenswrapper[4789]: I1122 08:49:35.941084 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Nov 22 08:49:35 crc kubenswrapper[4789]: I1122 08:49:35.941499 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Nov 22 08:49:35 crc kubenswrapper[4789]: I1122 08:49:35.942525 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ceph-nova" Nov 22 08:49:35 crc kubenswrapper[4789]: I1122 08:49:35.942744 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Nov 22 08:49:35 crc kubenswrapper[4789]: I1122 08:49:35.943099 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 22 08:49:35 crc kubenswrapper[4789]: I1122 08:49:35.944518 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-glmr7" Nov 22 08:49:35 crc kubenswrapper[4789]: I1122 08:49:35.947000 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b47tx"] Nov 22 08:49:36 crc kubenswrapper[4789]: I1122 08:49:36.092931 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/2c231e91-12f1-47e4-8947-1bf03bcae704-ceph-nova-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b47tx\" (UID: \"2c231e91-12f1-47e4-8947-1bf03bcae704\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b47tx" Nov 22 08:49:36 crc kubenswrapper[4789]: I1122 08:49:36.092999 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/2c231e91-12f1-47e4-8947-1bf03bcae704-nova-extra-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b47tx\" (UID: \"2c231e91-12f1-47e4-8947-1bf03bcae704\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b47tx" Nov 22 08:49:36 crc kubenswrapper[4789]: I1122 08:49:36.093125 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/2c231e91-12f1-47e4-8947-1bf03bcae704-nova-cell1-compute-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b47tx\" (UID: \"2c231e91-12f1-47e4-8947-1bf03bcae704\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b47tx" Nov 22 08:49:36 crc kubenswrapper[4789]: I1122 08:49:36.093188 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xzm9h\" (UniqueName: \"kubernetes.io/projected/2c231e91-12f1-47e4-8947-1bf03bcae704-kube-api-access-xzm9h\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b47tx\" (UID: \"2c231e91-12f1-47e4-8947-1bf03bcae704\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b47tx" Nov 22 08:49:36 crc kubenswrapper[4789]: I1122 08:49:36.093216 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2c231e91-12f1-47e4-8947-1bf03bcae704-ssh-key\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b47tx\" (UID: \"2c231e91-12f1-47e4-8947-1bf03bcae704\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b47tx" Nov 22 08:49:36 crc kubenswrapper[4789]: I1122 08:49:36.093311 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c231e91-12f1-47e4-8947-1bf03bcae704-nova-custom-ceph-combined-ca-bundle\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b47tx\" (UID: \"2c231e91-12f1-47e4-8947-1bf03bcae704\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b47tx" Nov 22 08:49:36 crc kubenswrapper[4789]: I1122 08:49:36.093375 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/2c231e91-12f1-47e4-8947-1bf03bcae704-nova-cell1-compute-config-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b47tx\" (UID: \"2c231e91-12f1-47e4-8947-1bf03bcae704\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b47tx" Nov 22 08:49:36 crc kubenswrapper[4789]: I1122 08:49:36.093461 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2c231e91-12f1-47e4-8947-1bf03bcae704-inventory\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b47tx\" (UID: \"2c231e91-12f1-47e4-8947-1bf03bcae704\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b47tx" Nov 22 08:49:36 crc kubenswrapper[4789]: I1122 08:49:36.093513 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/2c231e91-12f1-47e4-8947-1bf03bcae704-nova-migration-ssh-key-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b47tx\" (UID: \"2c231e91-12f1-47e4-8947-1bf03bcae704\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b47tx" Nov 22 08:49:36 crc kubenswrapper[4789]: I1122 08:49:36.093563 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/2c231e91-12f1-47e4-8947-1bf03bcae704-nova-migration-ssh-key-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b47tx\" (UID: \"2c231e91-12f1-47e4-8947-1bf03bcae704\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b47tx" Nov 22 08:49:36 crc kubenswrapper[4789]: I1122 08:49:36.093641 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2c231e91-12f1-47e4-8947-1bf03bcae704-ceph\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b47tx\" (UID: \"2c231e91-12f1-47e4-8947-1bf03bcae704\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b47tx" Nov 22 08:49:36 crc kubenswrapper[4789]: I1122 08:49:36.196069 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/2c231e91-12f1-47e4-8947-1bf03bcae704-ceph-nova-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b47tx\" (UID: \"2c231e91-12f1-47e4-8947-1bf03bcae704\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b47tx" Nov 22 08:49:36 crc kubenswrapper[4789]: I1122 08:49:36.196151 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/2c231e91-12f1-47e4-8947-1bf03bcae704-nova-extra-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b47tx\" (UID: \"2c231e91-12f1-47e4-8947-1bf03bcae704\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b47tx" Nov 22 08:49:36 crc kubenswrapper[4789]: I1122 08:49:36.196194 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/2c231e91-12f1-47e4-8947-1bf03bcae704-nova-cell1-compute-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b47tx\" (UID: \"2c231e91-12f1-47e4-8947-1bf03bcae704\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b47tx" Nov 22 08:49:36 crc kubenswrapper[4789]: I1122 08:49:36.196222 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xzm9h\" (UniqueName: \"kubernetes.io/projected/2c231e91-12f1-47e4-8947-1bf03bcae704-kube-api-access-xzm9h\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b47tx\" (UID: \"2c231e91-12f1-47e4-8947-1bf03bcae704\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b47tx" Nov 22 08:49:36 crc kubenswrapper[4789]: I1122 08:49:36.196246 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2c231e91-12f1-47e4-8947-1bf03bcae704-ssh-key\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b47tx\" (UID: \"2c231e91-12f1-47e4-8947-1bf03bcae704\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b47tx" Nov 22 08:49:36 crc kubenswrapper[4789]: I1122 08:49:36.196292 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c231e91-12f1-47e4-8947-1bf03bcae704-nova-custom-ceph-combined-ca-bundle\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b47tx\" (UID: \"2c231e91-12f1-47e4-8947-1bf03bcae704\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b47tx" Nov 22 08:49:36 crc kubenswrapper[4789]: I1122 08:49:36.196332 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/2c231e91-12f1-47e4-8947-1bf03bcae704-nova-cell1-compute-config-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b47tx\" (UID: \"2c231e91-12f1-47e4-8947-1bf03bcae704\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b47tx" Nov 22 08:49:36 crc kubenswrapper[4789]: I1122 08:49:36.196367 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2c231e91-12f1-47e4-8947-1bf03bcae704-inventory\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b47tx\" (UID: \"2c231e91-12f1-47e4-8947-1bf03bcae704\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b47tx" Nov 22 08:49:36 crc kubenswrapper[4789]: I1122 08:49:36.196403 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/2c231e91-12f1-47e4-8947-1bf03bcae704-nova-migration-ssh-key-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b47tx\" (UID: \"2c231e91-12f1-47e4-8947-1bf03bcae704\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b47tx" Nov 22 08:49:36 crc kubenswrapper[4789]: I1122 08:49:36.197214 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/2c231e91-12f1-47e4-8947-1bf03bcae704-nova-migration-ssh-key-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b47tx\" (UID: \"2c231e91-12f1-47e4-8947-1bf03bcae704\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b47tx" Nov 22 08:49:36 crc kubenswrapper[4789]: I1122 08:49:36.197236 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/2c231e91-12f1-47e4-8947-1bf03bcae704-nova-extra-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b47tx\" (UID: \"2c231e91-12f1-47e4-8947-1bf03bcae704\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b47tx" Nov 22 08:49:36 crc kubenswrapper[4789]: I1122 08:49:36.197278 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2c231e91-12f1-47e4-8947-1bf03bcae704-ceph\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b47tx\" (UID: \"2c231e91-12f1-47e4-8947-1bf03bcae704\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b47tx" Nov 22 08:49:36 crc kubenswrapper[4789]: I1122 08:49:36.198257 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/2c231e91-12f1-47e4-8947-1bf03bcae704-ceph-nova-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b47tx\" (UID: \"2c231e91-12f1-47e4-8947-1bf03bcae704\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b47tx" Nov 22 08:49:36 crc kubenswrapper[4789]: I1122 08:49:36.201506 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/2c231e91-12f1-47e4-8947-1bf03bcae704-nova-cell1-compute-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b47tx\" (UID: \"2c231e91-12f1-47e4-8947-1bf03bcae704\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b47tx" Nov 22 08:49:36 crc kubenswrapper[4789]: I1122 08:49:36.201621 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2c231e91-12f1-47e4-8947-1bf03bcae704-ssh-key\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b47tx\" (UID: \"2c231e91-12f1-47e4-8947-1bf03bcae704\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b47tx" Nov 22 08:49:36 crc kubenswrapper[4789]: I1122 08:49:36.201876 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c231e91-12f1-47e4-8947-1bf03bcae704-nova-custom-ceph-combined-ca-bundle\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b47tx\" (UID: \"2c231e91-12f1-47e4-8947-1bf03bcae704\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b47tx" Nov 22 08:49:36 crc kubenswrapper[4789]: I1122 08:49:36.202611 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/2c231e91-12f1-47e4-8947-1bf03bcae704-nova-migration-ssh-key-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b47tx\" (UID: \"2c231e91-12f1-47e4-8947-1bf03bcae704\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b47tx" Nov 22 08:49:36 crc kubenswrapper[4789]: I1122 08:49:36.203102 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/2c231e91-12f1-47e4-8947-1bf03bcae704-nova-migration-ssh-key-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b47tx\" (UID: \"2c231e91-12f1-47e4-8947-1bf03bcae704\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b47tx" Nov 22 08:49:36 crc kubenswrapper[4789]: I1122 08:49:36.204030 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2c231e91-12f1-47e4-8947-1bf03bcae704-inventory\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b47tx\" (UID: \"2c231e91-12f1-47e4-8947-1bf03bcae704\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b47tx" Nov 22 08:49:36 crc kubenswrapper[4789]: I1122 08:49:36.204038 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2c231e91-12f1-47e4-8947-1bf03bcae704-ceph\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b47tx\" (UID: \"2c231e91-12f1-47e4-8947-1bf03bcae704\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b47tx" Nov 22 08:49:36 crc kubenswrapper[4789]: I1122 08:49:36.205214 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/2c231e91-12f1-47e4-8947-1bf03bcae704-nova-cell1-compute-config-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b47tx\" (UID: \"2c231e91-12f1-47e4-8947-1bf03bcae704\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b47tx" Nov 22 08:49:36 crc kubenswrapper[4789]: I1122 08:49:36.213902 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xzm9h\" (UniqueName: \"kubernetes.io/projected/2c231e91-12f1-47e4-8947-1bf03bcae704-kube-api-access-xzm9h\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b47tx\" (UID: \"2c231e91-12f1-47e4-8947-1bf03bcae704\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b47tx" Nov 22 08:49:36 crc kubenswrapper[4789]: I1122 08:49:36.256728 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b47tx" Nov 22 08:49:36 crc kubenswrapper[4789]: I1122 08:49:36.820827 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b47tx"] Nov 22 08:49:36 crc kubenswrapper[4789]: I1122 08:49:36.825792 4789 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 22 08:49:36 crc kubenswrapper[4789]: I1122 08:49:36.859850 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b47tx" event={"ID":"2c231e91-12f1-47e4-8947-1bf03bcae704","Type":"ContainerStarted","Data":"55cfd923b3d8343569a7116305dd23df3aebd9cf05fd463a23613a771f7c5c00"} Nov 22 08:49:38 crc kubenswrapper[4789]: I1122 08:49:38.896658 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b47tx" event={"ID":"2c231e91-12f1-47e4-8947-1bf03bcae704","Type":"ContainerStarted","Data":"10e60a23d3a242d246e6ca5a07b7cad8d65dc2f0d72f533b5609d93e5c4ee559"} Nov 22 08:49:38 crc kubenswrapper[4789]: I1122 08:49:38.917527 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b47tx" podStartSLOduration=2.973157137 podStartE2EDuration="3.917506448s" podCreationTimestamp="2025-11-22 08:49:35 +0000 UTC" firstStartedPulling="2025-11-22 08:49:36.825565827 +0000 UTC m=+3191.059966100" lastFinishedPulling="2025-11-22 08:49:37.769915138 +0000 UTC m=+3192.004315411" observedRunningTime="2025-11-22 08:49:38.914290571 +0000 UTC m=+3193.148690884" watchObservedRunningTime="2025-11-22 08:49:38.917506448 +0000 UTC m=+3193.151906721" Nov 22 08:49:42 crc kubenswrapper[4789]: I1122 08:49:42.965435 4789 scope.go:117] "RemoveContainer" containerID="e6fcfa0e8fd700efa0433d598c23ce626ca7a007b2b45b45ee839e17923bad02" Nov 22 08:49:42 crc kubenswrapper[4789]: E1122 08:49:42.966255 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 08:49:53 crc kubenswrapper[4789]: I1122 08:49:53.965971 4789 scope.go:117] "RemoveContainer" containerID="e6fcfa0e8fd700efa0433d598c23ce626ca7a007b2b45b45ee839e17923bad02" Nov 22 08:49:53 crc kubenswrapper[4789]: E1122 08:49:53.966712 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 08:50:05 crc kubenswrapper[4789]: I1122 08:50:05.965617 4789 scope.go:117] "RemoveContainer" containerID="e6fcfa0e8fd700efa0433d598c23ce626ca7a007b2b45b45ee839e17923bad02" Nov 22 08:50:05 crc kubenswrapper[4789]: E1122 08:50:05.966431 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 08:50:18 crc kubenswrapper[4789]: I1122 08:50:18.965977 4789 scope.go:117] "RemoveContainer" containerID="e6fcfa0e8fd700efa0433d598c23ce626ca7a007b2b45b45ee839e17923bad02" Nov 22 08:50:18 crc kubenswrapper[4789]: E1122 08:50:18.966745 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 08:50:27 crc kubenswrapper[4789]: I1122 08:50:27.801447 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="b3e0f7a4-cde7-4da6-8339-615054807f46" containerName="ceilometer-central-agent" probeResult="failure" output="command timed out" Nov 22 08:50:27 crc kubenswrapper[4789]: I1122 08:50:27.801462 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="b3e0f7a4-cde7-4da6-8339-615054807f46" containerName="ceilometer-notification-agent" probeResult="failure" output="command timed out" Nov 22 08:50:28 crc kubenswrapper[4789]: I1122 08:50:28.609662 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="b3e0f7a4-cde7-4da6-8339-615054807f46" containerName="ceilometer-central-agent" probeResult="failure" output=< Nov 22 08:50:28 crc kubenswrapper[4789]: Unkown error: Expecting value: line 1 column 1 (char 0) Nov 22 08:50:28 crc kubenswrapper[4789]: > Nov 22 08:50:32 crc kubenswrapper[4789]: I1122 08:50:32.598536 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="b3e0f7a4-cde7-4da6-8339-615054807f46" containerName="ceilometer-central-agent" probeResult="failure" output=< Nov 22 08:50:32 crc kubenswrapper[4789]: Unkown error: Expecting value: line 1 column 1 (char 0) Nov 22 08:50:32 crc kubenswrapper[4789]: > Nov 22 08:50:32 crc kubenswrapper[4789]: I1122 08:50:32.599738 4789 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/ceilometer-0" Nov 22 08:50:32 crc kubenswrapper[4789]: I1122 08:50:32.601345 4789 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="ceilometer-central-agent" containerStatusID={"Type":"cri-o","ID":"a3569e8f73bbb94022be853ca4eca0f31e32f0be828470dce1bedafa7097f19c"} pod="openstack/ceilometer-0" containerMessage="Container ceilometer-central-agent failed liveness probe, will be restarted" Nov 22 08:50:32 crc kubenswrapper[4789]: I1122 08:50:32.601487 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b3e0f7a4-cde7-4da6-8339-615054807f46" containerName="ceilometer-central-agent" containerID="cri-o://a3569e8f73bbb94022be853ca4eca0f31e32f0be828470dce1bedafa7097f19c" gracePeriod=30 Nov 22 08:50:33 crc kubenswrapper[4789]: I1122 08:50:33.363838 4789 generic.go:334] "Generic (PLEG): container finished" podID="b3e0f7a4-cde7-4da6-8339-615054807f46" containerID="a3569e8f73bbb94022be853ca4eca0f31e32f0be828470dce1bedafa7097f19c" exitCode=0 Nov 22 08:50:33 crc kubenswrapper[4789]: I1122 08:50:33.363924 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b3e0f7a4-cde7-4da6-8339-615054807f46","Type":"ContainerDied","Data":"a3569e8f73bbb94022be853ca4eca0f31e32f0be828470dce1bedafa7097f19c"} Nov 22 08:50:33 crc kubenswrapper[4789]: I1122 08:50:33.965694 4789 scope.go:117] "RemoveContainer" containerID="e6fcfa0e8fd700efa0433d598c23ce626ca7a007b2b45b45ee839e17923bad02" Nov 22 08:50:33 crc kubenswrapper[4789]: E1122 08:50:33.966248 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 08:50:36 crc kubenswrapper[4789]: I1122 08:50:36.398065 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b3e0f7a4-cde7-4da6-8339-615054807f46","Type":"ContainerStarted","Data":"57bc1e114749de45fd56a9c528a09c54ebfb4ff2b2cda822e7b1411b31acac80"} Nov 22 08:50:45 crc kubenswrapper[4789]: I1122 08:50:45.965262 4789 scope.go:117] "RemoveContainer" containerID="e6fcfa0e8fd700efa0433d598c23ce626ca7a007b2b45b45ee839e17923bad02" Nov 22 08:50:46 crc kubenswrapper[4789]: I1122 08:50:46.485228 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" event={"ID":"58aa8071-7ecc-4692-8789-2db04ec70647","Type":"ContainerStarted","Data":"2e2387a756e8b093bdd9ccbfb31650bf2886cf3ee1392229517961fc3889046d"} Nov 22 08:52:24 crc kubenswrapper[4789]: I1122 08:52:24.277290 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-hs28n"] Nov 22 08:52:24 crc kubenswrapper[4789]: I1122 08:52:24.280032 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hs28n" Nov 22 08:52:24 crc kubenswrapper[4789]: I1122 08:52:24.287658 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hs28n"] Nov 22 08:52:24 crc kubenswrapper[4789]: I1122 08:52:24.448261 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/43964ece-d130-4c53-a61e-ecf30c9d62f1-catalog-content\") pod \"community-operators-hs28n\" (UID: \"43964ece-d130-4c53-a61e-ecf30c9d62f1\") " pod="openshift-marketplace/community-operators-hs28n" Nov 22 08:52:24 crc kubenswrapper[4789]: I1122 08:52:24.448385 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/43964ece-d130-4c53-a61e-ecf30c9d62f1-utilities\") pod \"community-operators-hs28n\" (UID: \"43964ece-d130-4c53-a61e-ecf30c9d62f1\") " pod="openshift-marketplace/community-operators-hs28n" Nov 22 08:52:24 crc kubenswrapper[4789]: I1122 08:52:24.448599 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xh8t6\" (UniqueName: \"kubernetes.io/projected/43964ece-d130-4c53-a61e-ecf30c9d62f1-kube-api-access-xh8t6\") pod \"community-operators-hs28n\" (UID: \"43964ece-d130-4c53-a61e-ecf30c9d62f1\") " pod="openshift-marketplace/community-operators-hs28n" Nov 22 08:52:24 crc kubenswrapper[4789]: I1122 08:52:24.550324 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/43964ece-d130-4c53-a61e-ecf30c9d62f1-catalog-content\") pod \"community-operators-hs28n\" (UID: \"43964ece-d130-4c53-a61e-ecf30c9d62f1\") " pod="openshift-marketplace/community-operators-hs28n" Nov 22 08:52:24 crc kubenswrapper[4789]: I1122 08:52:24.550410 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/43964ece-d130-4c53-a61e-ecf30c9d62f1-utilities\") pod \"community-operators-hs28n\" (UID: \"43964ece-d130-4c53-a61e-ecf30c9d62f1\") " pod="openshift-marketplace/community-operators-hs28n" Nov 22 08:52:24 crc kubenswrapper[4789]: I1122 08:52:24.550442 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xh8t6\" (UniqueName: \"kubernetes.io/projected/43964ece-d130-4c53-a61e-ecf30c9d62f1-kube-api-access-xh8t6\") pod \"community-operators-hs28n\" (UID: \"43964ece-d130-4c53-a61e-ecf30c9d62f1\") " pod="openshift-marketplace/community-operators-hs28n" Nov 22 08:52:24 crc kubenswrapper[4789]: I1122 08:52:24.550890 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/43964ece-d130-4c53-a61e-ecf30c9d62f1-catalog-content\") pod \"community-operators-hs28n\" (UID: \"43964ece-d130-4c53-a61e-ecf30c9d62f1\") " pod="openshift-marketplace/community-operators-hs28n" Nov 22 08:52:24 crc kubenswrapper[4789]: I1122 08:52:24.550944 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/43964ece-d130-4c53-a61e-ecf30c9d62f1-utilities\") pod \"community-operators-hs28n\" (UID: \"43964ece-d130-4c53-a61e-ecf30c9d62f1\") " pod="openshift-marketplace/community-operators-hs28n" Nov 22 08:52:24 crc kubenswrapper[4789]: I1122 08:52:24.574502 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xh8t6\" (UniqueName: \"kubernetes.io/projected/43964ece-d130-4c53-a61e-ecf30c9d62f1-kube-api-access-xh8t6\") pod \"community-operators-hs28n\" (UID: \"43964ece-d130-4c53-a61e-ecf30c9d62f1\") " pod="openshift-marketplace/community-operators-hs28n" Nov 22 08:52:24 crc kubenswrapper[4789]: I1122 08:52:24.602082 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hs28n" Nov 22 08:52:25 crc kubenswrapper[4789]: I1122 08:52:25.135085 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hs28n"] Nov 22 08:52:25 crc kubenswrapper[4789]: W1122 08:52:25.138330 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod43964ece_d130_4c53_a61e_ecf30c9d62f1.slice/crio-4c81001d5b86252b430e32ca0264b6848bfada46751c7d5b566511b6d11fb1c0 WatchSource:0}: Error finding container 4c81001d5b86252b430e32ca0264b6848bfada46751c7d5b566511b6d11fb1c0: Status 404 returned error can't find the container with id 4c81001d5b86252b430e32ca0264b6848bfada46751c7d5b566511b6d11fb1c0 Nov 22 08:52:25 crc kubenswrapper[4789]: I1122 08:52:25.317818 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hs28n" event={"ID":"43964ece-d130-4c53-a61e-ecf30c9d62f1","Type":"ContainerStarted","Data":"4c81001d5b86252b430e32ca0264b6848bfada46751c7d5b566511b6d11fb1c0"} Nov 22 08:52:26 crc kubenswrapper[4789]: I1122 08:52:26.329481 4789 generic.go:334] "Generic (PLEG): container finished" podID="43964ece-d130-4c53-a61e-ecf30c9d62f1" containerID="8b70ed0796b5b84fcce1fce4ce7d68c983efb0d81c74e2b5957c8a888e0acdbd" exitCode=0 Nov 22 08:52:26 crc kubenswrapper[4789]: I1122 08:52:26.329618 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hs28n" event={"ID":"43964ece-d130-4c53-a61e-ecf30c9d62f1","Type":"ContainerDied","Data":"8b70ed0796b5b84fcce1fce4ce7d68c983efb0d81c74e2b5957c8a888e0acdbd"} Nov 22 08:52:29 crc kubenswrapper[4789]: I1122 08:52:29.355228 4789 generic.go:334] "Generic (PLEG): container finished" podID="43964ece-d130-4c53-a61e-ecf30c9d62f1" containerID="278b664b212ead0cb707759a357289ad71d0828a606bd9d751bb0f94b1ed8db0" exitCode=0 Nov 22 08:52:29 crc kubenswrapper[4789]: I1122 08:52:29.355296 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hs28n" event={"ID":"43964ece-d130-4c53-a61e-ecf30c9d62f1","Type":"ContainerDied","Data":"278b664b212ead0cb707759a357289ad71d0828a606bd9d751bb0f94b1ed8db0"} Nov 22 08:52:29 crc kubenswrapper[4789]: I1122 08:52:29.657412 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-nhd9t"] Nov 22 08:52:29 crc kubenswrapper[4789]: I1122 08:52:29.659433 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nhd9t" Nov 22 08:52:29 crc kubenswrapper[4789]: I1122 08:52:29.675264 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nhd9t"] Nov 22 08:52:29 crc kubenswrapper[4789]: I1122 08:52:29.739399 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4548f47d-11cc-4762-bc6f-22d31fc2202e-utilities\") pod \"redhat-marketplace-nhd9t\" (UID: \"4548f47d-11cc-4762-bc6f-22d31fc2202e\") " pod="openshift-marketplace/redhat-marketplace-nhd9t" Nov 22 08:52:29 crc kubenswrapper[4789]: I1122 08:52:29.739488 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n6fv7\" (UniqueName: \"kubernetes.io/projected/4548f47d-11cc-4762-bc6f-22d31fc2202e-kube-api-access-n6fv7\") pod \"redhat-marketplace-nhd9t\" (UID: \"4548f47d-11cc-4762-bc6f-22d31fc2202e\") " pod="openshift-marketplace/redhat-marketplace-nhd9t" Nov 22 08:52:29 crc kubenswrapper[4789]: I1122 08:52:29.739637 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4548f47d-11cc-4762-bc6f-22d31fc2202e-catalog-content\") pod \"redhat-marketplace-nhd9t\" (UID: \"4548f47d-11cc-4762-bc6f-22d31fc2202e\") " pod="openshift-marketplace/redhat-marketplace-nhd9t" Nov 22 08:52:29 crc kubenswrapper[4789]: I1122 08:52:29.841836 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4548f47d-11cc-4762-bc6f-22d31fc2202e-catalog-content\") pod \"redhat-marketplace-nhd9t\" (UID: \"4548f47d-11cc-4762-bc6f-22d31fc2202e\") " pod="openshift-marketplace/redhat-marketplace-nhd9t" Nov 22 08:52:29 crc kubenswrapper[4789]: I1122 08:52:29.841967 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4548f47d-11cc-4762-bc6f-22d31fc2202e-utilities\") pod \"redhat-marketplace-nhd9t\" (UID: \"4548f47d-11cc-4762-bc6f-22d31fc2202e\") " pod="openshift-marketplace/redhat-marketplace-nhd9t" Nov 22 08:52:29 crc kubenswrapper[4789]: I1122 08:52:29.842082 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n6fv7\" (UniqueName: \"kubernetes.io/projected/4548f47d-11cc-4762-bc6f-22d31fc2202e-kube-api-access-n6fv7\") pod \"redhat-marketplace-nhd9t\" (UID: \"4548f47d-11cc-4762-bc6f-22d31fc2202e\") " pod="openshift-marketplace/redhat-marketplace-nhd9t" Nov 22 08:52:29 crc kubenswrapper[4789]: I1122 08:52:29.842412 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4548f47d-11cc-4762-bc6f-22d31fc2202e-catalog-content\") pod \"redhat-marketplace-nhd9t\" (UID: \"4548f47d-11cc-4762-bc6f-22d31fc2202e\") " pod="openshift-marketplace/redhat-marketplace-nhd9t" Nov 22 08:52:29 crc kubenswrapper[4789]: I1122 08:52:29.842573 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4548f47d-11cc-4762-bc6f-22d31fc2202e-utilities\") pod \"redhat-marketplace-nhd9t\" (UID: \"4548f47d-11cc-4762-bc6f-22d31fc2202e\") " pod="openshift-marketplace/redhat-marketplace-nhd9t" Nov 22 08:52:29 crc kubenswrapper[4789]: I1122 08:52:29.866565 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n6fv7\" (UniqueName: \"kubernetes.io/projected/4548f47d-11cc-4762-bc6f-22d31fc2202e-kube-api-access-n6fv7\") pod \"redhat-marketplace-nhd9t\" (UID: \"4548f47d-11cc-4762-bc6f-22d31fc2202e\") " pod="openshift-marketplace/redhat-marketplace-nhd9t" Nov 22 08:52:29 crc kubenswrapper[4789]: I1122 08:52:29.981023 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nhd9t" Nov 22 08:52:30 crc kubenswrapper[4789]: I1122 08:52:30.578648 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nhd9t"] Nov 22 08:52:30 crc kubenswrapper[4789]: W1122 08:52:30.579986 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4548f47d_11cc_4762_bc6f_22d31fc2202e.slice/crio-e8d878216f7ee10536e7c80aa7acb94cb5a870cbaacc01dca45610f42b0c4a1d WatchSource:0}: Error finding container e8d878216f7ee10536e7c80aa7acb94cb5a870cbaacc01dca45610f42b0c4a1d: Status 404 returned error can't find the container with id e8d878216f7ee10536e7c80aa7acb94cb5a870cbaacc01dca45610f42b0c4a1d Nov 22 08:52:31 crc kubenswrapper[4789]: I1122 08:52:31.374710 4789 generic.go:334] "Generic (PLEG): container finished" podID="4548f47d-11cc-4762-bc6f-22d31fc2202e" containerID="daaa4db34dda9555b56b0188d8c71a7c46a73b0e95de9a0ad1647ce042964af8" exitCode=0 Nov 22 08:52:31 crc kubenswrapper[4789]: I1122 08:52:31.374814 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nhd9t" event={"ID":"4548f47d-11cc-4762-bc6f-22d31fc2202e","Type":"ContainerDied","Data":"daaa4db34dda9555b56b0188d8c71a7c46a73b0e95de9a0ad1647ce042964af8"} Nov 22 08:52:31 crc kubenswrapper[4789]: I1122 08:52:31.375085 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nhd9t" event={"ID":"4548f47d-11cc-4762-bc6f-22d31fc2202e","Type":"ContainerStarted","Data":"e8d878216f7ee10536e7c80aa7acb94cb5a870cbaacc01dca45610f42b0c4a1d"} Nov 22 08:52:32 crc kubenswrapper[4789]: I1122 08:52:32.389245 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hs28n" event={"ID":"43964ece-d130-4c53-a61e-ecf30c9d62f1","Type":"ContainerStarted","Data":"1838642f044e8e25b0e5e07950a95a97b656ce186532b2a638485f92fba7cf8f"} Nov 22 08:52:32 crc kubenswrapper[4789]: I1122 08:52:32.408413 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-hs28n" podStartSLOduration=3.507783421 podStartE2EDuration="8.408393144s" podCreationTimestamp="2025-11-22 08:52:24 +0000 UTC" firstStartedPulling="2025-11-22 08:52:26.332154272 +0000 UTC m=+3360.566554545" lastFinishedPulling="2025-11-22 08:52:31.232763995 +0000 UTC m=+3365.467164268" observedRunningTime="2025-11-22 08:52:32.404701534 +0000 UTC m=+3366.639101817" watchObservedRunningTime="2025-11-22 08:52:32.408393144 +0000 UTC m=+3366.642793417" Nov 22 08:52:33 crc kubenswrapper[4789]: I1122 08:52:33.400974 4789 generic.go:334] "Generic (PLEG): container finished" podID="4548f47d-11cc-4762-bc6f-22d31fc2202e" containerID="05c2302c8972fe10df8022dd43fa9505ee24a11c31e22357d7f01de6050d4267" exitCode=0 Nov 22 08:52:33 crc kubenswrapper[4789]: I1122 08:52:33.401331 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nhd9t" event={"ID":"4548f47d-11cc-4762-bc6f-22d31fc2202e","Type":"ContainerDied","Data":"05c2302c8972fe10df8022dd43fa9505ee24a11c31e22357d7f01de6050d4267"} Nov 22 08:52:34 crc kubenswrapper[4789]: I1122 08:52:34.413951 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nhd9t" event={"ID":"4548f47d-11cc-4762-bc6f-22d31fc2202e","Type":"ContainerStarted","Data":"94d0a627619c8be698ef3052f7c956528b57bbe2b00f2d05e2736ab5b652913f"} Nov 22 08:52:34 crc kubenswrapper[4789]: I1122 08:52:34.443642 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-nhd9t" podStartSLOduration=2.920830743 podStartE2EDuration="5.44361644s" podCreationTimestamp="2025-11-22 08:52:29 +0000 UTC" firstStartedPulling="2025-11-22 08:52:31.376501955 +0000 UTC m=+3365.610902228" lastFinishedPulling="2025-11-22 08:52:33.899287642 +0000 UTC m=+3368.133687925" observedRunningTime="2025-11-22 08:52:34.430821694 +0000 UTC m=+3368.665221967" watchObservedRunningTime="2025-11-22 08:52:34.44361644 +0000 UTC m=+3368.678016733" Nov 22 08:52:34 crc kubenswrapper[4789]: I1122 08:52:34.602918 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-hs28n" Nov 22 08:52:34 crc kubenswrapper[4789]: I1122 08:52:34.602994 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-hs28n" Nov 22 08:52:34 crc kubenswrapper[4789]: I1122 08:52:34.650302 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-hs28n" Nov 22 08:52:39 crc kubenswrapper[4789]: I1122 08:52:39.981431 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-nhd9t" Nov 22 08:52:39 crc kubenswrapper[4789]: I1122 08:52:39.981967 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-nhd9t" Nov 22 08:52:40 crc kubenswrapper[4789]: I1122 08:52:40.029747 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-nhd9t" Nov 22 08:52:40 crc kubenswrapper[4789]: I1122 08:52:40.507296 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-nhd9t" Nov 22 08:52:40 crc kubenswrapper[4789]: I1122 08:52:40.566844 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nhd9t"] Nov 22 08:52:42 crc kubenswrapper[4789]: I1122 08:52:42.480950 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-nhd9t" podUID="4548f47d-11cc-4762-bc6f-22d31fc2202e" containerName="registry-server" containerID="cri-o://94d0a627619c8be698ef3052f7c956528b57bbe2b00f2d05e2736ab5b652913f" gracePeriod=2 Nov 22 08:52:42 crc kubenswrapper[4789]: I1122 08:52:42.901140 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nhd9t" Nov 22 08:52:42 crc kubenswrapper[4789]: I1122 08:52:42.989924 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n6fv7\" (UniqueName: \"kubernetes.io/projected/4548f47d-11cc-4762-bc6f-22d31fc2202e-kube-api-access-n6fv7\") pod \"4548f47d-11cc-4762-bc6f-22d31fc2202e\" (UID: \"4548f47d-11cc-4762-bc6f-22d31fc2202e\") " Nov 22 08:52:42 crc kubenswrapper[4789]: I1122 08:52:42.990049 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4548f47d-11cc-4762-bc6f-22d31fc2202e-utilities\") pod \"4548f47d-11cc-4762-bc6f-22d31fc2202e\" (UID: \"4548f47d-11cc-4762-bc6f-22d31fc2202e\") " Nov 22 08:52:42 crc kubenswrapper[4789]: I1122 08:52:42.990164 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4548f47d-11cc-4762-bc6f-22d31fc2202e-catalog-content\") pod \"4548f47d-11cc-4762-bc6f-22d31fc2202e\" (UID: \"4548f47d-11cc-4762-bc6f-22d31fc2202e\") " Nov 22 08:52:42 crc kubenswrapper[4789]: I1122 08:52:42.991326 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4548f47d-11cc-4762-bc6f-22d31fc2202e-utilities" (OuterVolumeSpecName: "utilities") pod "4548f47d-11cc-4762-bc6f-22d31fc2202e" (UID: "4548f47d-11cc-4762-bc6f-22d31fc2202e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:52:42 crc kubenswrapper[4789]: I1122 08:52:42.997232 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4548f47d-11cc-4762-bc6f-22d31fc2202e-kube-api-access-n6fv7" (OuterVolumeSpecName: "kube-api-access-n6fv7") pod "4548f47d-11cc-4762-bc6f-22d31fc2202e" (UID: "4548f47d-11cc-4762-bc6f-22d31fc2202e"). InnerVolumeSpecName "kube-api-access-n6fv7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:52:43 crc kubenswrapper[4789]: I1122 08:52:43.010365 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4548f47d-11cc-4762-bc6f-22d31fc2202e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4548f47d-11cc-4762-bc6f-22d31fc2202e" (UID: "4548f47d-11cc-4762-bc6f-22d31fc2202e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:52:43 crc kubenswrapper[4789]: I1122 08:52:43.092645 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4548f47d-11cc-4762-bc6f-22d31fc2202e-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 08:52:43 crc kubenswrapper[4789]: I1122 08:52:43.092681 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4548f47d-11cc-4762-bc6f-22d31fc2202e-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 08:52:43 crc kubenswrapper[4789]: I1122 08:52:43.092697 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n6fv7\" (UniqueName: \"kubernetes.io/projected/4548f47d-11cc-4762-bc6f-22d31fc2202e-kube-api-access-n6fv7\") on node \"crc\" DevicePath \"\"" Nov 22 08:52:43 crc kubenswrapper[4789]: I1122 08:52:43.491106 4789 generic.go:334] "Generic (PLEG): container finished" podID="4548f47d-11cc-4762-bc6f-22d31fc2202e" containerID="94d0a627619c8be698ef3052f7c956528b57bbe2b00f2d05e2736ab5b652913f" exitCode=0 Nov 22 08:52:43 crc kubenswrapper[4789]: I1122 08:52:43.491159 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nhd9t" event={"ID":"4548f47d-11cc-4762-bc6f-22d31fc2202e","Type":"ContainerDied","Data":"94d0a627619c8be698ef3052f7c956528b57bbe2b00f2d05e2736ab5b652913f"} Nov 22 08:52:43 crc kubenswrapper[4789]: I1122 08:52:43.491196 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nhd9t" event={"ID":"4548f47d-11cc-4762-bc6f-22d31fc2202e","Type":"ContainerDied","Data":"e8d878216f7ee10536e7c80aa7acb94cb5a870cbaacc01dca45610f42b0c4a1d"} Nov 22 08:52:43 crc kubenswrapper[4789]: I1122 08:52:43.491210 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nhd9t" Nov 22 08:52:43 crc kubenswrapper[4789]: I1122 08:52:43.491227 4789 scope.go:117] "RemoveContainer" containerID="94d0a627619c8be698ef3052f7c956528b57bbe2b00f2d05e2736ab5b652913f" Nov 22 08:52:43 crc kubenswrapper[4789]: I1122 08:52:43.518191 4789 scope.go:117] "RemoveContainer" containerID="05c2302c8972fe10df8022dd43fa9505ee24a11c31e22357d7f01de6050d4267" Nov 22 08:52:43 crc kubenswrapper[4789]: I1122 08:52:43.527066 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nhd9t"] Nov 22 08:52:43 crc kubenswrapper[4789]: I1122 08:52:43.534998 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-nhd9t"] Nov 22 08:52:43 crc kubenswrapper[4789]: I1122 08:52:43.546492 4789 scope.go:117] "RemoveContainer" containerID="daaa4db34dda9555b56b0188d8c71a7c46a73b0e95de9a0ad1647ce042964af8" Nov 22 08:52:43 crc kubenswrapper[4789]: I1122 08:52:43.585727 4789 scope.go:117] "RemoveContainer" containerID="94d0a627619c8be698ef3052f7c956528b57bbe2b00f2d05e2736ab5b652913f" Nov 22 08:52:43 crc kubenswrapper[4789]: E1122 08:52:43.586311 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"94d0a627619c8be698ef3052f7c956528b57bbe2b00f2d05e2736ab5b652913f\": container with ID starting with 94d0a627619c8be698ef3052f7c956528b57bbe2b00f2d05e2736ab5b652913f not found: ID does not exist" containerID="94d0a627619c8be698ef3052f7c956528b57bbe2b00f2d05e2736ab5b652913f" Nov 22 08:52:43 crc kubenswrapper[4789]: I1122 08:52:43.586365 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"94d0a627619c8be698ef3052f7c956528b57bbe2b00f2d05e2736ab5b652913f"} err="failed to get container status \"94d0a627619c8be698ef3052f7c956528b57bbe2b00f2d05e2736ab5b652913f\": rpc error: code = NotFound desc = could not find container \"94d0a627619c8be698ef3052f7c956528b57bbe2b00f2d05e2736ab5b652913f\": container with ID starting with 94d0a627619c8be698ef3052f7c956528b57bbe2b00f2d05e2736ab5b652913f not found: ID does not exist" Nov 22 08:52:43 crc kubenswrapper[4789]: I1122 08:52:43.586410 4789 scope.go:117] "RemoveContainer" containerID="05c2302c8972fe10df8022dd43fa9505ee24a11c31e22357d7f01de6050d4267" Nov 22 08:52:43 crc kubenswrapper[4789]: E1122 08:52:43.586916 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"05c2302c8972fe10df8022dd43fa9505ee24a11c31e22357d7f01de6050d4267\": container with ID starting with 05c2302c8972fe10df8022dd43fa9505ee24a11c31e22357d7f01de6050d4267 not found: ID does not exist" containerID="05c2302c8972fe10df8022dd43fa9505ee24a11c31e22357d7f01de6050d4267" Nov 22 08:52:43 crc kubenswrapper[4789]: I1122 08:52:43.586956 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"05c2302c8972fe10df8022dd43fa9505ee24a11c31e22357d7f01de6050d4267"} err="failed to get container status \"05c2302c8972fe10df8022dd43fa9505ee24a11c31e22357d7f01de6050d4267\": rpc error: code = NotFound desc = could not find container \"05c2302c8972fe10df8022dd43fa9505ee24a11c31e22357d7f01de6050d4267\": container with ID starting with 05c2302c8972fe10df8022dd43fa9505ee24a11c31e22357d7f01de6050d4267 not found: ID does not exist" Nov 22 08:52:43 crc kubenswrapper[4789]: I1122 08:52:43.587002 4789 scope.go:117] "RemoveContainer" containerID="daaa4db34dda9555b56b0188d8c71a7c46a73b0e95de9a0ad1647ce042964af8" Nov 22 08:52:43 crc kubenswrapper[4789]: E1122 08:52:43.587329 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"daaa4db34dda9555b56b0188d8c71a7c46a73b0e95de9a0ad1647ce042964af8\": container with ID starting with daaa4db34dda9555b56b0188d8c71a7c46a73b0e95de9a0ad1647ce042964af8 not found: ID does not exist" containerID="daaa4db34dda9555b56b0188d8c71a7c46a73b0e95de9a0ad1647ce042964af8" Nov 22 08:52:43 crc kubenswrapper[4789]: I1122 08:52:43.587366 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"daaa4db34dda9555b56b0188d8c71a7c46a73b0e95de9a0ad1647ce042964af8"} err="failed to get container status \"daaa4db34dda9555b56b0188d8c71a7c46a73b0e95de9a0ad1647ce042964af8\": rpc error: code = NotFound desc = could not find container \"daaa4db34dda9555b56b0188d8c71a7c46a73b0e95de9a0ad1647ce042964af8\": container with ID starting with daaa4db34dda9555b56b0188d8c71a7c46a73b0e95de9a0ad1647ce042964af8 not found: ID does not exist" Nov 22 08:52:43 crc kubenswrapper[4789]: I1122 08:52:43.977740 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4548f47d-11cc-4762-bc6f-22d31fc2202e" path="/var/lib/kubelet/pods/4548f47d-11cc-4762-bc6f-22d31fc2202e/volumes" Nov 22 08:52:44 crc kubenswrapper[4789]: I1122 08:52:44.651486 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-hs28n" Nov 22 08:52:45 crc kubenswrapper[4789]: I1122 08:52:45.662335 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hs28n"] Nov 22 08:52:45 crc kubenswrapper[4789]: I1122 08:52:45.662940 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-hs28n" podUID="43964ece-d130-4c53-a61e-ecf30c9d62f1" containerName="registry-server" containerID="cri-o://1838642f044e8e25b0e5e07950a95a97b656ce186532b2a638485f92fba7cf8f" gracePeriod=2 Nov 22 08:52:46 crc kubenswrapper[4789]: I1122 08:52:46.125649 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hs28n" Nov 22 08:52:46 crc kubenswrapper[4789]: I1122 08:52:46.243546 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/43964ece-d130-4c53-a61e-ecf30c9d62f1-catalog-content\") pod \"43964ece-d130-4c53-a61e-ecf30c9d62f1\" (UID: \"43964ece-d130-4c53-a61e-ecf30c9d62f1\") " Nov 22 08:52:46 crc kubenswrapper[4789]: I1122 08:52:46.243603 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xh8t6\" (UniqueName: \"kubernetes.io/projected/43964ece-d130-4c53-a61e-ecf30c9d62f1-kube-api-access-xh8t6\") pod \"43964ece-d130-4c53-a61e-ecf30c9d62f1\" (UID: \"43964ece-d130-4c53-a61e-ecf30c9d62f1\") " Nov 22 08:52:46 crc kubenswrapper[4789]: I1122 08:52:46.243770 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/43964ece-d130-4c53-a61e-ecf30c9d62f1-utilities\") pod \"43964ece-d130-4c53-a61e-ecf30c9d62f1\" (UID: \"43964ece-d130-4c53-a61e-ecf30c9d62f1\") " Nov 22 08:52:46 crc kubenswrapper[4789]: I1122 08:52:46.244587 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/43964ece-d130-4c53-a61e-ecf30c9d62f1-utilities" (OuterVolumeSpecName: "utilities") pod "43964ece-d130-4c53-a61e-ecf30c9d62f1" (UID: "43964ece-d130-4c53-a61e-ecf30c9d62f1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:52:46 crc kubenswrapper[4789]: I1122 08:52:46.249802 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43964ece-d130-4c53-a61e-ecf30c9d62f1-kube-api-access-xh8t6" (OuterVolumeSpecName: "kube-api-access-xh8t6") pod "43964ece-d130-4c53-a61e-ecf30c9d62f1" (UID: "43964ece-d130-4c53-a61e-ecf30c9d62f1"). InnerVolumeSpecName "kube-api-access-xh8t6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:52:46 crc kubenswrapper[4789]: I1122 08:52:46.296827 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/43964ece-d130-4c53-a61e-ecf30c9d62f1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "43964ece-d130-4c53-a61e-ecf30c9d62f1" (UID: "43964ece-d130-4c53-a61e-ecf30c9d62f1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:52:46 crc kubenswrapper[4789]: I1122 08:52:46.345990 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/43964ece-d130-4c53-a61e-ecf30c9d62f1-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 08:52:46 crc kubenswrapper[4789]: I1122 08:52:46.346019 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/43964ece-d130-4c53-a61e-ecf30c9d62f1-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 08:52:46 crc kubenswrapper[4789]: I1122 08:52:46.346030 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xh8t6\" (UniqueName: \"kubernetes.io/projected/43964ece-d130-4c53-a61e-ecf30c9d62f1-kube-api-access-xh8t6\") on node \"crc\" DevicePath \"\"" Nov 22 08:52:46 crc kubenswrapper[4789]: I1122 08:52:46.520606 4789 generic.go:334] "Generic (PLEG): container finished" podID="43964ece-d130-4c53-a61e-ecf30c9d62f1" containerID="1838642f044e8e25b0e5e07950a95a97b656ce186532b2a638485f92fba7cf8f" exitCode=0 Nov 22 08:52:46 crc kubenswrapper[4789]: I1122 08:52:46.520660 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hs28n" event={"ID":"43964ece-d130-4c53-a61e-ecf30c9d62f1","Type":"ContainerDied","Data":"1838642f044e8e25b0e5e07950a95a97b656ce186532b2a638485f92fba7cf8f"} Nov 22 08:52:46 crc kubenswrapper[4789]: I1122 08:52:46.520666 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hs28n" Nov 22 08:52:46 crc kubenswrapper[4789]: I1122 08:52:46.520699 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hs28n" event={"ID":"43964ece-d130-4c53-a61e-ecf30c9d62f1","Type":"ContainerDied","Data":"4c81001d5b86252b430e32ca0264b6848bfada46751c7d5b566511b6d11fb1c0"} Nov 22 08:52:46 crc kubenswrapper[4789]: I1122 08:52:46.520725 4789 scope.go:117] "RemoveContainer" containerID="1838642f044e8e25b0e5e07950a95a97b656ce186532b2a638485f92fba7cf8f" Nov 22 08:52:46 crc kubenswrapper[4789]: I1122 08:52:46.540250 4789 scope.go:117] "RemoveContainer" containerID="278b664b212ead0cb707759a357289ad71d0828a606bd9d751bb0f94b1ed8db0" Nov 22 08:52:46 crc kubenswrapper[4789]: I1122 08:52:46.556518 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hs28n"] Nov 22 08:52:46 crc kubenswrapper[4789]: I1122 08:52:46.564878 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-hs28n"] Nov 22 08:52:46 crc kubenswrapper[4789]: I1122 08:52:46.575957 4789 scope.go:117] "RemoveContainer" containerID="8b70ed0796b5b84fcce1fce4ce7d68c983efb0d81c74e2b5957c8a888e0acdbd" Nov 22 08:52:46 crc kubenswrapper[4789]: I1122 08:52:46.608968 4789 scope.go:117] "RemoveContainer" containerID="1838642f044e8e25b0e5e07950a95a97b656ce186532b2a638485f92fba7cf8f" Nov 22 08:52:46 crc kubenswrapper[4789]: E1122 08:52:46.609463 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1838642f044e8e25b0e5e07950a95a97b656ce186532b2a638485f92fba7cf8f\": container with ID starting with 1838642f044e8e25b0e5e07950a95a97b656ce186532b2a638485f92fba7cf8f not found: ID does not exist" containerID="1838642f044e8e25b0e5e07950a95a97b656ce186532b2a638485f92fba7cf8f" Nov 22 08:52:46 crc kubenswrapper[4789]: I1122 08:52:46.609531 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1838642f044e8e25b0e5e07950a95a97b656ce186532b2a638485f92fba7cf8f"} err="failed to get container status \"1838642f044e8e25b0e5e07950a95a97b656ce186532b2a638485f92fba7cf8f\": rpc error: code = NotFound desc = could not find container \"1838642f044e8e25b0e5e07950a95a97b656ce186532b2a638485f92fba7cf8f\": container with ID starting with 1838642f044e8e25b0e5e07950a95a97b656ce186532b2a638485f92fba7cf8f not found: ID does not exist" Nov 22 08:52:46 crc kubenswrapper[4789]: I1122 08:52:46.609560 4789 scope.go:117] "RemoveContainer" containerID="278b664b212ead0cb707759a357289ad71d0828a606bd9d751bb0f94b1ed8db0" Nov 22 08:52:46 crc kubenswrapper[4789]: E1122 08:52:46.609907 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"278b664b212ead0cb707759a357289ad71d0828a606bd9d751bb0f94b1ed8db0\": container with ID starting with 278b664b212ead0cb707759a357289ad71d0828a606bd9d751bb0f94b1ed8db0 not found: ID does not exist" containerID="278b664b212ead0cb707759a357289ad71d0828a606bd9d751bb0f94b1ed8db0" Nov 22 08:52:46 crc kubenswrapper[4789]: I1122 08:52:46.609942 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"278b664b212ead0cb707759a357289ad71d0828a606bd9d751bb0f94b1ed8db0"} err="failed to get container status \"278b664b212ead0cb707759a357289ad71d0828a606bd9d751bb0f94b1ed8db0\": rpc error: code = NotFound desc = could not find container \"278b664b212ead0cb707759a357289ad71d0828a606bd9d751bb0f94b1ed8db0\": container with ID starting with 278b664b212ead0cb707759a357289ad71d0828a606bd9d751bb0f94b1ed8db0 not found: ID does not exist" Nov 22 08:52:46 crc kubenswrapper[4789]: I1122 08:52:46.609969 4789 scope.go:117] "RemoveContainer" containerID="8b70ed0796b5b84fcce1fce4ce7d68c983efb0d81c74e2b5957c8a888e0acdbd" Nov 22 08:52:46 crc kubenswrapper[4789]: E1122 08:52:46.610294 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8b70ed0796b5b84fcce1fce4ce7d68c983efb0d81c74e2b5957c8a888e0acdbd\": container with ID starting with 8b70ed0796b5b84fcce1fce4ce7d68c983efb0d81c74e2b5957c8a888e0acdbd not found: ID does not exist" containerID="8b70ed0796b5b84fcce1fce4ce7d68c983efb0d81c74e2b5957c8a888e0acdbd" Nov 22 08:52:46 crc kubenswrapper[4789]: I1122 08:52:46.610320 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b70ed0796b5b84fcce1fce4ce7d68c983efb0d81c74e2b5957c8a888e0acdbd"} err="failed to get container status \"8b70ed0796b5b84fcce1fce4ce7d68c983efb0d81c74e2b5957c8a888e0acdbd\": rpc error: code = NotFound desc = could not find container \"8b70ed0796b5b84fcce1fce4ce7d68c983efb0d81c74e2b5957c8a888e0acdbd\": container with ID starting with 8b70ed0796b5b84fcce1fce4ce7d68c983efb0d81c74e2b5957c8a888e0acdbd not found: ID does not exist" Nov 22 08:52:47 crc kubenswrapper[4789]: I1122 08:52:47.974544 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43964ece-d130-4c53-a61e-ecf30c9d62f1" path="/var/lib/kubelet/pods/43964ece-d130-4c53-a61e-ecf30c9d62f1/volumes" Nov 22 08:53:00 crc kubenswrapper[4789]: I1122 08:53:00.647124 4789 generic.go:334] "Generic (PLEG): container finished" podID="2c231e91-12f1-47e4-8947-1bf03bcae704" containerID="10e60a23d3a242d246e6ca5a07b7cad8d65dc2f0d72f533b5609d93e5c4ee559" exitCode=0 Nov 22 08:53:00 crc kubenswrapper[4789]: I1122 08:53:00.647230 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b47tx" event={"ID":"2c231e91-12f1-47e4-8947-1bf03bcae704","Type":"ContainerDied","Data":"10e60a23d3a242d246e6ca5a07b7cad8d65dc2f0d72f533b5609d93e5c4ee559"} Nov 22 08:53:02 crc kubenswrapper[4789]: I1122 08:53:02.074252 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b47tx" Nov 22 08:53:02 crc kubenswrapper[4789]: I1122 08:53:02.137129 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2c231e91-12f1-47e4-8947-1bf03bcae704-ssh-key\") pod \"2c231e91-12f1-47e4-8947-1bf03bcae704\" (UID: \"2c231e91-12f1-47e4-8947-1bf03bcae704\") " Nov 22 08:53:02 crc kubenswrapper[4789]: I1122 08:53:02.137213 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c231e91-12f1-47e4-8947-1bf03bcae704-nova-custom-ceph-combined-ca-bundle\") pod \"2c231e91-12f1-47e4-8947-1bf03bcae704\" (UID: \"2c231e91-12f1-47e4-8947-1bf03bcae704\") " Nov 22 08:53:02 crc kubenswrapper[4789]: I1122 08:53:02.137252 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/2c231e91-12f1-47e4-8947-1bf03bcae704-nova-migration-ssh-key-0\") pod \"2c231e91-12f1-47e4-8947-1bf03bcae704\" (UID: \"2c231e91-12f1-47e4-8947-1bf03bcae704\") " Nov 22 08:53:02 crc kubenswrapper[4789]: I1122 08:53:02.137312 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2c231e91-12f1-47e4-8947-1bf03bcae704-inventory\") pod \"2c231e91-12f1-47e4-8947-1bf03bcae704\" (UID: \"2c231e91-12f1-47e4-8947-1bf03bcae704\") " Nov 22 08:53:02 crc kubenswrapper[4789]: I1122 08:53:02.137351 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2c231e91-12f1-47e4-8947-1bf03bcae704-ceph\") pod \"2c231e91-12f1-47e4-8947-1bf03bcae704\" (UID: \"2c231e91-12f1-47e4-8947-1bf03bcae704\") " Nov 22 08:53:02 crc kubenswrapper[4789]: I1122 08:53:02.137383 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/2c231e91-12f1-47e4-8947-1bf03bcae704-ceph-nova-0\") pod \"2c231e91-12f1-47e4-8947-1bf03bcae704\" (UID: \"2c231e91-12f1-47e4-8947-1bf03bcae704\") " Nov 22 08:53:02 crc kubenswrapper[4789]: I1122 08:53:02.137489 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/2c231e91-12f1-47e4-8947-1bf03bcae704-nova-cell1-compute-config-0\") pod \"2c231e91-12f1-47e4-8947-1bf03bcae704\" (UID: \"2c231e91-12f1-47e4-8947-1bf03bcae704\") " Nov 22 08:53:02 crc kubenswrapper[4789]: I1122 08:53:02.137534 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xzm9h\" (UniqueName: \"kubernetes.io/projected/2c231e91-12f1-47e4-8947-1bf03bcae704-kube-api-access-xzm9h\") pod \"2c231e91-12f1-47e4-8947-1bf03bcae704\" (UID: \"2c231e91-12f1-47e4-8947-1bf03bcae704\") " Nov 22 08:53:02 crc kubenswrapper[4789]: I1122 08:53:02.137571 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/2c231e91-12f1-47e4-8947-1bf03bcae704-nova-cell1-compute-config-1\") pod \"2c231e91-12f1-47e4-8947-1bf03bcae704\" (UID: \"2c231e91-12f1-47e4-8947-1bf03bcae704\") " Nov 22 08:53:02 crc kubenswrapper[4789]: I1122 08:53:02.137601 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/2c231e91-12f1-47e4-8947-1bf03bcae704-nova-migration-ssh-key-1\") pod \"2c231e91-12f1-47e4-8947-1bf03bcae704\" (UID: \"2c231e91-12f1-47e4-8947-1bf03bcae704\") " Nov 22 08:53:02 crc kubenswrapper[4789]: I1122 08:53:02.137630 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/2c231e91-12f1-47e4-8947-1bf03bcae704-nova-extra-config-0\") pod \"2c231e91-12f1-47e4-8947-1bf03bcae704\" (UID: \"2c231e91-12f1-47e4-8947-1bf03bcae704\") " Nov 22 08:53:02 crc kubenswrapper[4789]: I1122 08:53:02.144358 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c231e91-12f1-47e4-8947-1bf03bcae704-ceph" (OuterVolumeSpecName: "ceph") pod "2c231e91-12f1-47e4-8947-1bf03bcae704" (UID: "2c231e91-12f1-47e4-8947-1bf03bcae704"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:53:02 crc kubenswrapper[4789]: I1122 08:53:02.144977 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c231e91-12f1-47e4-8947-1bf03bcae704-nova-custom-ceph-combined-ca-bundle" (OuterVolumeSpecName: "nova-custom-ceph-combined-ca-bundle") pod "2c231e91-12f1-47e4-8947-1bf03bcae704" (UID: "2c231e91-12f1-47e4-8947-1bf03bcae704"). InnerVolumeSpecName "nova-custom-ceph-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:53:02 crc kubenswrapper[4789]: I1122 08:53:02.145544 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c231e91-12f1-47e4-8947-1bf03bcae704-kube-api-access-xzm9h" (OuterVolumeSpecName: "kube-api-access-xzm9h") pod "2c231e91-12f1-47e4-8947-1bf03bcae704" (UID: "2c231e91-12f1-47e4-8947-1bf03bcae704"). InnerVolumeSpecName "kube-api-access-xzm9h". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:53:02 crc kubenswrapper[4789]: I1122 08:53:02.166046 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c231e91-12f1-47e4-8947-1bf03bcae704-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "2c231e91-12f1-47e4-8947-1bf03bcae704" (UID: "2c231e91-12f1-47e4-8947-1bf03bcae704"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:53:02 crc kubenswrapper[4789]: I1122 08:53:02.166192 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2c231e91-12f1-47e4-8947-1bf03bcae704-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "2c231e91-12f1-47e4-8947-1bf03bcae704" (UID: "2c231e91-12f1-47e4-8947-1bf03bcae704"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:53:02 crc kubenswrapper[4789]: I1122 08:53:02.168543 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c231e91-12f1-47e4-8947-1bf03bcae704-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2c231e91-12f1-47e4-8947-1bf03bcae704" (UID: "2c231e91-12f1-47e4-8947-1bf03bcae704"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:53:02 crc kubenswrapper[4789]: I1122 08:53:02.168566 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c231e91-12f1-47e4-8947-1bf03bcae704-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "2c231e91-12f1-47e4-8947-1bf03bcae704" (UID: "2c231e91-12f1-47e4-8947-1bf03bcae704"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:53:02 crc kubenswrapper[4789]: I1122 08:53:02.170787 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2c231e91-12f1-47e4-8947-1bf03bcae704-ceph-nova-0" (OuterVolumeSpecName: "ceph-nova-0") pod "2c231e91-12f1-47e4-8947-1bf03bcae704" (UID: "2c231e91-12f1-47e4-8947-1bf03bcae704"). InnerVolumeSpecName "ceph-nova-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:53:02 crc kubenswrapper[4789]: I1122 08:53:02.174135 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c231e91-12f1-47e4-8947-1bf03bcae704-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "2c231e91-12f1-47e4-8947-1bf03bcae704" (UID: "2c231e91-12f1-47e4-8947-1bf03bcae704"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:53:02 crc kubenswrapper[4789]: I1122 08:53:02.177150 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c231e91-12f1-47e4-8947-1bf03bcae704-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "2c231e91-12f1-47e4-8947-1bf03bcae704" (UID: "2c231e91-12f1-47e4-8947-1bf03bcae704"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:53:02 crc kubenswrapper[4789]: I1122 08:53:02.180263 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c231e91-12f1-47e4-8947-1bf03bcae704-inventory" (OuterVolumeSpecName: "inventory") pod "2c231e91-12f1-47e4-8947-1bf03bcae704" (UID: "2c231e91-12f1-47e4-8947-1bf03bcae704"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:53:02 crc kubenswrapper[4789]: I1122 08:53:02.239731 4789 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2c231e91-12f1-47e4-8947-1bf03bcae704-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 22 08:53:02 crc kubenswrapper[4789]: I1122 08:53:02.241063 4789 reconciler_common.go:293] "Volume detached for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c231e91-12f1-47e4-8947-1bf03bcae704-nova-custom-ceph-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:53:02 crc kubenswrapper[4789]: I1122 08:53:02.241094 4789 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/2c231e91-12f1-47e4-8947-1bf03bcae704-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Nov 22 08:53:02 crc kubenswrapper[4789]: I1122 08:53:02.241107 4789 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2c231e91-12f1-47e4-8947-1bf03bcae704-inventory\") on node \"crc\" DevicePath \"\"" Nov 22 08:53:02 crc kubenswrapper[4789]: I1122 08:53:02.241118 4789 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2c231e91-12f1-47e4-8947-1bf03bcae704-ceph\") on node \"crc\" DevicePath \"\"" Nov 22 08:53:02 crc kubenswrapper[4789]: I1122 08:53:02.241128 4789 reconciler_common.go:293] "Volume detached for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/2c231e91-12f1-47e4-8947-1bf03bcae704-ceph-nova-0\") on node \"crc\" DevicePath \"\"" Nov 22 08:53:02 crc kubenswrapper[4789]: I1122 08:53:02.241141 4789 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/2c231e91-12f1-47e4-8947-1bf03bcae704-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Nov 22 08:53:02 crc kubenswrapper[4789]: I1122 08:53:02.241150 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xzm9h\" (UniqueName: \"kubernetes.io/projected/2c231e91-12f1-47e4-8947-1bf03bcae704-kube-api-access-xzm9h\") on node \"crc\" DevicePath \"\"" Nov 22 08:53:02 crc kubenswrapper[4789]: I1122 08:53:02.241161 4789 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/2c231e91-12f1-47e4-8947-1bf03bcae704-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Nov 22 08:53:02 crc kubenswrapper[4789]: I1122 08:53:02.241171 4789 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/2c231e91-12f1-47e4-8947-1bf03bcae704-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Nov 22 08:53:02 crc kubenswrapper[4789]: I1122 08:53:02.241180 4789 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/2c231e91-12f1-47e4-8947-1bf03bcae704-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Nov 22 08:53:02 crc kubenswrapper[4789]: I1122 08:53:02.664555 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b47tx" event={"ID":"2c231e91-12f1-47e4-8947-1bf03bcae704","Type":"ContainerDied","Data":"55cfd923b3d8343569a7116305dd23df3aebd9cf05fd463a23613a771f7c5c00"} Nov 22 08:53:02 crc kubenswrapper[4789]: I1122 08:53:02.664600 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b47tx" Nov 22 08:53:02 crc kubenswrapper[4789]: I1122 08:53:02.664603 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="55cfd923b3d8343569a7116305dd23df3aebd9cf05fd463a23613a771f7c5c00" Nov 22 08:53:05 crc kubenswrapper[4789]: I1122 08:53:05.372704 4789 patch_prober.go:28] interesting pod/machine-config-daemon-pxvrm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 08:53:05 crc kubenswrapper[4789]: I1122 08:53:05.374285 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 08:53:16 crc kubenswrapper[4789]: I1122 08:53:16.994058 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-volume-volume1-0"] Nov 22 08:53:16 crc kubenswrapper[4789]: E1122 08:53:16.995048 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4548f47d-11cc-4762-bc6f-22d31fc2202e" containerName="extract-content" Nov 22 08:53:16 crc kubenswrapper[4789]: I1122 08:53:16.995064 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="4548f47d-11cc-4762-bc6f-22d31fc2202e" containerName="extract-content" Nov 22 08:53:16 crc kubenswrapper[4789]: E1122 08:53:16.995081 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4548f47d-11cc-4762-bc6f-22d31fc2202e" containerName="extract-utilities" Nov 22 08:53:16 crc kubenswrapper[4789]: I1122 08:53:16.995088 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="4548f47d-11cc-4762-bc6f-22d31fc2202e" containerName="extract-utilities" Nov 22 08:53:16 crc kubenswrapper[4789]: E1122 08:53:16.995097 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43964ece-d130-4c53-a61e-ecf30c9d62f1" containerName="registry-server" Nov 22 08:53:16 crc kubenswrapper[4789]: I1122 08:53:16.995103 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="43964ece-d130-4c53-a61e-ecf30c9d62f1" containerName="registry-server" Nov 22 08:53:16 crc kubenswrapper[4789]: E1122 08:53:16.995122 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43964ece-d130-4c53-a61e-ecf30c9d62f1" containerName="extract-utilities" Nov 22 08:53:16 crc kubenswrapper[4789]: I1122 08:53:16.995127 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="43964ece-d130-4c53-a61e-ecf30c9d62f1" containerName="extract-utilities" Nov 22 08:53:16 crc kubenswrapper[4789]: E1122 08:53:16.995142 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4548f47d-11cc-4762-bc6f-22d31fc2202e" containerName="registry-server" Nov 22 08:53:16 crc kubenswrapper[4789]: I1122 08:53:16.995147 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="4548f47d-11cc-4762-bc6f-22d31fc2202e" containerName="registry-server" Nov 22 08:53:16 crc kubenswrapper[4789]: E1122 08:53:16.995160 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c231e91-12f1-47e4-8947-1bf03bcae704" containerName="nova-custom-ceph-edpm-deployment-openstack-edpm-ipam" Nov 22 08:53:16 crc kubenswrapper[4789]: I1122 08:53:16.995169 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c231e91-12f1-47e4-8947-1bf03bcae704" containerName="nova-custom-ceph-edpm-deployment-openstack-edpm-ipam" Nov 22 08:53:16 crc kubenswrapper[4789]: E1122 08:53:16.995181 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43964ece-d130-4c53-a61e-ecf30c9d62f1" containerName="extract-content" Nov 22 08:53:16 crc kubenswrapper[4789]: I1122 08:53:16.995186 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="43964ece-d130-4c53-a61e-ecf30c9d62f1" containerName="extract-content" Nov 22 08:53:16 crc kubenswrapper[4789]: I1122 08:53:16.995352 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="4548f47d-11cc-4762-bc6f-22d31fc2202e" containerName="registry-server" Nov 22 08:53:16 crc kubenswrapper[4789]: I1122 08:53:16.995379 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c231e91-12f1-47e4-8947-1bf03bcae704" containerName="nova-custom-ceph-edpm-deployment-openstack-edpm-ipam" Nov 22 08:53:16 crc kubenswrapper[4789]: I1122 08:53:16.995391 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="43964ece-d130-4c53-a61e-ecf30c9d62f1" containerName="registry-server" Nov 22 08:53:16 crc kubenswrapper[4789]: I1122 08:53:16.996393 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-volume1-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:16.999314 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.000062 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-volume-volume1-config-data" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.006897 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-volume1-0"] Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.061432 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-backup-0"] Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.063043 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.067860 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-backup-config-data" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.082061 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.102842 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2b1ad6b9-9f67-4533-bafd-5933ade7f938-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"2b1ad6b9-9f67-4533-bafd-5933ade7f938\") " pod="openstack/cinder-volume-volume1-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.102989 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2b1ad6b9-9f67-4533-bafd-5933ade7f938-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"2b1ad6b9-9f67-4533-bafd-5933ade7f938\") " pod="openstack/cinder-volume-volume1-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.103071 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/2b1ad6b9-9f67-4533-bafd-5933ade7f938-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"2b1ad6b9-9f67-4533-bafd-5933ade7f938\") " pod="openstack/cinder-volume-volume1-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.103096 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/2b1ad6b9-9f67-4533-bafd-5933ade7f938-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"2b1ad6b9-9f67-4533-bafd-5933ade7f938\") " pod="openstack/cinder-volume-volume1-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.103127 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/2b1ad6b9-9f67-4533-bafd-5933ade7f938-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"2b1ad6b9-9f67-4533-bafd-5933ade7f938\") " pod="openstack/cinder-volume-volume1-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.103152 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/2b1ad6b9-9f67-4533-bafd-5933ade7f938-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"2b1ad6b9-9f67-4533-bafd-5933ade7f938\") " pod="openstack/cinder-volume-volume1-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.103169 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/2b1ad6b9-9f67-4533-bafd-5933ade7f938-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"2b1ad6b9-9f67-4533-bafd-5933ade7f938\") " pod="openstack/cinder-volume-volume1-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.103248 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/2b1ad6b9-9f67-4533-bafd-5933ade7f938-dev\") pod \"cinder-volume-volume1-0\" (UID: \"2b1ad6b9-9f67-4533-bafd-5933ade7f938\") " pod="openstack/cinder-volume-volume1-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.103289 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b1ad6b9-9f67-4533-bafd-5933ade7f938-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"2b1ad6b9-9f67-4533-bafd-5933ade7f938\") " pod="openstack/cinder-volume-volume1-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.103404 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/2b1ad6b9-9f67-4533-bafd-5933ade7f938-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"2b1ad6b9-9f67-4533-bafd-5933ade7f938\") " pod="openstack/cinder-volume-volume1-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.103448 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b1ad6b9-9f67-4533-bafd-5933ade7f938-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"2b1ad6b9-9f67-4533-bafd-5933ade7f938\") " pod="openstack/cinder-volume-volume1-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.103473 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bqdp9\" (UniqueName: \"kubernetes.io/projected/2b1ad6b9-9f67-4533-bafd-5933ade7f938-kube-api-access-bqdp9\") pod \"cinder-volume-volume1-0\" (UID: \"2b1ad6b9-9f67-4533-bafd-5933ade7f938\") " pod="openstack/cinder-volume-volume1-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.103509 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/2b1ad6b9-9f67-4533-bafd-5933ade7f938-sys\") pod \"cinder-volume-volume1-0\" (UID: \"2b1ad6b9-9f67-4533-bafd-5933ade7f938\") " pod="openstack/cinder-volume-volume1-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.103542 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/2b1ad6b9-9f67-4533-bafd-5933ade7f938-run\") pod \"cinder-volume-volume1-0\" (UID: \"2b1ad6b9-9f67-4533-bafd-5933ade7f938\") " pod="openstack/cinder-volume-volume1-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.103635 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/2b1ad6b9-9f67-4533-bafd-5933ade7f938-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"2b1ad6b9-9f67-4533-bafd-5933ade7f938\") " pod="openstack/cinder-volume-volume1-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.103665 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2b1ad6b9-9f67-4533-bafd-5933ade7f938-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"2b1ad6b9-9f67-4533-bafd-5933ade7f938\") " pod="openstack/cinder-volume-volume1-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.205028 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zfkfl\" (UniqueName: \"kubernetes.io/projected/62901ed9-929f-443e-ae1b-dc75cb2e1514-kube-api-access-zfkfl\") pod \"cinder-backup-0\" (UID: \"62901ed9-929f-443e-ae1b-dc75cb2e1514\") " pod="openstack/cinder-backup-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.205079 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/2b1ad6b9-9f67-4533-bafd-5933ade7f938-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"2b1ad6b9-9f67-4533-bafd-5933ade7f938\") " pod="openstack/cinder-volume-volume1-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.205102 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/2b1ad6b9-9f67-4533-bafd-5933ade7f938-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"2b1ad6b9-9f67-4533-bafd-5933ade7f938\") " pod="openstack/cinder-volume-volume1-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.205126 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/62901ed9-929f-443e-ae1b-dc75cb2e1514-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"62901ed9-929f-443e-ae1b-dc75cb2e1514\") " pod="openstack/cinder-backup-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.205146 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/62901ed9-929f-443e-ae1b-dc75cb2e1514-config-data-custom\") pod \"cinder-backup-0\" (UID: \"62901ed9-929f-443e-ae1b-dc75cb2e1514\") " pod="openstack/cinder-backup-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.205162 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/2b1ad6b9-9f67-4533-bafd-5933ade7f938-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"2b1ad6b9-9f67-4533-bafd-5933ade7f938\") " pod="openstack/cinder-volume-volume1-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.205179 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/2b1ad6b9-9f67-4533-bafd-5933ade7f938-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"2b1ad6b9-9f67-4533-bafd-5933ade7f938\") " pod="openstack/cinder-volume-volume1-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.205219 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62901ed9-929f-443e-ae1b-dc75cb2e1514-config-data\") pod \"cinder-backup-0\" (UID: \"62901ed9-929f-443e-ae1b-dc75cb2e1514\") " pod="openstack/cinder-backup-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.205281 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/2b1ad6b9-9f67-4533-bafd-5933ade7f938-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"2b1ad6b9-9f67-4533-bafd-5933ade7f938\") " pod="openstack/cinder-volume-volume1-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.205321 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/62901ed9-929f-443e-ae1b-dc75cb2e1514-dev\") pod \"cinder-backup-0\" (UID: \"62901ed9-929f-443e-ae1b-dc75cb2e1514\") " pod="openstack/cinder-backup-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.205355 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/2b1ad6b9-9f67-4533-bafd-5933ade7f938-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"2b1ad6b9-9f67-4533-bafd-5933ade7f938\") " pod="openstack/cinder-volume-volume1-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.205428 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/2b1ad6b9-9f67-4533-bafd-5933ade7f938-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"2b1ad6b9-9f67-4533-bafd-5933ade7f938\") " pod="openstack/cinder-volume-volume1-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.205519 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/62901ed9-929f-443e-ae1b-dc75cb2e1514-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"62901ed9-929f-443e-ae1b-dc75cb2e1514\") " pod="openstack/cinder-backup-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.205523 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/2b1ad6b9-9f67-4533-bafd-5933ade7f938-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"2b1ad6b9-9f67-4533-bafd-5933ade7f938\") " pod="openstack/cinder-volume-volume1-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.205570 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/62901ed9-929f-443e-ae1b-dc75cb2e1514-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"62901ed9-929f-443e-ae1b-dc75cb2e1514\") " pod="openstack/cinder-backup-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.205610 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/2b1ad6b9-9f67-4533-bafd-5933ade7f938-dev\") pod \"cinder-volume-volume1-0\" (UID: \"2b1ad6b9-9f67-4533-bafd-5933ade7f938\") " pod="openstack/cinder-volume-volume1-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.205662 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/2b1ad6b9-9f67-4533-bafd-5933ade7f938-dev\") pod \"cinder-volume-volume1-0\" (UID: \"2b1ad6b9-9f67-4533-bafd-5933ade7f938\") " pod="openstack/cinder-volume-volume1-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.205682 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b1ad6b9-9f67-4533-bafd-5933ade7f938-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"2b1ad6b9-9f67-4533-bafd-5933ade7f938\") " pod="openstack/cinder-volume-volume1-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.205716 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/62901ed9-929f-443e-ae1b-dc75cb2e1514-lib-modules\") pod \"cinder-backup-0\" (UID: \"62901ed9-929f-443e-ae1b-dc75cb2e1514\") " pod="openstack/cinder-backup-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.205743 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/62901ed9-929f-443e-ae1b-dc75cb2e1514-sys\") pod \"cinder-backup-0\" (UID: \"62901ed9-929f-443e-ae1b-dc75cb2e1514\") " pod="openstack/cinder-backup-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.205905 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/2b1ad6b9-9f67-4533-bafd-5933ade7f938-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"2b1ad6b9-9f67-4533-bafd-5933ade7f938\") " pod="openstack/cinder-volume-volume1-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.205930 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/62901ed9-929f-443e-ae1b-dc75cb2e1514-etc-nvme\") pod \"cinder-backup-0\" (UID: \"62901ed9-929f-443e-ae1b-dc75cb2e1514\") " pod="openstack/cinder-backup-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.205941 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/2b1ad6b9-9f67-4533-bafd-5933ade7f938-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"2b1ad6b9-9f67-4533-bafd-5933ade7f938\") " pod="openstack/cinder-volume-volume1-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.206001 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b1ad6b9-9f67-4533-bafd-5933ade7f938-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"2b1ad6b9-9f67-4533-bafd-5933ade7f938\") " pod="openstack/cinder-volume-volume1-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.206021 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62901ed9-929f-443e-ae1b-dc75cb2e1514-scripts\") pod \"cinder-backup-0\" (UID: \"62901ed9-929f-443e-ae1b-dc75cb2e1514\") " pod="openstack/cinder-backup-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.206039 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bqdp9\" (UniqueName: \"kubernetes.io/projected/2b1ad6b9-9f67-4533-bafd-5933ade7f938-kube-api-access-bqdp9\") pod \"cinder-volume-volume1-0\" (UID: \"2b1ad6b9-9f67-4533-bafd-5933ade7f938\") " pod="openstack/cinder-volume-volume1-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.206192 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/2b1ad6b9-9f67-4533-bafd-5933ade7f938-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"2b1ad6b9-9f67-4533-bafd-5933ade7f938\") " pod="openstack/cinder-volume-volume1-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.206306 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/62901ed9-929f-443e-ae1b-dc75cb2e1514-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"62901ed9-929f-443e-ae1b-dc75cb2e1514\") " pod="openstack/cinder-backup-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.206329 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/62901ed9-929f-443e-ae1b-dc75cb2e1514-run\") pod \"cinder-backup-0\" (UID: \"62901ed9-929f-443e-ae1b-dc75cb2e1514\") " pod="openstack/cinder-backup-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.206351 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/2b1ad6b9-9f67-4533-bafd-5933ade7f938-sys\") pod \"cinder-volume-volume1-0\" (UID: \"2b1ad6b9-9f67-4533-bafd-5933ade7f938\") " pod="openstack/cinder-volume-volume1-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.206376 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/2b1ad6b9-9f67-4533-bafd-5933ade7f938-run\") pod \"cinder-volume-volume1-0\" (UID: \"2b1ad6b9-9f67-4533-bafd-5933ade7f938\") " pod="openstack/cinder-volume-volume1-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.206407 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/2b1ad6b9-9f67-4533-bafd-5933ade7f938-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"2b1ad6b9-9f67-4533-bafd-5933ade7f938\") " pod="openstack/cinder-volume-volume1-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.206427 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2b1ad6b9-9f67-4533-bafd-5933ade7f938-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"2b1ad6b9-9f67-4533-bafd-5933ade7f938\") " pod="openstack/cinder-volume-volume1-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.206444 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62901ed9-929f-443e-ae1b-dc75cb2e1514-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"62901ed9-929f-443e-ae1b-dc75cb2e1514\") " pod="openstack/cinder-backup-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.206466 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2b1ad6b9-9f67-4533-bafd-5933ade7f938-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"2b1ad6b9-9f67-4533-bafd-5933ade7f938\") " pod="openstack/cinder-volume-volume1-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.206483 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/62901ed9-929f-443e-ae1b-dc75cb2e1514-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"62901ed9-929f-443e-ae1b-dc75cb2e1514\") " pod="openstack/cinder-backup-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.206498 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/62901ed9-929f-443e-ae1b-dc75cb2e1514-ceph\") pod \"cinder-backup-0\" (UID: \"62901ed9-929f-443e-ae1b-dc75cb2e1514\") " pod="openstack/cinder-backup-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.206536 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2b1ad6b9-9f67-4533-bafd-5933ade7f938-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"2b1ad6b9-9f67-4533-bafd-5933ade7f938\") " pod="openstack/cinder-volume-volume1-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.207305 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/2b1ad6b9-9f67-4533-bafd-5933ade7f938-sys\") pod \"cinder-volume-volume1-0\" (UID: \"2b1ad6b9-9f67-4533-bafd-5933ade7f938\") " pod="openstack/cinder-volume-volume1-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.207515 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2b1ad6b9-9f67-4533-bafd-5933ade7f938-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"2b1ad6b9-9f67-4533-bafd-5933ade7f938\") " pod="openstack/cinder-volume-volume1-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.207624 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/2b1ad6b9-9f67-4533-bafd-5933ade7f938-run\") pod \"cinder-volume-volume1-0\" (UID: \"2b1ad6b9-9f67-4533-bafd-5933ade7f938\") " pod="openstack/cinder-volume-volume1-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.207802 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/2b1ad6b9-9f67-4533-bafd-5933ade7f938-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"2b1ad6b9-9f67-4533-bafd-5933ade7f938\") " pod="openstack/cinder-volume-volume1-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.214148 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2b1ad6b9-9f67-4533-bafd-5933ade7f938-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"2b1ad6b9-9f67-4533-bafd-5933ade7f938\") " pod="openstack/cinder-volume-volume1-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.214425 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b1ad6b9-9f67-4533-bafd-5933ade7f938-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"2b1ad6b9-9f67-4533-bafd-5933ade7f938\") " pod="openstack/cinder-volume-volume1-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.214243 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/2b1ad6b9-9f67-4533-bafd-5933ade7f938-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"2b1ad6b9-9f67-4533-bafd-5933ade7f938\") " pod="openstack/cinder-volume-volume1-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.214656 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b1ad6b9-9f67-4533-bafd-5933ade7f938-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"2b1ad6b9-9f67-4533-bafd-5933ade7f938\") " pod="openstack/cinder-volume-volume1-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.215824 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2b1ad6b9-9f67-4533-bafd-5933ade7f938-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"2b1ad6b9-9f67-4533-bafd-5933ade7f938\") " pod="openstack/cinder-volume-volume1-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.233333 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bqdp9\" (UniqueName: \"kubernetes.io/projected/2b1ad6b9-9f67-4533-bafd-5933ade7f938-kube-api-access-bqdp9\") pod \"cinder-volume-volume1-0\" (UID: \"2b1ad6b9-9f67-4533-bafd-5933ade7f938\") " pod="openstack/cinder-volume-volume1-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.307938 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/62901ed9-929f-443e-ae1b-dc75cb2e1514-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"62901ed9-929f-443e-ae1b-dc75cb2e1514\") " pod="openstack/cinder-backup-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.308037 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/62901ed9-929f-443e-ae1b-dc75cb2e1514-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"62901ed9-929f-443e-ae1b-dc75cb2e1514\") " pod="openstack/cinder-backup-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.308198 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/62901ed9-929f-443e-ae1b-dc75cb2e1514-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"62901ed9-929f-443e-ae1b-dc75cb2e1514\") " pod="openstack/cinder-backup-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.308284 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/62901ed9-929f-443e-ae1b-dc75cb2e1514-lib-modules\") pod \"cinder-backup-0\" (UID: \"62901ed9-929f-443e-ae1b-dc75cb2e1514\") " pod="openstack/cinder-backup-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.308310 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/62901ed9-929f-443e-ae1b-dc75cb2e1514-sys\") pod \"cinder-backup-0\" (UID: \"62901ed9-929f-443e-ae1b-dc75cb2e1514\") " pod="openstack/cinder-backup-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.308361 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/62901ed9-929f-443e-ae1b-dc75cb2e1514-etc-nvme\") pod \"cinder-backup-0\" (UID: \"62901ed9-929f-443e-ae1b-dc75cb2e1514\") " pod="openstack/cinder-backup-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.308402 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62901ed9-929f-443e-ae1b-dc75cb2e1514-scripts\") pod \"cinder-backup-0\" (UID: \"62901ed9-929f-443e-ae1b-dc75cb2e1514\") " pod="openstack/cinder-backup-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.308432 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/62901ed9-929f-443e-ae1b-dc75cb2e1514-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"62901ed9-929f-443e-ae1b-dc75cb2e1514\") " pod="openstack/cinder-backup-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.308442 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/62901ed9-929f-443e-ae1b-dc75cb2e1514-sys\") pod \"cinder-backup-0\" (UID: \"62901ed9-929f-443e-ae1b-dc75cb2e1514\") " pod="openstack/cinder-backup-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.308475 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/62901ed9-929f-443e-ae1b-dc75cb2e1514-run\") pod \"cinder-backup-0\" (UID: \"62901ed9-929f-443e-ae1b-dc75cb2e1514\") " pod="openstack/cinder-backup-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.308442 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/62901ed9-929f-443e-ae1b-dc75cb2e1514-lib-modules\") pod \"cinder-backup-0\" (UID: \"62901ed9-929f-443e-ae1b-dc75cb2e1514\") " pod="openstack/cinder-backup-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.308451 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/62901ed9-929f-443e-ae1b-dc75cb2e1514-run\") pod \"cinder-backup-0\" (UID: \"62901ed9-929f-443e-ae1b-dc75cb2e1514\") " pod="openstack/cinder-backup-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.308533 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/62901ed9-929f-443e-ae1b-dc75cb2e1514-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"62901ed9-929f-443e-ae1b-dc75cb2e1514\") " pod="openstack/cinder-backup-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.308577 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/62901ed9-929f-443e-ae1b-dc75cb2e1514-etc-nvme\") pod \"cinder-backup-0\" (UID: \"62901ed9-929f-443e-ae1b-dc75cb2e1514\") " pod="openstack/cinder-backup-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.308688 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62901ed9-929f-443e-ae1b-dc75cb2e1514-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"62901ed9-929f-443e-ae1b-dc75cb2e1514\") " pod="openstack/cinder-backup-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.308767 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/62901ed9-929f-443e-ae1b-dc75cb2e1514-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"62901ed9-929f-443e-ae1b-dc75cb2e1514\") " pod="openstack/cinder-backup-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.308800 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/62901ed9-929f-443e-ae1b-dc75cb2e1514-ceph\") pod \"cinder-backup-0\" (UID: \"62901ed9-929f-443e-ae1b-dc75cb2e1514\") " pod="openstack/cinder-backup-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.308876 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/62901ed9-929f-443e-ae1b-dc75cb2e1514-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"62901ed9-929f-443e-ae1b-dc75cb2e1514\") " pod="openstack/cinder-backup-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.308925 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zfkfl\" (UniqueName: \"kubernetes.io/projected/62901ed9-929f-443e-ae1b-dc75cb2e1514-kube-api-access-zfkfl\") pod \"cinder-backup-0\" (UID: \"62901ed9-929f-443e-ae1b-dc75cb2e1514\") " pod="openstack/cinder-backup-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.308975 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/62901ed9-929f-443e-ae1b-dc75cb2e1514-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"62901ed9-929f-443e-ae1b-dc75cb2e1514\") " pod="openstack/cinder-backup-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.309013 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/62901ed9-929f-443e-ae1b-dc75cb2e1514-config-data-custom\") pod \"cinder-backup-0\" (UID: \"62901ed9-929f-443e-ae1b-dc75cb2e1514\") " pod="openstack/cinder-backup-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.309044 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62901ed9-929f-443e-ae1b-dc75cb2e1514-config-data\") pod \"cinder-backup-0\" (UID: \"62901ed9-929f-443e-ae1b-dc75cb2e1514\") " pod="openstack/cinder-backup-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.309057 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/62901ed9-929f-443e-ae1b-dc75cb2e1514-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"62901ed9-929f-443e-ae1b-dc75cb2e1514\") " pod="openstack/cinder-backup-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.309068 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/62901ed9-929f-443e-ae1b-dc75cb2e1514-dev\") pod \"cinder-backup-0\" (UID: \"62901ed9-929f-443e-ae1b-dc75cb2e1514\") " pod="openstack/cinder-backup-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.309174 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/62901ed9-929f-443e-ae1b-dc75cb2e1514-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"62901ed9-929f-443e-ae1b-dc75cb2e1514\") " pod="openstack/cinder-backup-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.309192 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/62901ed9-929f-443e-ae1b-dc75cb2e1514-dev\") pod \"cinder-backup-0\" (UID: \"62901ed9-929f-443e-ae1b-dc75cb2e1514\") " pod="openstack/cinder-backup-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.313026 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62901ed9-929f-443e-ae1b-dc75cb2e1514-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"62901ed9-929f-443e-ae1b-dc75cb2e1514\") " pod="openstack/cinder-backup-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.313128 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/62901ed9-929f-443e-ae1b-dc75cb2e1514-config-data-custom\") pod \"cinder-backup-0\" (UID: \"62901ed9-929f-443e-ae1b-dc75cb2e1514\") " pod="openstack/cinder-backup-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.314076 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62901ed9-929f-443e-ae1b-dc75cb2e1514-scripts\") pod \"cinder-backup-0\" (UID: \"62901ed9-929f-443e-ae1b-dc75cb2e1514\") " pod="openstack/cinder-backup-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.315193 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/62901ed9-929f-443e-ae1b-dc75cb2e1514-ceph\") pod \"cinder-backup-0\" (UID: \"62901ed9-929f-443e-ae1b-dc75cb2e1514\") " pod="openstack/cinder-backup-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.316427 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62901ed9-929f-443e-ae1b-dc75cb2e1514-config-data\") pod \"cinder-backup-0\" (UID: \"62901ed9-929f-443e-ae1b-dc75cb2e1514\") " pod="openstack/cinder-backup-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.329888 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zfkfl\" (UniqueName: \"kubernetes.io/projected/62901ed9-929f-443e-ae1b-dc75cb2e1514-kube-api-access-zfkfl\") pod \"cinder-backup-0\" (UID: \"62901ed9-929f-443e-ae1b-dc75cb2e1514\") " pod="openstack/cinder-backup-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.333452 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-volume1-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.382105 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.540205 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-db-create-2rtqh"] Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.557227 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-create-2rtqh"] Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.557327 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-2rtqh" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.618634 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cx8qc\" (UniqueName: \"kubernetes.io/projected/fc0bbde6-8d1e-4bd6-ab42-2c0549fa8792-kube-api-access-cx8qc\") pod \"manila-db-create-2rtqh\" (UID: \"fc0bbde6-8d1e-4bd6-ab42-2c0549fa8792\") " pod="openstack/manila-db-create-2rtqh" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.618721 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fc0bbde6-8d1e-4bd6-ab42-2c0549fa8792-operator-scripts\") pod \"manila-db-create-2rtqh\" (UID: \"fc0bbde6-8d1e-4bd6-ab42-2c0549fa8792\") " pod="openstack/manila-db-create-2rtqh" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.631595 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-b552-account-create-js8sh"] Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.633302 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-b552-account-create-js8sh" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.642088 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-db-secret" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.645223 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-b552-account-create-js8sh"] Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.720551 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cx8qc\" (UniqueName: \"kubernetes.io/projected/fc0bbde6-8d1e-4bd6-ab42-2c0549fa8792-kube-api-access-cx8qc\") pod \"manila-db-create-2rtqh\" (UID: \"fc0bbde6-8d1e-4bd6-ab42-2c0549fa8792\") " pod="openstack/manila-db-create-2rtqh" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.720613 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b2l9x\" (UniqueName: \"kubernetes.io/projected/5aa41617-1c85-4ccf-b21c-cc5bd6da4902-kube-api-access-b2l9x\") pod \"manila-b552-account-create-js8sh\" (UID: \"5aa41617-1c85-4ccf-b21c-cc5bd6da4902\") " pod="openstack/manila-b552-account-create-js8sh" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.720661 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5aa41617-1c85-4ccf-b21c-cc5bd6da4902-operator-scripts\") pod \"manila-b552-account-create-js8sh\" (UID: \"5aa41617-1c85-4ccf-b21c-cc5bd6da4902\") " pod="openstack/manila-b552-account-create-js8sh" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.720711 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fc0bbde6-8d1e-4bd6-ab42-2c0549fa8792-operator-scripts\") pod \"manila-db-create-2rtqh\" (UID: \"fc0bbde6-8d1e-4bd6-ab42-2c0549fa8792\") " pod="openstack/manila-db-create-2rtqh" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.721479 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fc0bbde6-8d1e-4bd6-ab42-2c0549fa8792-operator-scripts\") pod \"manila-db-create-2rtqh\" (UID: \"fc0bbde6-8d1e-4bd6-ab42-2c0549fa8792\") " pod="openstack/manila-db-create-2rtqh" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.744533 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cx8qc\" (UniqueName: \"kubernetes.io/projected/fc0bbde6-8d1e-4bd6-ab42-2c0549fa8792-kube-api-access-cx8qc\") pod \"manila-db-create-2rtqh\" (UID: \"fc0bbde6-8d1e-4bd6-ab42-2c0549fa8792\") " pod="openstack/manila-db-create-2rtqh" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.818047 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.819631 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.823614 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b2l9x\" (UniqueName: \"kubernetes.io/projected/5aa41617-1c85-4ccf-b21c-cc5bd6da4902-kube-api-access-b2l9x\") pod \"manila-b552-account-create-js8sh\" (UID: \"5aa41617-1c85-4ccf-b21c-cc5bd6da4902\") " pod="openstack/manila-b552-account-create-js8sh" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.823689 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5aa41617-1c85-4ccf-b21c-cc5bd6da4902-operator-scripts\") pod \"manila-b552-account-create-js8sh\" (UID: \"5aa41617-1c85-4ccf-b21c-cc5bd6da4902\") " pod="openstack/manila-b552-account-create-js8sh" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.824585 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5aa41617-1c85-4ccf-b21c-cc5bd6da4902-operator-scripts\") pod \"manila-b552-account-create-js8sh\" (UID: \"5aa41617-1c85-4ccf-b21c-cc5bd6da4902\") " pod="openstack/manila-b552-account-create-js8sh" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.825668 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.826309 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.826385 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.826665 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.827044 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-pd2j7" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.846891 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b2l9x\" (UniqueName: \"kubernetes.io/projected/5aa41617-1c85-4ccf-b21c-cc5bd6da4902-kube-api-access-b2l9x\") pod \"manila-b552-account-create-js8sh\" (UID: \"5aa41617-1c85-4ccf-b21c-cc5bd6da4902\") " pod="openstack/manila-b552-account-create-js8sh" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.871070 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.873023 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.875846 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.875953 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.889192 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-2rtqh" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.902276 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.927256 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4dktw\" (UniqueName: \"kubernetes.io/projected/9b280529-6f38-4c33-8f60-e2aad8fbb2a1-kube-api-access-4dktw\") pod \"glance-default-external-api-0\" (UID: \"9b280529-6f38-4c33-8f60-e2aad8fbb2a1\") " pod="openstack/glance-default-external-api-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.927914 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b280529-6f38-4c33-8f60-e2aad8fbb2a1-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"9b280529-6f38-4c33-8f60-e2aad8fbb2a1\") " pod="openstack/glance-default-external-api-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.927976 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"8660ff9f-23fb-4be2-93f4-330ffdab1566\") " pod="openstack/glance-default-internal-api-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.928001 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8660ff9f-23fb-4be2-93f4-330ffdab1566-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"8660ff9f-23fb-4be2-93f4-330ffdab1566\") " pod="openstack/glance-default-internal-api-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.928026 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"9b280529-6f38-4c33-8f60-e2aad8fbb2a1\") " pod="openstack/glance-default-external-api-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.928129 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9b280529-6f38-4c33-8f60-e2aad8fbb2a1-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"9b280529-6f38-4c33-8f60-e2aad8fbb2a1\") " pod="openstack/glance-default-external-api-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.928165 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8660ff9f-23fb-4be2-93f4-330ffdab1566-config-data\") pod \"glance-default-internal-api-0\" (UID: \"8660ff9f-23fb-4be2-93f4-330ffdab1566\") " pod="openstack/glance-default-internal-api-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.928192 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8660ff9f-23fb-4be2-93f4-330ffdab1566-logs\") pod \"glance-default-internal-api-0\" (UID: \"8660ff9f-23fb-4be2-93f4-330ffdab1566\") " pod="openstack/glance-default-internal-api-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.928218 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9b280529-6f38-4c33-8f60-e2aad8fbb2a1-scripts\") pod \"glance-default-external-api-0\" (UID: \"9b280529-6f38-4c33-8f60-e2aad8fbb2a1\") " pod="openstack/glance-default-external-api-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.928257 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9b280529-6f38-4c33-8f60-e2aad8fbb2a1-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"9b280529-6f38-4c33-8f60-e2aad8fbb2a1\") " pod="openstack/glance-default-external-api-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.928314 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9b280529-6f38-4c33-8f60-e2aad8fbb2a1-config-data\") pod \"glance-default-external-api-0\" (UID: \"9b280529-6f38-4c33-8f60-e2aad8fbb2a1\") " pod="openstack/glance-default-external-api-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.928348 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8660ff9f-23fb-4be2-93f4-330ffdab1566-scripts\") pod \"glance-default-internal-api-0\" (UID: \"8660ff9f-23fb-4be2-93f4-330ffdab1566\") " pod="openstack/glance-default-internal-api-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.928595 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-47x8x\" (UniqueName: \"kubernetes.io/projected/8660ff9f-23fb-4be2-93f4-330ffdab1566-kube-api-access-47x8x\") pod \"glance-default-internal-api-0\" (UID: \"8660ff9f-23fb-4be2-93f4-330ffdab1566\") " pod="openstack/glance-default-internal-api-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.928676 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9b280529-6f38-4c33-8f60-e2aad8fbb2a1-logs\") pod \"glance-default-external-api-0\" (UID: \"9b280529-6f38-4c33-8f60-e2aad8fbb2a1\") " pod="openstack/glance-default-external-api-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.928746 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8660ff9f-23fb-4be2-93f4-330ffdab1566-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"8660ff9f-23fb-4be2-93f4-330ffdab1566\") " pod="openstack/glance-default-internal-api-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.928869 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8660ff9f-23fb-4be2-93f4-330ffdab1566-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"8660ff9f-23fb-4be2-93f4-330ffdab1566\") " pod="openstack/glance-default-internal-api-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.928927 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/8660ff9f-23fb-4be2-93f4-330ffdab1566-ceph\") pod \"glance-default-internal-api-0\" (UID: \"8660ff9f-23fb-4be2-93f4-330ffdab1566\") " pod="openstack/glance-default-internal-api-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.928979 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/9b280529-6f38-4c33-8f60-e2aad8fbb2a1-ceph\") pod \"glance-default-external-api-0\" (UID: \"9b280529-6f38-4c33-8f60-e2aad8fbb2a1\") " pod="openstack/glance-default-external-api-0" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.960775 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-b552-account-create-js8sh" Nov 22 08:53:17 crc kubenswrapper[4789]: I1122 08:53:17.994566 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-volume1-0"] Nov 22 08:53:18 crc kubenswrapper[4789]: W1122 08:53:18.011332 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2b1ad6b9_9f67_4533_bafd_5933ade7f938.slice/crio-ffd3f97c2a0dbd8339363e6b3f827ad266fc679f9c0b45dd2ec59e9679242fc3 WatchSource:0}: Error finding container ffd3f97c2a0dbd8339363e6b3f827ad266fc679f9c0b45dd2ec59e9679242fc3: Status 404 returned error can't find the container with id ffd3f97c2a0dbd8339363e6b3f827ad266fc679f9c0b45dd2ec59e9679242fc3 Nov 22 08:53:18 crc kubenswrapper[4789]: I1122 08:53:18.030793 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9b280529-6f38-4c33-8f60-e2aad8fbb2a1-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"9b280529-6f38-4c33-8f60-e2aad8fbb2a1\") " pod="openstack/glance-default-external-api-0" Nov 22 08:53:18 crc kubenswrapper[4789]: I1122 08:53:18.030844 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8660ff9f-23fb-4be2-93f4-330ffdab1566-config-data\") pod \"glance-default-internal-api-0\" (UID: \"8660ff9f-23fb-4be2-93f4-330ffdab1566\") " pod="openstack/glance-default-internal-api-0" Nov 22 08:53:18 crc kubenswrapper[4789]: I1122 08:53:18.030859 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8660ff9f-23fb-4be2-93f4-330ffdab1566-logs\") pod \"glance-default-internal-api-0\" (UID: \"8660ff9f-23fb-4be2-93f4-330ffdab1566\") " pod="openstack/glance-default-internal-api-0" Nov 22 08:53:18 crc kubenswrapper[4789]: I1122 08:53:18.030875 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9b280529-6f38-4c33-8f60-e2aad8fbb2a1-scripts\") pod \"glance-default-external-api-0\" (UID: \"9b280529-6f38-4c33-8f60-e2aad8fbb2a1\") " pod="openstack/glance-default-external-api-0" Nov 22 08:53:18 crc kubenswrapper[4789]: I1122 08:53:18.030897 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9b280529-6f38-4c33-8f60-e2aad8fbb2a1-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"9b280529-6f38-4c33-8f60-e2aad8fbb2a1\") " pod="openstack/glance-default-external-api-0" Nov 22 08:53:18 crc kubenswrapper[4789]: I1122 08:53:18.030922 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9b280529-6f38-4c33-8f60-e2aad8fbb2a1-config-data\") pod \"glance-default-external-api-0\" (UID: \"9b280529-6f38-4c33-8f60-e2aad8fbb2a1\") " pod="openstack/glance-default-external-api-0" Nov 22 08:53:18 crc kubenswrapper[4789]: I1122 08:53:18.030939 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8660ff9f-23fb-4be2-93f4-330ffdab1566-scripts\") pod \"glance-default-internal-api-0\" (UID: \"8660ff9f-23fb-4be2-93f4-330ffdab1566\") " pod="openstack/glance-default-internal-api-0" Nov 22 08:53:18 crc kubenswrapper[4789]: I1122 08:53:18.030963 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-47x8x\" (UniqueName: \"kubernetes.io/projected/8660ff9f-23fb-4be2-93f4-330ffdab1566-kube-api-access-47x8x\") pod \"glance-default-internal-api-0\" (UID: \"8660ff9f-23fb-4be2-93f4-330ffdab1566\") " pod="openstack/glance-default-internal-api-0" Nov 22 08:53:18 crc kubenswrapper[4789]: I1122 08:53:18.031005 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9b280529-6f38-4c33-8f60-e2aad8fbb2a1-logs\") pod \"glance-default-external-api-0\" (UID: \"9b280529-6f38-4c33-8f60-e2aad8fbb2a1\") " pod="openstack/glance-default-external-api-0" Nov 22 08:53:18 crc kubenswrapper[4789]: I1122 08:53:18.031035 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8660ff9f-23fb-4be2-93f4-330ffdab1566-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"8660ff9f-23fb-4be2-93f4-330ffdab1566\") " pod="openstack/glance-default-internal-api-0" Nov 22 08:53:18 crc kubenswrapper[4789]: I1122 08:53:18.031080 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8660ff9f-23fb-4be2-93f4-330ffdab1566-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"8660ff9f-23fb-4be2-93f4-330ffdab1566\") " pod="openstack/glance-default-internal-api-0" Nov 22 08:53:18 crc kubenswrapper[4789]: I1122 08:53:18.031125 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/8660ff9f-23fb-4be2-93f4-330ffdab1566-ceph\") pod \"glance-default-internal-api-0\" (UID: \"8660ff9f-23fb-4be2-93f4-330ffdab1566\") " pod="openstack/glance-default-internal-api-0" Nov 22 08:53:18 crc kubenswrapper[4789]: I1122 08:53:18.031157 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/9b280529-6f38-4c33-8f60-e2aad8fbb2a1-ceph\") pod \"glance-default-external-api-0\" (UID: \"9b280529-6f38-4c33-8f60-e2aad8fbb2a1\") " pod="openstack/glance-default-external-api-0" Nov 22 08:53:18 crc kubenswrapper[4789]: I1122 08:53:18.031195 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4dktw\" (UniqueName: \"kubernetes.io/projected/9b280529-6f38-4c33-8f60-e2aad8fbb2a1-kube-api-access-4dktw\") pod \"glance-default-external-api-0\" (UID: \"9b280529-6f38-4c33-8f60-e2aad8fbb2a1\") " pod="openstack/glance-default-external-api-0" Nov 22 08:53:18 crc kubenswrapper[4789]: I1122 08:53:18.031225 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b280529-6f38-4c33-8f60-e2aad8fbb2a1-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"9b280529-6f38-4c33-8f60-e2aad8fbb2a1\") " pod="openstack/glance-default-external-api-0" Nov 22 08:53:18 crc kubenswrapper[4789]: I1122 08:53:18.031257 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"8660ff9f-23fb-4be2-93f4-330ffdab1566\") " pod="openstack/glance-default-internal-api-0" Nov 22 08:53:18 crc kubenswrapper[4789]: I1122 08:53:18.031273 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8660ff9f-23fb-4be2-93f4-330ffdab1566-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"8660ff9f-23fb-4be2-93f4-330ffdab1566\") " pod="openstack/glance-default-internal-api-0" Nov 22 08:53:18 crc kubenswrapper[4789]: I1122 08:53:18.031291 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"9b280529-6f38-4c33-8f60-e2aad8fbb2a1\") " pod="openstack/glance-default-external-api-0" Nov 22 08:53:18 crc kubenswrapper[4789]: I1122 08:53:18.031669 4789 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"9b280529-6f38-4c33-8f60-e2aad8fbb2a1\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/glance-default-external-api-0" Nov 22 08:53:18 crc kubenswrapper[4789]: I1122 08:53:18.033007 4789 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"8660ff9f-23fb-4be2-93f4-330ffdab1566\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/glance-default-internal-api-0" Nov 22 08:53:18 crc kubenswrapper[4789]: I1122 08:53:18.033924 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8660ff9f-23fb-4be2-93f4-330ffdab1566-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"8660ff9f-23fb-4be2-93f4-330ffdab1566\") " pod="openstack/glance-default-internal-api-0" Nov 22 08:53:18 crc kubenswrapper[4789]: I1122 08:53:18.035557 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9b280529-6f38-4c33-8f60-e2aad8fbb2a1-logs\") pod \"glance-default-external-api-0\" (UID: \"9b280529-6f38-4c33-8f60-e2aad8fbb2a1\") " pod="openstack/glance-default-external-api-0" Nov 22 08:53:18 crc kubenswrapper[4789]: I1122 08:53:18.035875 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8660ff9f-23fb-4be2-93f4-330ffdab1566-logs\") pod \"glance-default-internal-api-0\" (UID: \"8660ff9f-23fb-4be2-93f4-330ffdab1566\") " pod="openstack/glance-default-internal-api-0" Nov 22 08:53:18 crc kubenswrapper[4789]: I1122 08:53:18.039020 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9b280529-6f38-4c33-8f60-e2aad8fbb2a1-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"9b280529-6f38-4c33-8f60-e2aad8fbb2a1\") " pod="openstack/glance-default-external-api-0" Nov 22 08:53:18 crc kubenswrapper[4789]: I1122 08:53:18.043131 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8660ff9f-23fb-4be2-93f4-330ffdab1566-scripts\") pod \"glance-default-internal-api-0\" (UID: \"8660ff9f-23fb-4be2-93f4-330ffdab1566\") " pod="openstack/glance-default-internal-api-0" Nov 22 08:53:18 crc kubenswrapper[4789]: I1122 08:53:18.043565 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9b280529-6f38-4c33-8f60-e2aad8fbb2a1-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"9b280529-6f38-4c33-8f60-e2aad8fbb2a1\") " pod="openstack/glance-default-external-api-0" Nov 22 08:53:18 crc kubenswrapper[4789]: I1122 08:53:18.043671 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/8660ff9f-23fb-4be2-93f4-330ffdab1566-ceph\") pod \"glance-default-internal-api-0\" (UID: \"8660ff9f-23fb-4be2-93f4-330ffdab1566\") " pod="openstack/glance-default-internal-api-0" Nov 22 08:53:18 crc kubenswrapper[4789]: I1122 08:53:18.044257 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8660ff9f-23fb-4be2-93f4-330ffdab1566-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"8660ff9f-23fb-4be2-93f4-330ffdab1566\") " pod="openstack/glance-default-internal-api-0" Nov 22 08:53:18 crc kubenswrapper[4789]: I1122 08:53:18.044304 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9b280529-6f38-4c33-8f60-e2aad8fbb2a1-config-data\") pod \"glance-default-external-api-0\" (UID: \"9b280529-6f38-4c33-8f60-e2aad8fbb2a1\") " pod="openstack/glance-default-external-api-0" Nov 22 08:53:18 crc kubenswrapper[4789]: I1122 08:53:18.044844 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9b280529-6f38-4c33-8f60-e2aad8fbb2a1-scripts\") pod \"glance-default-external-api-0\" (UID: \"9b280529-6f38-4c33-8f60-e2aad8fbb2a1\") " pod="openstack/glance-default-external-api-0" Nov 22 08:53:18 crc kubenswrapper[4789]: I1122 08:53:18.047894 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/9b280529-6f38-4c33-8f60-e2aad8fbb2a1-ceph\") pod \"glance-default-external-api-0\" (UID: \"9b280529-6f38-4c33-8f60-e2aad8fbb2a1\") " pod="openstack/glance-default-external-api-0" Nov 22 08:53:18 crc kubenswrapper[4789]: I1122 08:53:18.048794 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8660ff9f-23fb-4be2-93f4-330ffdab1566-config-data\") pod \"glance-default-internal-api-0\" (UID: \"8660ff9f-23fb-4be2-93f4-330ffdab1566\") " pod="openstack/glance-default-internal-api-0" Nov 22 08:53:18 crc kubenswrapper[4789]: I1122 08:53:18.056132 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8660ff9f-23fb-4be2-93f4-330ffdab1566-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"8660ff9f-23fb-4be2-93f4-330ffdab1566\") " pod="openstack/glance-default-internal-api-0" Nov 22 08:53:18 crc kubenswrapper[4789]: I1122 08:53:18.056725 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4dktw\" (UniqueName: \"kubernetes.io/projected/9b280529-6f38-4c33-8f60-e2aad8fbb2a1-kube-api-access-4dktw\") pod \"glance-default-external-api-0\" (UID: \"9b280529-6f38-4c33-8f60-e2aad8fbb2a1\") " pod="openstack/glance-default-external-api-0" Nov 22 08:53:18 crc kubenswrapper[4789]: I1122 08:53:18.064419 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b280529-6f38-4c33-8f60-e2aad8fbb2a1-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"9b280529-6f38-4c33-8f60-e2aad8fbb2a1\") " pod="openstack/glance-default-external-api-0" Nov 22 08:53:18 crc kubenswrapper[4789]: I1122 08:53:18.064998 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-47x8x\" (UniqueName: \"kubernetes.io/projected/8660ff9f-23fb-4be2-93f4-330ffdab1566-kube-api-access-47x8x\") pod \"glance-default-internal-api-0\" (UID: \"8660ff9f-23fb-4be2-93f4-330ffdab1566\") " pod="openstack/glance-default-internal-api-0" Nov 22 08:53:18 crc kubenswrapper[4789]: I1122 08:53:18.085532 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Nov 22 08:53:18 crc kubenswrapper[4789]: I1122 08:53:18.091569 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"9b280529-6f38-4c33-8f60-e2aad8fbb2a1\") " pod="openstack/glance-default-external-api-0" Nov 22 08:53:18 crc kubenswrapper[4789]: I1122 08:53:18.094622 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"8660ff9f-23fb-4be2-93f4-330ffdab1566\") " pod="openstack/glance-default-internal-api-0" Nov 22 08:53:18 crc kubenswrapper[4789]: I1122 08:53:18.144519 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 22 08:53:18 crc kubenswrapper[4789]: I1122 08:53:18.203410 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 22 08:53:18 crc kubenswrapper[4789]: I1122 08:53:18.397245 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-create-2rtqh"] Nov 22 08:53:18 crc kubenswrapper[4789]: W1122 08:53:18.404429 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfc0bbde6_8d1e_4bd6_ab42_2c0549fa8792.slice/crio-e5dcab6868eca0380fa3d79fadadf5a2506b5c840096755cdf8e33317ca9958b WatchSource:0}: Error finding container e5dcab6868eca0380fa3d79fadadf5a2506b5c840096755cdf8e33317ca9958b: Status 404 returned error can't find the container with id e5dcab6868eca0380fa3d79fadadf5a2506b5c840096755cdf8e33317ca9958b Nov 22 08:53:18 crc kubenswrapper[4789]: I1122 08:53:18.481181 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-b552-account-create-js8sh"] Nov 22 08:53:18 crc kubenswrapper[4789]: W1122 08:53:18.481847 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5aa41617_1c85_4ccf_b21c_cc5bd6da4902.slice/crio-df3df802e83b7b92885362aa32d90170d4d2c7007fa9329d1e017711f34366c0 WatchSource:0}: Error finding container df3df802e83b7b92885362aa32d90170d4d2c7007fa9329d1e017711f34366c0: Status 404 returned error can't find the container with id df3df802e83b7b92885362aa32d90170d4d2c7007fa9329d1e017711f34366c0 Nov 22 08:53:18 crc kubenswrapper[4789]: I1122 08:53:18.712681 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 22 08:53:18 crc kubenswrapper[4789]: W1122 08:53:18.717569 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9b280529_6f38_4c33_8f60_e2aad8fbb2a1.slice/crio-f89388ec32f81b8ceac0c9474aa547a54584eaa59b241a5fc8b87692cdb38f59 WatchSource:0}: Error finding container f89388ec32f81b8ceac0c9474aa547a54584eaa59b241a5fc8b87692cdb38f59: Status 404 returned error can't find the container with id f89388ec32f81b8ceac0c9474aa547a54584eaa59b241a5fc8b87692cdb38f59 Nov 22 08:53:18 crc kubenswrapper[4789]: I1122 08:53:18.813113 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-b552-account-create-js8sh" event={"ID":"5aa41617-1c85-4ccf-b21c-cc5bd6da4902","Type":"ContainerStarted","Data":"0d9ed7230b97e612c06435edfe9acc03c8828ac2700f3e2d6269238323be0f0f"} Nov 22 08:53:18 crc kubenswrapper[4789]: I1122 08:53:18.813157 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-b552-account-create-js8sh" event={"ID":"5aa41617-1c85-4ccf-b21c-cc5bd6da4902","Type":"ContainerStarted","Data":"df3df802e83b7b92885362aa32d90170d4d2c7007fa9329d1e017711f34366c0"} Nov 22 08:53:18 crc kubenswrapper[4789]: I1122 08:53:18.823693 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"62901ed9-929f-443e-ae1b-dc75cb2e1514","Type":"ContainerStarted","Data":"62c549deb39b9af73f39de233f622f8a1aa8c1c37ca9d56e6e4ac30a41054b38"} Nov 22 08:53:18 crc kubenswrapper[4789]: I1122 08:53:18.825117 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9b280529-6f38-4c33-8f60-e2aad8fbb2a1","Type":"ContainerStarted","Data":"f89388ec32f81b8ceac0c9474aa547a54584eaa59b241a5fc8b87692cdb38f59"} Nov 22 08:53:18 crc kubenswrapper[4789]: I1122 08:53:18.826556 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-2rtqh" event={"ID":"fc0bbde6-8d1e-4bd6-ab42-2c0549fa8792","Type":"ContainerStarted","Data":"51c3198d42a0b4bc7b03f39d5f91f5e164345794c02ae80a3f8806a0b1a4ef07"} Nov 22 08:53:18 crc kubenswrapper[4789]: I1122 08:53:18.826648 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-2rtqh" event={"ID":"fc0bbde6-8d1e-4bd6-ab42-2c0549fa8792","Type":"ContainerStarted","Data":"e5dcab6868eca0380fa3d79fadadf5a2506b5c840096755cdf8e33317ca9958b"} Nov 22 08:53:18 crc kubenswrapper[4789]: I1122 08:53:18.829142 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"2b1ad6b9-9f67-4533-bafd-5933ade7f938","Type":"ContainerStarted","Data":"ffd3f97c2a0dbd8339363e6b3f827ad266fc679f9c0b45dd2ec59e9679242fc3"} Nov 22 08:53:18 crc kubenswrapper[4789]: I1122 08:53:18.830113 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 22 08:53:18 crc kubenswrapper[4789]: W1122 08:53:18.837413 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8660ff9f_23fb_4be2_93f4_330ffdab1566.slice/crio-ea91d231902b06cdb3418a1cbfb55d58bbfa9caef45994127cd0c929f008a188 WatchSource:0}: Error finding container ea91d231902b06cdb3418a1cbfb55d58bbfa9caef45994127cd0c929f008a188: Status 404 returned error can't find the container with id ea91d231902b06cdb3418a1cbfb55d58bbfa9caef45994127cd0c929f008a188 Nov 22 08:53:19 crc kubenswrapper[4789]: I1122 08:53:19.838087 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"8660ff9f-23fb-4be2-93f4-330ffdab1566","Type":"ContainerStarted","Data":"ea91d231902b06cdb3418a1cbfb55d58bbfa9caef45994127cd0c929f008a188"} Nov 22 08:53:19 crc kubenswrapper[4789]: I1122 08:53:19.840424 4789 generic.go:334] "Generic (PLEG): container finished" podID="fc0bbde6-8d1e-4bd6-ab42-2c0549fa8792" containerID="51c3198d42a0b4bc7b03f39d5f91f5e164345794c02ae80a3f8806a0b1a4ef07" exitCode=0 Nov 22 08:53:19 crc kubenswrapper[4789]: I1122 08:53:19.840499 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-2rtqh" event={"ID":"fc0bbde6-8d1e-4bd6-ab42-2c0549fa8792","Type":"ContainerDied","Data":"51c3198d42a0b4bc7b03f39d5f91f5e164345794c02ae80a3f8806a0b1a4ef07"} Nov 22 08:53:19 crc kubenswrapper[4789]: I1122 08:53:19.860983 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-b552-account-create-js8sh" podStartSLOduration=2.860850756 podStartE2EDuration="2.860850756s" podCreationTimestamp="2025-11-22 08:53:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:53:19.854983037 +0000 UTC m=+3414.089383310" watchObservedRunningTime="2025-11-22 08:53:19.860850756 +0000 UTC m=+3414.095251029" Nov 22 08:53:20 crc kubenswrapper[4789]: I1122 08:53:20.850035 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"8660ff9f-23fb-4be2-93f4-330ffdab1566","Type":"ContainerStarted","Data":"d55d1a7b39925afeeca72138debfa45ac42c8b62ca4fb09f550e0d367581b155"} Nov 22 08:53:20 crc kubenswrapper[4789]: I1122 08:53:20.852430 4789 generic.go:334] "Generic (PLEG): container finished" podID="5aa41617-1c85-4ccf-b21c-cc5bd6da4902" containerID="0d9ed7230b97e612c06435edfe9acc03c8828ac2700f3e2d6269238323be0f0f" exitCode=0 Nov 22 08:53:20 crc kubenswrapper[4789]: I1122 08:53:20.852486 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-b552-account-create-js8sh" event={"ID":"5aa41617-1c85-4ccf-b21c-cc5bd6da4902","Type":"ContainerDied","Data":"0d9ed7230b97e612c06435edfe9acc03c8828ac2700f3e2d6269238323be0f0f"} Nov 22 08:53:20 crc kubenswrapper[4789]: I1122 08:53:20.854285 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"62901ed9-929f-443e-ae1b-dc75cb2e1514","Type":"ContainerStarted","Data":"803487acc17b691ffaf01065aef83b94965e8af27cde97f225c51de18baa9fb2"} Nov 22 08:53:20 crc kubenswrapper[4789]: I1122 08:53:20.854319 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"62901ed9-929f-443e-ae1b-dc75cb2e1514","Type":"ContainerStarted","Data":"60fbbdb7c30637e0499a6af4c308dbf9c900cbd2e0eadc547ebbba78be2b2067"} Nov 22 08:53:20 crc kubenswrapper[4789]: I1122 08:53:20.861741 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9b280529-6f38-4c33-8f60-e2aad8fbb2a1","Type":"ContainerStarted","Data":"40ead2b10dba116f2634dc21458cf96e5e43eaac7a5fed7e06d3ffe7a3f98797"} Nov 22 08:53:20 crc kubenswrapper[4789]: I1122 08:53:20.864576 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"2b1ad6b9-9f67-4533-bafd-5933ade7f938","Type":"ContainerStarted","Data":"955aee6e72ecb2c64f0f45b7e2c221b714a3ad97065504c2d3e07dd527265606"} Nov 22 08:53:20 crc kubenswrapper[4789]: I1122 08:53:20.864621 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"2b1ad6b9-9f67-4533-bafd-5933ade7f938","Type":"ContainerStarted","Data":"7e7b648a44d0b359c204843d0f141b1b018b434f92f75da7ce827f441adf063d"} Nov 22 08:53:20 crc kubenswrapper[4789]: I1122 08:53:20.903407 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-backup-0" podStartSLOduration=1.777730881 podStartE2EDuration="3.903385444s" podCreationTimestamp="2025-11-22 08:53:17 +0000 UTC" firstStartedPulling="2025-11-22 08:53:18.112384379 +0000 UTC m=+3412.346784652" lastFinishedPulling="2025-11-22 08:53:20.238038952 +0000 UTC m=+3414.472439215" observedRunningTime="2025-11-22 08:53:20.896291452 +0000 UTC m=+3415.130691745" watchObservedRunningTime="2025-11-22 08:53:20.903385444 +0000 UTC m=+3415.137785717" Nov 22 08:53:20 crc kubenswrapper[4789]: I1122 08:53:20.934947 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-volume-volume1-0" podStartSLOduration=2.712418734 podStartE2EDuration="4.934901926s" podCreationTimestamp="2025-11-22 08:53:16 +0000 UTC" firstStartedPulling="2025-11-22 08:53:18.014546732 +0000 UTC m=+3412.248947005" lastFinishedPulling="2025-11-22 08:53:20.237029924 +0000 UTC m=+3414.471430197" observedRunningTime="2025-11-22 08:53:20.927126836 +0000 UTC m=+3415.161527119" watchObservedRunningTime="2025-11-22 08:53:20.934901926 +0000 UTC m=+3415.169302189" Nov 22 08:53:21 crc kubenswrapper[4789]: I1122 08:53:21.178737 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-2rtqh" Nov 22 08:53:21 crc kubenswrapper[4789]: I1122 08:53:21.204964 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cx8qc\" (UniqueName: \"kubernetes.io/projected/fc0bbde6-8d1e-4bd6-ab42-2c0549fa8792-kube-api-access-cx8qc\") pod \"fc0bbde6-8d1e-4bd6-ab42-2c0549fa8792\" (UID: \"fc0bbde6-8d1e-4bd6-ab42-2c0549fa8792\") " Nov 22 08:53:21 crc kubenswrapper[4789]: I1122 08:53:21.205115 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fc0bbde6-8d1e-4bd6-ab42-2c0549fa8792-operator-scripts\") pod \"fc0bbde6-8d1e-4bd6-ab42-2c0549fa8792\" (UID: \"fc0bbde6-8d1e-4bd6-ab42-2c0549fa8792\") " Nov 22 08:53:21 crc kubenswrapper[4789]: I1122 08:53:21.206432 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fc0bbde6-8d1e-4bd6-ab42-2c0549fa8792-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "fc0bbde6-8d1e-4bd6-ab42-2c0549fa8792" (UID: "fc0bbde6-8d1e-4bd6-ab42-2c0549fa8792"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:53:21 crc kubenswrapper[4789]: I1122 08:53:21.213096 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc0bbde6-8d1e-4bd6-ab42-2c0549fa8792-kube-api-access-cx8qc" (OuterVolumeSpecName: "kube-api-access-cx8qc") pod "fc0bbde6-8d1e-4bd6-ab42-2c0549fa8792" (UID: "fc0bbde6-8d1e-4bd6-ab42-2c0549fa8792"). InnerVolumeSpecName "kube-api-access-cx8qc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:53:21 crc kubenswrapper[4789]: I1122 08:53:21.307687 4789 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fc0bbde6-8d1e-4bd6-ab42-2c0549fa8792-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:53:21 crc kubenswrapper[4789]: I1122 08:53:21.308113 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cx8qc\" (UniqueName: \"kubernetes.io/projected/fc0bbde6-8d1e-4bd6-ab42-2c0549fa8792-kube-api-access-cx8qc\") on node \"crc\" DevicePath \"\"" Nov 22 08:53:21 crc kubenswrapper[4789]: I1122 08:53:21.875553 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9b280529-6f38-4c33-8f60-e2aad8fbb2a1","Type":"ContainerStarted","Data":"c9fa35eee2ca07425c72bdae061209d5d60bc63bbd3a1f17729fdffe4c3cee5c"} Nov 22 08:53:21 crc kubenswrapper[4789]: I1122 08:53:21.878762 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-2rtqh" Nov 22 08:53:21 crc kubenswrapper[4789]: I1122 08:53:21.879725 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-2rtqh" event={"ID":"fc0bbde6-8d1e-4bd6-ab42-2c0549fa8792","Type":"ContainerDied","Data":"e5dcab6868eca0380fa3d79fadadf5a2506b5c840096755cdf8e33317ca9958b"} Nov 22 08:53:21 crc kubenswrapper[4789]: I1122 08:53:21.879802 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e5dcab6868eca0380fa3d79fadadf5a2506b5c840096755cdf8e33317ca9958b" Nov 22 08:53:21 crc kubenswrapper[4789]: I1122 08:53:21.883097 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"8660ff9f-23fb-4be2-93f4-330ffdab1566","Type":"ContainerStarted","Data":"f1ab4a0aebe84f05b6a1df9dc6632895207a2cea091b59f23b95de34cf39bd90"} Nov 22 08:53:21 crc kubenswrapper[4789]: I1122 08:53:21.927519 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=5.927482042 podStartE2EDuration="5.927482042s" podCreationTimestamp="2025-11-22 08:53:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:53:21.895195608 +0000 UTC m=+3416.129595921" watchObservedRunningTime="2025-11-22 08:53:21.927482042 +0000 UTC m=+3416.161882315" Nov 22 08:53:21 crc kubenswrapper[4789]: I1122 08:53:21.937698 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=5.9376756870000005 podStartE2EDuration="5.937675687s" podCreationTimestamp="2025-11-22 08:53:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:53:21.926699671 +0000 UTC m=+3416.161099954" watchObservedRunningTime="2025-11-22 08:53:21.937675687 +0000 UTC m=+3416.172075960" Nov 22 08:53:22 crc kubenswrapper[4789]: I1122 08:53:22.240896 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-b552-account-create-js8sh" Nov 22 08:53:22 crc kubenswrapper[4789]: I1122 08:53:22.334219 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-volume-volume1-0" Nov 22 08:53:22 crc kubenswrapper[4789]: I1122 08:53:22.383599 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-backup-0" Nov 22 08:53:22 crc kubenswrapper[4789]: I1122 08:53:22.432805 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5aa41617-1c85-4ccf-b21c-cc5bd6da4902-operator-scripts\") pod \"5aa41617-1c85-4ccf-b21c-cc5bd6da4902\" (UID: \"5aa41617-1c85-4ccf-b21c-cc5bd6da4902\") " Nov 22 08:53:22 crc kubenswrapper[4789]: I1122 08:53:22.432900 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b2l9x\" (UniqueName: \"kubernetes.io/projected/5aa41617-1c85-4ccf-b21c-cc5bd6da4902-kube-api-access-b2l9x\") pod \"5aa41617-1c85-4ccf-b21c-cc5bd6da4902\" (UID: \"5aa41617-1c85-4ccf-b21c-cc5bd6da4902\") " Nov 22 08:53:22 crc kubenswrapper[4789]: I1122 08:53:22.433296 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5aa41617-1c85-4ccf-b21c-cc5bd6da4902-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5aa41617-1c85-4ccf-b21c-cc5bd6da4902" (UID: "5aa41617-1c85-4ccf-b21c-cc5bd6da4902"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:53:22 crc kubenswrapper[4789]: I1122 08:53:22.433546 4789 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5aa41617-1c85-4ccf-b21c-cc5bd6da4902-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:53:22 crc kubenswrapper[4789]: I1122 08:53:22.447043 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5aa41617-1c85-4ccf-b21c-cc5bd6da4902-kube-api-access-b2l9x" (OuterVolumeSpecName: "kube-api-access-b2l9x") pod "5aa41617-1c85-4ccf-b21c-cc5bd6da4902" (UID: "5aa41617-1c85-4ccf-b21c-cc5bd6da4902"). InnerVolumeSpecName "kube-api-access-b2l9x". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:53:22 crc kubenswrapper[4789]: I1122 08:53:22.536886 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b2l9x\" (UniqueName: \"kubernetes.io/projected/5aa41617-1c85-4ccf-b21c-cc5bd6da4902-kube-api-access-b2l9x\") on node \"crc\" DevicePath \"\"" Nov 22 08:53:22 crc kubenswrapper[4789]: I1122 08:53:22.896732 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-b552-account-create-js8sh" Nov 22 08:53:22 crc kubenswrapper[4789]: I1122 08:53:22.898368 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-b552-account-create-js8sh" event={"ID":"5aa41617-1c85-4ccf-b21c-cc5bd6da4902","Type":"ContainerDied","Data":"df3df802e83b7b92885362aa32d90170d4d2c7007fa9329d1e017711f34366c0"} Nov 22 08:53:22 crc kubenswrapper[4789]: I1122 08:53:22.898407 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="df3df802e83b7b92885362aa32d90170d4d2c7007fa9329d1e017711f34366c0" Nov 22 08:53:27 crc kubenswrapper[4789]: I1122 08:53:27.539247 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-volume-volume1-0" Nov 22 08:53:27 crc kubenswrapper[4789]: I1122 08:53:27.574976 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-backup-0" Nov 22 08:53:28 crc kubenswrapper[4789]: I1122 08:53:28.078783 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-db-sync-jbx7v"] Nov 22 08:53:28 crc kubenswrapper[4789]: E1122 08:53:28.079481 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5aa41617-1c85-4ccf-b21c-cc5bd6da4902" containerName="mariadb-account-create" Nov 22 08:53:28 crc kubenswrapper[4789]: I1122 08:53:28.079505 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="5aa41617-1c85-4ccf-b21c-cc5bd6da4902" containerName="mariadb-account-create" Nov 22 08:53:28 crc kubenswrapper[4789]: E1122 08:53:28.079532 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc0bbde6-8d1e-4bd6-ab42-2c0549fa8792" containerName="mariadb-database-create" Nov 22 08:53:28 crc kubenswrapper[4789]: I1122 08:53:28.079539 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc0bbde6-8d1e-4bd6-ab42-2c0549fa8792" containerName="mariadb-database-create" Nov 22 08:53:28 crc kubenswrapper[4789]: I1122 08:53:28.079716 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="5aa41617-1c85-4ccf-b21c-cc5bd6da4902" containerName="mariadb-account-create" Nov 22 08:53:28 crc kubenswrapper[4789]: I1122 08:53:28.079741 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc0bbde6-8d1e-4bd6-ab42-2c0549fa8792" containerName="mariadb-database-create" Nov 22 08:53:28 crc kubenswrapper[4789]: I1122 08:53:28.080485 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-jbx7v" Nov 22 08:53:28 crc kubenswrapper[4789]: I1122 08:53:28.085936 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-manila-dockercfg-fzw4x" Nov 22 08:53:28 crc kubenswrapper[4789]: I1122 08:53:28.086111 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-config-data" Nov 22 08:53:28 crc kubenswrapper[4789]: I1122 08:53:28.090942 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-sync-jbx7v"] Nov 22 08:53:28 crc kubenswrapper[4789]: I1122 08:53:28.144648 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 22 08:53:28 crc kubenswrapper[4789]: I1122 08:53:28.144685 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 22 08:53:28 crc kubenswrapper[4789]: I1122 08:53:28.174456 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 22 08:53:28 crc kubenswrapper[4789]: I1122 08:53:28.184313 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 22 08:53:28 crc kubenswrapper[4789]: I1122 08:53:28.204980 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 22 08:53:28 crc kubenswrapper[4789]: I1122 08:53:28.205035 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 22 08:53:28 crc kubenswrapper[4789]: I1122 08:53:28.240025 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tp9wg\" (UniqueName: \"kubernetes.io/projected/6ff92fe7-35b8-43a0-9d62-cc825720ae76-kube-api-access-tp9wg\") pod \"manila-db-sync-jbx7v\" (UID: \"6ff92fe7-35b8-43a0-9d62-cc825720ae76\") " pod="openstack/manila-db-sync-jbx7v" Nov 22 08:53:28 crc kubenswrapper[4789]: I1122 08:53:28.240103 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ff92fe7-35b8-43a0-9d62-cc825720ae76-combined-ca-bundle\") pod \"manila-db-sync-jbx7v\" (UID: \"6ff92fe7-35b8-43a0-9d62-cc825720ae76\") " pod="openstack/manila-db-sync-jbx7v" Nov 22 08:53:28 crc kubenswrapper[4789]: I1122 08:53:28.240941 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ff92fe7-35b8-43a0-9d62-cc825720ae76-config-data\") pod \"manila-db-sync-jbx7v\" (UID: \"6ff92fe7-35b8-43a0-9d62-cc825720ae76\") " pod="openstack/manila-db-sync-jbx7v" Nov 22 08:53:28 crc kubenswrapper[4789]: I1122 08:53:28.241449 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/6ff92fe7-35b8-43a0-9d62-cc825720ae76-job-config-data\") pod \"manila-db-sync-jbx7v\" (UID: \"6ff92fe7-35b8-43a0-9d62-cc825720ae76\") " pod="openstack/manila-db-sync-jbx7v" Nov 22 08:53:28 crc kubenswrapper[4789]: I1122 08:53:28.241735 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 22 08:53:28 crc kubenswrapper[4789]: I1122 08:53:28.247233 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 22 08:53:28 crc kubenswrapper[4789]: I1122 08:53:28.343812 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ff92fe7-35b8-43a0-9d62-cc825720ae76-combined-ca-bundle\") pod \"manila-db-sync-jbx7v\" (UID: \"6ff92fe7-35b8-43a0-9d62-cc825720ae76\") " pod="openstack/manila-db-sync-jbx7v" Nov 22 08:53:28 crc kubenswrapper[4789]: I1122 08:53:28.343937 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ff92fe7-35b8-43a0-9d62-cc825720ae76-config-data\") pod \"manila-db-sync-jbx7v\" (UID: \"6ff92fe7-35b8-43a0-9d62-cc825720ae76\") " pod="openstack/manila-db-sync-jbx7v" Nov 22 08:53:28 crc kubenswrapper[4789]: I1122 08:53:28.344121 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/6ff92fe7-35b8-43a0-9d62-cc825720ae76-job-config-data\") pod \"manila-db-sync-jbx7v\" (UID: \"6ff92fe7-35b8-43a0-9d62-cc825720ae76\") " pod="openstack/manila-db-sync-jbx7v" Nov 22 08:53:28 crc kubenswrapper[4789]: I1122 08:53:28.344155 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tp9wg\" (UniqueName: \"kubernetes.io/projected/6ff92fe7-35b8-43a0-9d62-cc825720ae76-kube-api-access-tp9wg\") pod \"manila-db-sync-jbx7v\" (UID: \"6ff92fe7-35b8-43a0-9d62-cc825720ae76\") " pod="openstack/manila-db-sync-jbx7v" Nov 22 08:53:28 crc kubenswrapper[4789]: I1122 08:53:28.350718 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ff92fe7-35b8-43a0-9d62-cc825720ae76-combined-ca-bundle\") pod \"manila-db-sync-jbx7v\" (UID: \"6ff92fe7-35b8-43a0-9d62-cc825720ae76\") " pod="openstack/manila-db-sync-jbx7v" Nov 22 08:53:28 crc kubenswrapper[4789]: I1122 08:53:28.364403 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/6ff92fe7-35b8-43a0-9d62-cc825720ae76-job-config-data\") pod \"manila-db-sync-jbx7v\" (UID: \"6ff92fe7-35b8-43a0-9d62-cc825720ae76\") " pod="openstack/manila-db-sync-jbx7v" Nov 22 08:53:28 crc kubenswrapper[4789]: I1122 08:53:28.365742 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ff92fe7-35b8-43a0-9d62-cc825720ae76-config-data\") pod \"manila-db-sync-jbx7v\" (UID: \"6ff92fe7-35b8-43a0-9d62-cc825720ae76\") " pod="openstack/manila-db-sync-jbx7v" Nov 22 08:53:28 crc kubenswrapper[4789]: I1122 08:53:28.376557 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tp9wg\" (UniqueName: \"kubernetes.io/projected/6ff92fe7-35b8-43a0-9d62-cc825720ae76-kube-api-access-tp9wg\") pod \"manila-db-sync-jbx7v\" (UID: \"6ff92fe7-35b8-43a0-9d62-cc825720ae76\") " pod="openstack/manila-db-sync-jbx7v" Nov 22 08:53:28 crc kubenswrapper[4789]: I1122 08:53:28.409097 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-jbx7v" Nov 22 08:53:28 crc kubenswrapper[4789]: I1122 08:53:28.948730 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 22 08:53:28 crc kubenswrapper[4789]: I1122 08:53:28.949050 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 22 08:53:28 crc kubenswrapper[4789]: I1122 08:53:28.949063 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 22 08:53:28 crc kubenswrapper[4789]: I1122 08:53:28.949075 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 22 08:53:28 crc kubenswrapper[4789]: I1122 08:53:28.962295 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-sync-jbx7v"] Nov 22 08:53:28 crc kubenswrapper[4789]: W1122 08:53:28.966166 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6ff92fe7_35b8_43a0_9d62_cc825720ae76.slice/crio-c69327f7dcb82fe722234369e12d05710841200b38520974c490d8dfb8c3f99d WatchSource:0}: Error finding container c69327f7dcb82fe722234369e12d05710841200b38520974c490d8dfb8c3f99d: Status 404 returned error can't find the container with id c69327f7dcb82fe722234369e12d05710841200b38520974c490d8dfb8c3f99d Nov 22 08:53:29 crc kubenswrapper[4789]: I1122 08:53:29.966113 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-jbx7v" event={"ID":"6ff92fe7-35b8-43a0-9d62-cc825720ae76","Type":"ContainerStarted","Data":"c69327f7dcb82fe722234369e12d05710841200b38520974c490d8dfb8c3f99d"} Nov 22 08:53:31 crc kubenswrapper[4789]: I1122 08:53:31.075852 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 22 08:53:31 crc kubenswrapper[4789]: I1122 08:53:31.076246 4789 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 22 08:53:31 crc kubenswrapper[4789]: I1122 08:53:31.081839 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 22 08:53:31 crc kubenswrapper[4789]: I1122 08:53:31.088582 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 22 08:53:31 crc kubenswrapper[4789]: I1122 08:53:31.088720 4789 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 22 08:53:31 crc kubenswrapper[4789]: I1122 08:53:31.092159 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 22 08:53:34 crc kubenswrapper[4789]: I1122 08:53:34.004159 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-jbx7v" event={"ID":"6ff92fe7-35b8-43a0-9d62-cc825720ae76","Type":"ContainerStarted","Data":"b979cfbd66f08683cb938e63ddf2a4a2cc18432260fbd2a605f434b33ba54a0d"} Nov 22 08:53:34 crc kubenswrapper[4789]: I1122 08:53:34.025157 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-db-sync-jbx7v" podStartSLOduration=1.689324593 podStartE2EDuration="6.02513499s" podCreationTimestamp="2025-11-22 08:53:28 +0000 UTC" firstStartedPulling="2025-11-22 08:53:28.969117573 +0000 UTC m=+3423.203517846" lastFinishedPulling="2025-11-22 08:53:33.30492797 +0000 UTC m=+3427.539328243" observedRunningTime="2025-11-22 08:53:34.018394429 +0000 UTC m=+3428.252794702" watchObservedRunningTime="2025-11-22 08:53:34.02513499 +0000 UTC m=+3428.259535263" Nov 22 08:53:35 crc kubenswrapper[4789]: I1122 08:53:35.371967 4789 patch_prober.go:28] interesting pod/machine-config-daemon-pxvrm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 08:53:35 crc kubenswrapper[4789]: I1122 08:53:35.372307 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 08:53:52 crc kubenswrapper[4789]: I1122 08:53:52.161058 4789 generic.go:334] "Generic (PLEG): container finished" podID="6ff92fe7-35b8-43a0-9d62-cc825720ae76" containerID="b979cfbd66f08683cb938e63ddf2a4a2cc18432260fbd2a605f434b33ba54a0d" exitCode=0 Nov 22 08:53:52 crc kubenswrapper[4789]: I1122 08:53:52.161166 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-jbx7v" event={"ID":"6ff92fe7-35b8-43a0-9d62-cc825720ae76","Type":"ContainerDied","Data":"b979cfbd66f08683cb938e63ddf2a4a2cc18432260fbd2a605f434b33ba54a0d"} Nov 22 08:53:53 crc kubenswrapper[4789]: I1122 08:53:53.549217 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-jbx7v" Nov 22 08:53:53 crc kubenswrapper[4789]: I1122 08:53:53.669241 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ff92fe7-35b8-43a0-9d62-cc825720ae76-config-data\") pod \"6ff92fe7-35b8-43a0-9d62-cc825720ae76\" (UID: \"6ff92fe7-35b8-43a0-9d62-cc825720ae76\") " Nov 22 08:53:53 crc kubenswrapper[4789]: I1122 08:53:53.669431 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ff92fe7-35b8-43a0-9d62-cc825720ae76-combined-ca-bundle\") pod \"6ff92fe7-35b8-43a0-9d62-cc825720ae76\" (UID: \"6ff92fe7-35b8-43a0-9d62-cc825720ae76\") " Nov 22 08:53:53 crc kubenswrapper[4789]: I1122 08:53:53.669491 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tp9wg\" (UniqueName: \"kubernetes.io/projected/6ff92fe7-35b8-43a0-9d62-cc825720ae76-kube-api-access-tp9wg\") pod \"6ff92fe7-35b8-43a0-9d62-cc825720ae76\" (UID: \"6ff92fe7-35b8-43a0-9d62-cc825720ae76\") " Nov 22 08:53:53 crc kubenswrapper[4789]: I1122 08:53:53.669581 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/6ff92fe7-35b8-43a0-9d62-cc825720ae76-job-config-data\") pod \"6ff92fe7-35b8-43a0-9d62-cc825720ae76\" (UID: \"6ff92fe7-35b8-43a0-9d62-cc825720ae76\") " Nov 22 08:53:53 crc kubenswrapper[4789]: I1122 08:53:53.674979 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ff92fe7-35b8-43a0-9d62-cc825720ae76-job-config-data" (OuterVolumeSpecName: "job-config-data") pod "6ff92fe7-35b8-43a0-9d62-cc825720ae76" (UID: "6ff92fe7-35b8-43a0-9d62-cc825720ae76"). InnerVolumeSpecName "job-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:53:53 crc kubenswrapper[4789]: I1122 08:53:53.675084 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ff92fe7-35b8-43a0-9d62-cc825720ae76-kube-api-access-tp9wg" (OuterVolumeSpecName: "kube-api-access-tp9wg") pod "6ff92fe7-35b8-43a0-9d62-cc825720ae76" (UID: "6ff92fe7-35b8-43a0-9d62-cc825720ae76"). InnerVolumeSpecName "kube-api-access-tp9wg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:53:53 crc kubenswrapper[4789]: I1122 08:53:53.679058 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ff92fe7-35b8-43a0-9d62-cc825720ae76-config-data" (OuterVolumeSpecName: "config-data") pod "6ff92fe7-35b8-43a0-9d62-cc825720ae76" (UID: "6ff92fe7-35b8-43a0-9d62-cc825720ae76"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:53:53 crc kubenswrapper[4789]: I1122 08:53:53.698470 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ff92fe7-35b8-43a0-9d62-cc825720ae76-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6ff92fe7-35b8-43a0-9d62-cc825720ae76" (UID: "6ff92fe7-35b8-43a0-9d62-cc825720ae76"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:53:53 crc kubenswrapper[4789]: I1122 08:53:53.772027 4789 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ff92fe7-35b8-43a0-9d62-cc825720ae76-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:53:53 crc kubenswrapper[4789]: I1122 08:53:53.772362 4789 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ff92fe7-35b8-43a0-9d62-cc825720ae76-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:53:53 crc kubenswrapper[4789]: I1122 08:53:53.772454 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tp9wg\" (UniqueName: \"kubernetes.io/projected/6ff92fe7-35b8-43a0-9d62-cc825720ae76-kube-api-access-tp9wg\") on node \"crc\" DevicePath \"\"" Nov 22 08:53:53 crc kubenswrapper[4789]: I1122 08:53:53.772528 4789 reconciler_common.go:293] "Volume detached for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/6ff92fe7-35b8-43a0-9d62-cc825720ae76-job-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:53:54 crc kubenswrapper[4789]: I1122 08:53:54.181622 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-jbx7v" event={"ID":"6ff92fe7-35b8-43a0-9d62-cc825720ae76","Type":"ContainerDied","Data":"c69327f7dcb82fe722234369e12d05710841200b38520974c490d8dfb8c3f99d"} Nov 22 08:53:54 crc kubenswrapper[4789]: I1122 08:53:54.182374 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c69327f7dcb82fe722234369e12d05710841200b38520974c490d8dfb8c3f99d" Nov 22 08:53:54 crc kubenswrapper[4789]: I1122 08:53:54.181673 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-jbx7v" Nov 22 08:53:54 crc kubenswrapper[4789]: I1122 08:53:54.569159 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-share-share1-0"] Nov 22 08:53:54 crc kubenswrapper[4789]: E1122 08:53:54.569707 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ff92fe7-35b8-43a0-9d62-cc825720ae76" containerName="manila-db-sync" Nov 22 08:53:54 crc kubenswrapper[4789]: I1122 08:53:54.569728 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ff92fe7-35b8-43a0-9d62-cc825720ae76" containerName="manila-db-sync" Nov 22 08:53:54 crc kubenswrapper[4789]: I1122 08:53:54.569961 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ff92fe7-35b8-43a0-9d62-cc825720ae76" containerName="manila-db-sync" Nov 22 08:53:54 crc kubenswrapper[4789]: I1122 08:53:54.571434 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Nov 22 08:53:54 crc kubenswrapper[4789]: I1122 08:53:54.576042 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-config-data" Nov 22 08:53:54 crc kubenswrapper[4789]: I1122 08:53:54.576391 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scripts" Nov 22 08:53:54 crc kubenswrapper[4789]: I1122 08:53:54.576604 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-share-share1-config-data" Nov 22 08:53:54 crc kubenswrapper[4789]: I1122 08:53:54.576694 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-manila-dockercfg-fzw4x" Nov 22 08:53:54 crc kubenswrapper[4789]: I1122 08:53:54.586303 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-scheduler-0"] Nov 22 08:53:54 crc kubenswrapper[4789]: I1122 08:53:54.587987 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Nov 22 08:53:54 crc kubenswrapper[4789]: I1122 08:53:54.591297 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scheduler-config-data" Nov 22 08:53:54 crc kubenswrapper[4789]: I1122 08:53:54.610380 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Nov 22 08:53:54 crc kubenswrapper[4789]: I1122 08:53:54.632050 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Nov 22 08:53:54 crc kubenswrapper[4789]: I1122 08:53:54.687620 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6885d49d55-m8c6x"] Nov 22 08:53:54 crc kubenswrapper[4789]: I1122 08:53:54.689913 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6885d49d55-m8c6x" Nov 22 08:53:54 crc kubenswrapper[4789]: I1122 08:53:54.691395 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b679ee8-b6ac-4313-ac79-abd465d1b0c4-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"3b679ee8-b6ac-4313-ac79-abd465d1b0c4\") " pod="openstack/manila-share-share1-0" Nov 22 08:53:54 crc kubenswrapper[4789]: I1122 08:53:54.691464 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/3b679ee8-b6ac-4313-ac79-abd465d1b0c4-ceph\") pod \"manila-share-share1-0\" (UID: \"3b679ee8-b6ac-4313-ac79-abd465d1b0c4\") " pod="openstack/manila-share-share1-0" Nov 22 08:53:54 crc kubenswrapper[4789]: I1122 08:53:54.691489 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3b679ee8-b6ac-4313-ac79-abd465d1b0c4-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"3b679ee8-b6ac-4313-ac79-abd465d1b0c4\") " pod="openstack/manila-share-share1-0" Nov 22 08:53:54 crc kubenswrapper[4789]: I1122 08:53:54.691522 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h647z\" (UniqueName: \"kubernetes.io/projected/3b679ee8-b6ac-4313-ac79-abd465d1b0c4-kube-api-access-h647z\") pod \"manila-share-share1-0\" (UID: \"3b679ee8-b6ac-4313-ac79-abd465d1b0c4\") " pod="openstack/manila-share-share1-0" Nov 22 08:53:54 crc kubenswrapper[4789]: I1122 08:53:54.691561 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46af0a4f-fe66-47b9-b9d9-892c3eb3d97c-config-data\") pod \"manila-scheduler-0\" (UID: \"46af0a4f-fe66-47b9-b9d9-892c3eb3d97c\") " pod="openstack/manila-scheduler-0" Nov 22 08:53:54 crc kubenswrapper[4789]: I1122 08:53:54.691584 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r2mss\" (UniqueName: \"kubernetes.io/projected/46af0a4f-fe66-47b9-b9d9-892c3eb3d97c-kube-api-access-r2mss\") pod \"manila-scheduler-0\" (UID: \"46af0a4f-fe66-47b9-b9d9-892c3eb3d97c\") " pod="openstack/manila-scheduler-0" Nov 22 08:53:54 crc kubenswrapper[4789]: I1122 08:53:54.691618 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3b679ee8-b6ac-4313-ac79-abd465d1b0c4-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"3b679ee8-b6ac-4313-ac79-abd465d1b0c4\") " pod="openstack/manila-share-share1-0" Nov 22 08:53:54 crc kubenswrapper[4789]: I1122 08:53:54.691712 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b679ee8-b6ac-4313-ac79-abd465d1b0c4-config-data\") pod \"manila-share-share1-0\" (UID: \"3b679ee8-b6ac-4313-ac79-abd465d1b0c4\") " pod="openstack/manila-share-share1-0" Nov 22 08:53:54 crc kubenswrapper[4789]: I1122 08:53:54.691735 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/46af0a4f-fe66-47b9-b9d9-892c3eb3d97c-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"46af0a4f-fe66-47b9-b9d9-892c3eb3d97c\") " pod="openstack/manila-scheduler-0" Nov 22 08:53:54 crc kubenswrapper[4789]: I1122 08:53:54.691791 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46af0a4f-fe66-47b9-b9d9-892c3eb3d97c-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"46af0a4f-fe66-47b9-b9d9-892c3eb3d97c\") " pod="openstack/manila-scheduler-0" Nov 22 08:53:54 crc kubenswrapper[4789]: I1122 08:53:54.691817 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/46af0a4f-fe66-47b9-b9d9-892c3eb3d97c-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"46af0a4f-fe66-47b9-b9d9-892c3eb3d97c\") " pod="openstack/manila-scheduler-0" Nov 22 08:53:54 crc kubenswrapper[4789]: I1122 08:53:54.691847 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/46af0a4f-fe66-47b9-b9d9-892c3eb3d97c-scripts\") pod \"manila-scheduler-0\" (UID: \"46af0a4f-fe66-47b9-b9d9-892c3eb3d97c\") " pod="openstack/manila-scheduler-0" Nov 22 08:53:54 crc kubenswrapper[4789]: I1122 08:53:54.691894 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3b679ee8-b6ac-4313-ac79-abd465d1b0c4-scripts\") pod \"manila-share-share1-0\" (UID: \"3b679ee8-b6ac-4313-ac79-abd465d1b0c4\") " pod="openstack/manila-share-share1-0" Nov 22 08:53:54 crc kubenswrapper[4789]: I1122 08:53:54.691919 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/3b679ee8-b6ac-4313-ac79-abd465d1b0c4-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"3b679ee8-b6ac-4313-ac79-abd465d1b0c4\") " pod="openstack/manila-share-share1-0" Nov 22 08:53:54 crc kubenswrapper[4789]: I1122 08:53:54.714014 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6885d49d55-m8c6x"] Nov 22 08:53:54 crc kubenswrapper[4789]: I1122 08:53:54.793442 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/3b679ee8-b6ac-4313-ac79-abd465d1b0c4-ceph\") pod \"manila-share-share1-0\" (UID: \"3b679ee8-b6ac-4313-ac79-abd465d1b0c4\") " pod="openstack/manila-share-share1-0" Nov 22 08:53:54 crc kubenswrapper[4789]: I1122 08:53:54.793507 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3b679ee8-b6ac-4313-ac79-abd465d1b0c4-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"3b679ee8-b6ac-4313-ac79-abd465d1b0c4\") " pod="openstack/manila-share-share1-0" Nov 22 08:53:54 crc kubenswrapper[4789]: I1122 08:53:54.793548 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h647z\" (UniqueName: \"kubernetes.io/projected/3b679ee8-b6ac-4313-ac79-abd465d1b0c4-kube-api-access-h647z\") pod \"manila-share-share1-0\" (UID: \"3b679ee8-b6ac-4313-ac79-abd465d1b0c4\") " pod="openstack/manila-share-share1-0" Nov 22 08:53:54 crc kubenswrapper[4789]: I1122 08:53:54.793588 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46af0a4f-fe66-47b9-b9d9-892c3eb3d97c-config-data\") pod \"manila-scheduler-0\" (UID: \"46af0a4f-fe66-47b9-b9d9-892c3eb3d97c\") " pod="openstack/manila-scheduler-0" Nov 22 08:53:54 crc kubenswrapper[4789]: I1122 08:53:54.793618 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r2mss\" (UniqueName: \"kubernetes.io/projected/46af0a4f-fe66-47b9-b9d9-892c3eb3d97c-kube-api-access-r2mss\") pod \"manila-scheduler-0\" (UID: \"46af0a4f-fe66-47b9-b9d9-892c3eb3d97c\") " pod="openstack/manila-scheduler-0" Nov 22 08:53:54 crc kubenswrapper[4789]: I1122 08:53:54.793660 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3b679ee8-b6ac-4313-ac79-abd465d1b0c4-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"3b679ee8-b6ac-4313-ac79-abd465d1b0c4\") " pod="openstack/manila-share-share1-0" Nov 22 08:53:54 crc kubenswrapper[4789]: I1122 08:53:54.793705 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5e1d2ec8-1172-4bcc-a5e7-5cbbea574845-config\") pod \"dnsmasq-dns-6885d49d55-m8c6x\" (UID: \"5e1d2ec8-1172-4bcc-a5e7-5cbbea574845\") " pod="openstack/dnsmasq-dns-6885d49d55-m8c6x" Nov 22 08:53:54 crc kubenswrapper[4789]: I1122 08:53:54.793725 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m2lvp\" (UniqueName: \"kubernetes.io/projected/5e1d2ec8-1172-4bcc-a5e7-5cbbea574845-kube-api-access-m2lvp\") pod \"dnsmasq-dns-6885d49d55-m8c6x\" (UID: \"5e1d2ec8-1172-4bcc-a5e7-5cbbea574845\") " pod="openstack/dnsmasq-dns-6885d49d55-m8c6x" Nov 22 08:53:54 crc kubenswrapper[4789]: I1122 08:53:54.793801 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5e1d2ec8-1172-4bcc-a5e7-5cbbea574845-dns-svc\") pod \"dnsmasq-dns-6885d49d55-m8c6x\" (UID: \"5e1d2ec8-1172-4bcc-a5e7-5cbbea574845\") " pod="openstack/dnsmasq-dns-6885d49d55-m8c6x" Nov 22 08:53:54 crc kubenswrapper[4789]: I1122 08:53:54.793819 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5e1d2ec8-1172-4bcc-a5e7-5cbbea574845-ovsdbserver-sb\") pod \"dnsmasq-dns-6885d49d55-m8c6x\" (UID: \"5e1d2ec8-1172-4bcc-a5e7-5cbbea574845\") " pod="openstack/dnsmasq-dns-6885d49d55-m8c6x" Nov 22 08:53:54 crc kubenswrapper[4789]: I1122 08:53:54.793841 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b679ee8-b6ac-4313-ac79-abd465d1b0c4-config-data\") pod \"manila-share-share1-0\" (UID: \"3b679ee8-b6ac-4313-ac79-abd465d1b0c4\") " pod="openstack/manila-share-share1-0" Nov 22 08:53:54 crc kubenswrapper[4789]: I1122 08:53:54.793857 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/46af0a4f-fe66-47b9-b9d9-892c3eb3d97c-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"46af0a4f-fe66-47b9-b9d9-892c3eb3d97c\") " pod="openstack/manila-scheduler-0" Nov 22 08:53:54 crc kubenswrapper[4789]: I1122 08:53:54.793879 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5e1d2ec8-1172-4bcc-a5e7-5cbbea574845-ovsdbserver-nb\") pod \"dnsmasq-dns-6885d49d55-m8c6x\" (UID: \"5e1d2ec8-1172-4bcc-a5e7-5cbbea574845\") " pod="openstack/dnsmasq-dns-6885d49d55-m8c6x" Nov 22 08:53:54 crc kubenswrapper[4789]: I1122 08:53:54.794401 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3b679ee8-b6ac-4313-ac79-abd465d1b0c4-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"3b679ee8-b6ac-4313-ac79-abd465d1b0c4\") " pod="openstack/manila-share-share1-0" Nov 22 08:53:54 crc kubenswrapper[4789]: I1122 08:53:54.794541 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46af0a4f-fe66-47b9-b9d9-892c3eb3d97c-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"46af0a4f-fe66-47b9-b9d9-892c3eb3d97c\") " pod="openstack/manila-scheduler-0" Nov 22 08:53:54 crc kubenswrapper[4789]: I1122 08:53:54.794590 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/46af0a4f-fe66-47b9-b9d9-892c3eb3d97c-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"46af0a4f-fe66-47b9-b9d9-892c3eb3d97c\") " pod="openstack/manila-scheduler-0" Nov 22 08:53:54 crc kubenswrapper[4789]: I1122 08:53:54.794645 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/46af0a4f-fe66-47b9-b9d9-892c3eb3d97c-scripts\") pod \"manila-scheduler-0\" (UID: \"46af0a4f-fe66-47b9-b9d9-892c3eb3d97c\") " pod="openstack/manila-scheduler-0" Nov 22 08:53:54 crc kubenswrapper[4789]: I1122 08:53:54.794699 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/5e1d2ec8-1172-4bcc-a5e7-5cbbea574845-openstack-edpm-ipam\") pod \"dnsmasq-dns-6885d49d55-m8c6x\" (UID: \"5e1d2ec8-1172-4bcc-a5e7-5cbbea574845\") " pod="openstack/dnsmasq-dns-6885d49d55-m8c6x" Nov 22 08:53:54 crc kubenswrapper[4789]: I1122 08:53:54.794818 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3b679ee8-b6ac-4313-ac79-abd465d1b0c4-scripts\") pod \"manila-share-share1-0\" (UID: \"3b679ee8-b6ac-4313-ac79-abd465d1b0c4\") " pod="openstack/manila-share-share1-0" Nov 22 08:53:54 crc kubenswrapper[4789]: I1122 08:53:54.794866 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/3b679ee8-b6ac-4313-ac79-abd465d1b0c4-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"3b679ee8-b6ac-4313-ac79-abd465d1b0c4\") " pod="openstack/manila-share-share1-0" Nov 22 08:53:54 crc kubenswrapper[4789]: I1122 08:53:54.794956 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b679ee8-b6ac-4313-ac79-abd465d1b0c4-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"3b679ee8-b6ac-4313-ac79-abd465d1b0c4\") " pod="openstack/manila-share-share1-0" Nov 22 08:53:54 crc kubenswrapper[4789]: I1122 08:53:54.796621 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/46af0a4f-fe66-47b9-b9d9-892c3eb3d97c-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"46af0a4f-fe66-47b9-b9d9-892c3eb3d97c\") " pod="openstack/manila-scheduler-0" Nov 22 08:53:54 crc kubenswrapper[4789]: I1122 08:53:54.796795 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/3b679ee8-b6ac-4313-ac79-abd465d1b0c4-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"3b679ee8-b6ac-4313-ac79-abd465d1b0c4\") " pod="openstack/manila-share-share1-0" Nov 22 08:53:54 crc kubenswrapper[4789]: I1122 08:53:54.802811 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3b679ee8-b6ac-4313-ac79-abd465d1b0c4-scripts\") pod \"manila-share-share1-0\" (UID: \"3b679ee8-b6ac-4313-ac79-abd465d1b0c4\") " pod="openstack/manila-share-share1-0" Nov 22 08:53:54 crc kubenswrapper[4789]: I1122 08:53:54.802902 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3b679ee8-b6ac-4313-ac79-abd465d1b0c4-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"3b679ee8-b6ac-4313-ac79-abd465d1b0c4\") " pod="openstack/manila-share-share1-0" Nov 22 08:53:54 crc kubenswrapper[4789]: I1122 08:53:54.803262 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/46af0a4f-fe66-47b9-b9d9-892c3eb3d97c-scripts\") pod \"manila-scheduler-0\" (UID: \"46af0a4f-fe66-47b9-b9d9-892c3eb3d97c\") " pod="openstack/manila-scheduler-0" Nov 22 08:53:54 crc kubenswrapper[4789]: I1122 08:53:54.803388 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46af0a4f-fe66-47b9-b9d9-892c3eb3d97c-config-data\") pod \"manila-scheduler-0\" (UID: \"46af0a4f-fe66-47b9-b9d9-892c3eb3d97c\") " pod="openstack/manila-scheduler-0" Nov 22 08:53:54 crc kubenswrapper[4789]: I1122 08:53:54.803537 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b679ee8-b6ac-4313-ac79-abd465d1b0c4-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"3b679ee8-b6ac-4313-ac79-abd465d1b0c4\") " pod="openstack/manila-share-share1-0" Nov 22 08:53:54 crc kubenswrapper[4789]: I1122 08:53:54.808263 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/46af0a4f-fe66-47b9-b9d9-892c3eb3d97c-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"46af0a4f-fe66-47b9-b9d9-892c3eb3d97c\") " pod="openstack/manila-scheduler-0" Nov 22 08:53:54 crc kubenswrapper[4789]: I1122 08:53:54.810294 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b679ee8-b6ac-4313-ac79-abd465d1b0c4-config-data\") pod \"manila-share-share1-0\" (UID: \"3b679ee8-b6ac-4313-ac79-abd465d1b0c4\") " pod="openstack/manila-share-share1-0" Nov 22 08:53:54 crc kubenswrapper[4789]: I1122 08:53:54.835385 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46af0a4f-fe66-47b9-b9d9-892c3eb3d97c-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"46af0a4f-fe66-47b9-b9d9-892c3eb3d97c\") " pod="openstack/manila-scheduler-0" Nov 22 08:53:54 crc kubenswrapper[4789]: I1122 08:53:54.835494 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/3b679ee8-b6ac-4313-ac79-abd465d1b0c4-ceph\") pod \"manila-share-share1-0\" (UID: \"3b679ee8-b6ac-4313-ac79-abd465d1b0c4\") " pod="openstack/manila-share-share1-0" Nov 22 08:53:54 crc kubenswrapper[4789]: I1122 08:53:54.835539 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r2mss\" (UniqueName: \"kubernetes.io/projected/46af0a4f-fe66-47b9-b9d9-892c3eb3d97c-kube-api-access-r2mss\") pod \"manila-scheduler-0\" (UID: \"46af0a4f-fe66-47b9-b9d9-892c3eb3d97c\") " pod="openstack/manila-scheduler-0" Nov 22 08:53:54 crc kubenswrapper[4789]: I1122 08:53:54.839032 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h647z\" (UniqueName: \"kubernetes.io/projected/3b679ee8-b6ac-4313-ac79-abd465d1b0c4-kube-api-access-h647z\") pod \"manila-share-share1-0\" (UID: \"3b679ee8-b6ac-4313-ac79-abd465d1b0c4\") " pod="openstack/manila-share-share1-0" Nov 22 08:53:54 crc kubenswrapper[4789]: I1122 08:53:54.892546 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-api-0"] Nov 22 08:53:54 crc kubenswrapper[4789]: I1122 08:53:54.893963 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Nov 22 08:53:54 crc kubenswrapper[4789]: I1122 08:53:54.894159 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Nov 22 08:53:54 crc kubenswrapper[4789]: I1122 08:53:54.897990 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5e1d2ec8-1172-4bcc-a5e7-5cbbea574845-config\") pod \"dnsmasq-dns-6885d49d55-m8c6x\" (UID: \"5e1d2ec8-1172-4bcc-a5e7-5cbbea574845\") " pod="openstack/dnsmasq-dns-6885d49d55-m8c6x" Nov 22 08:53:54 crc kubenswrapper[4789]: I1122 08:53:54.898050 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m2lvp\" (UniqueName: \"kubernetes.io/projected/5e1d2ec8-1172-4bcc-a5e7-5cbbea574845-kube-api-access-m2lvp\") pod \"dnsmasq-dns-6885d49d55-m8c6x\" (UID: \"5e1d2ec8-1172-4bcc-a5e7-5cbbea574845\") " pod="openstack/dnsmasq-dns-6885d49d55-m8c6x" Nov 22 08:53:54 crc kubenswrapper[4789]: I1122 08:53:54.898104 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5e1d2ec8-1172-4bcc-a5e7-5cbbea574845-dns-svc\") pod \"dnsmasq-dns-6885d49d55-m8c6x\" (UID: \"5e1d2ec8-1172-4bcc-a5e7-5cbbea574845\") " pod="openstack/dnsmasq-dns-6885d49d55-m8c6x" Nov 22 08:53:54 crc kubenswrapper[4789]: I1122 08:53:54.898130 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5e1d2ec8-1172-4bcc-a5e7-5cbbea574845-ovsdbserver-sb\") pod \"dnsmasq-dns-6885d49d55-m8c6x\" (UID: \"5e1d2ec8-1172-4bcc-a5e7-5cbbea574845\") " pod="openstack/dnsmasq-dns-6885d49d55-m8c6x" Nov 22 08:53:54 crc kubenswrapper[4789]: I1122 08:53:54.898159 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5e1d2ec8-1172-4bcc-a5e7-5cbbea574845-ovsdbserver-nb\") pod \"dnsmasq-dns-6885d49d55-m8c6x\" (UID: \"5e1d2ec8-1172-4bcc-a5e7-5cbbea574845\") " pod="openstack/dnsmasq-dns-6885d49d55-m8c6x" Nov 22 08:53:54 crc kubenswrapper[4789]: I1122 08:53:54.898206 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/5e1d2ec8-1172-4bcc-a5e7-5cbbea574845-openstack-edpm-ipam\") pod \"dnsmasq-dns-6885d49d55-m8c6x\" (UID: \"5e1d2ec8-1172-4bcc-a5e7-5cbbea574845\") " pod="openstack/dnsmasq-dns-6885d49d55-m8c6x" Nov 22 08:53:54 crc kubenswrapper[4789]: I1122 08:53:54.899363 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/5e1d2ec8-1172-4bcc-a5e7-5cbbea574845-openstack-edpm-ipam\") pod \"dnsmasq-dns-6885d49d55-m8c6x\" (UID: \"5e1d2ec8-1172-4bcc-a5e7-5cbbea574845\") " pod="openstack/dnsmasq-dns-6885d49d55-m8c6x" Nov 22 08:53:54 crc kubenswrapper[4789]: I1122 08:53:54.900057 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5e1d2ec8-1172-4bcc-a5e7-5cbbea574845-config\") pod \"dnsmasq-dns-6885d49d55-m8c6x\" (UID: \"5e1d2ec8-1172-4bcc-a5e7-5cbbea574845\") " pod="openstack/dnsmasq-dns-6885d49d55-m8c6x" Nov 22 08:53:54 crc kubenswrapper[4789]: I1122 08:53:54.904214 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5e1d2ec8-1172-4bcc-a5e7-5cbbea574845-dns-svc\") pod \"dnsmasq-dns-6885d49d55-m8c6x\" (UID: \"5e1d2ec8-1172-4bcc-a5e7-5cbbea574845\") " pod="openstack/dnsmasq-dns-6885d49d55-m8c6x" Nov 22 08:53:54 crc kubenswrapper[4789]: I1122 08:53:54.905112 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-api-config-data" Nov 22 08:53:54 crc kubenswrapper[4789]: I1122 08:53:54.905325 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5e1d2ec8-1172-4bcc-a5e7-5cbbea574845-ovsdbserver-sb\") pod \"dnsmasq-dns-6885d49d55-m8c6x\" (UID: \"5e1d2ec8-1172-4bcc-a5e7-5cbbea574845\") " pod="openstack/dnsmasq-dns-6885d49d55-m8c6x" Nov 22 08:53:54 crc kubenswrapper[4789]: I1122 08:53:54.905649 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5e1d2ec8-1172-4bcc-a5e7-5cbbea574845-ovsdbserver-nb\") pod \"dnsmasq-dns-6885d49d55-m8c6x\" (UID: \"5e1d2ec8-1172-4bcc-a5e7-5cbbea574845\") " pod="openstack/dnsmasq-dns-6885d49d55-m8c6x" Nov 22 08:53:54 crc kubenswrapper[4789]: I1122 08:53:54.908032 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Nov 22 08:53:54 crc kubenswrapper[4789]: I1122 08:53:54.929817 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m2lvp\" (UniqueName: \"kubernetes.io/projected/5e1d2ec8-1172-4bcc-a5e7-5cbbea574845-kube-api-access-m2lvp\") pod \"dnsmasq-dns-6885d49d55-m8c6x\" (UID: \"5e1d2ec8-1172-4bcc-a5e7-5cbbea574845\") " pod="openstack/dnsmasq-dns-6885d49d55-m8c6x" Nov 22 08:53:54 crc kubenswrapper[4789]: I1122 08:53:54.934072 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Nov 22 08:53:55 crc kubenswrapper[4789]: I1122 08:53:55.001033 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9f40e5f0-7b57-47c4-8f31-bc041fb07b3c-config-data-custom\") pod \"manila-api-0\" (UID: \"9f40e5f0-7b57-47c4-8f31-bc041fb07b3c\") " pod="openstack/manila-api-0" Nov 22 08:53:55 crc kubenswrapper[4789]: I1122 08:53:55.001129 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9f40e5f0-7b57-47c4-8f31-bc041fb07b3c-scripts\") pod \"manila-api-0\" (UID: \"9f40e5f0-7b57-47c4-8f31-bc041fb07b3c\") " pod="openstack/manila-api-0" Nov 22 08:53:55 crc kubenswrapper[4789]: I1122 08:53:55.001166 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f40e5f0-7b57-47c4-8f31-bc041fb07b3c-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"9f40e5f0-7b57-47c4-8f31-bc041fb07b3c\") " pod="openstack/manila-api-0" Nov 22 08:53:55 crc kubenswrapper[4789]: I1122 08:53:55.001184 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9f40e5f0-7b57-47c4-8f31-bc041fb07b3c-logs\") pod \"manila-api-0\" (UID: \"9f40e5f0-7b57-47c4-8f31-bc041fb07b3c\") " pod="openstack/manila-api-0" Nov 22 08:53:55 crc kubenswrapper[4789]: I1122 08:53:55.001216 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pdnbw\" (UniqueName: \"kubernetes.io/projected/9f40e5f0-7b57-47c4-8f31-bc041fb07b3c-kube-api-access-pdnbw\") pod \"manila-api-0\" (UID: \"9f40e5f0-7b57-47c4-8f31-bc041fb07b3c\") " pod="openstack/manila-api-0" Nov 22 08:53:55 crc kubenswrapper[4789]: I1122 08:53:55.001245 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9f40e5f0-7b57-47c4-8f31-bc041fb07b3c-etc-machine-id\") pod \"manila-api-0\" (UID: \"9f40e5f0-7b57-47c4-8f31-bc041fb07b3c\") " pod="openstack/manila-api-0" Nov 22 08:53:55 crc kubenswrapper[4789]: I1122 08:53:55.001383 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f40e5f0-7b57-47c4-8f31-bc041fb07b3c-config-data\") pod \"manila-api-0\" (UID: \"9f40e5f0-7b57-47c4-8f31-bc041fb07b3c\") " pod="openstack/manila-api-0" Nov 22 08:53:55 crc kubenswrapper[4789]: I1122 08:53:55.020190 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6885d49d55-m8c6x" Nov 22 08:53:55 crc kubenswrapper[4789]: I1122 08:53:55.105213 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9f40e5f0-7b57-47c4-8f31-bc041fb07b3c-config-data-custom\") pod \"manila-api-0\" (UID: \"9f40e5f0-7b57-47c4-8f31-bc041fb07b3c\") " pod="openstack/manila-api-0" Nov 22 08:53:55 crc kubenswrapper[4789]: I1122 08:53:55.105349 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9f40e5f0-7b57-47c4-8f31-bc041fb07b3c-scripts\") pod \"manila-api-0\" (UID: \"9f40e5f0-7b57-47c4-8f31-bc041fb07b3c\") " pod="openstack/manila-api-0" Nov 22 08:53:55 crc kubenswrapper[4789]: I1122 08:53:55.105397 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f40e5f0-7b57-47c4-8f31-bc041fb07b3c-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"9f40e5f0-7b57-47c4-8f31-bc041fb07b3c\") " pod="openstack/manila-api-0" Nov 22 08:53:55 crc kubenswrapper[4789]: I1122 08:53:55.105472 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9f40e5f0-7b57-47c4-8f31-bc041fb07b3c-logs\") pod \"manila-api-0\" (UID: \"9f40e5f0-7b57-47c4-8f31-bc041fb07b3c\") " pod="openstack/manila-api-0" Nov 22 08:53:55 crc kubenswrapper[4789]: I1122 08:53:55.105534 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pdnbw\" (UniqueName: \"kubernetes.io/projected/9f40e5f0-7b57-47c4-8f31-bc041fb07b3c-kube-api-access-pdnbw\") pod \"manila-api-0\" (UID: \"9f40e5f0-7b57-47c4-8f31-bc041fb07b3c\") " pod="openstack/manila-api-0" Nov 22 08:53:55 crc kubenswrapper[4789]: I1122 08:53:55.105588 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9f40e5f0-7b57-47c4-8f31-bc041fb07b3c-etc-machine-id\") pod \"manila-api-0\" (UID: \"9f40e5f0-7b57-47c4-8f31-bc041fb07b3c\") " pod="openstack/manila-api-0" Nov 22 08:53:55 crc kubenswrapper[4789]: I1122 08:53:55.105616 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f40e5f0-7b57-47c4-8f31-bc041fb07b3c-config-data\") pod \"manila-api-0\" (UID: \"9f40e5f0-7b57-47c4-8f31-bc041fb07b3c\") " pod="openstack/manila-api-0" Nov 22 08:53:55 crc kubenswrapper[4789]: I1122 08:53:55.106836 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9f40e5f0-7b57-47c4-8f31-bc041fb07b3c-logs\") pod \"manila-api-0\" (UID: \"9f40e5f0-7b57-47c4-8f31-bc041fb07b3c\") " pod="openstack/manila-api-0" Nov 22 08:53:55 crc kubenswrapper[4789]: I1122 08:53:55.107971 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9f40e5f0-7b57-47c4-8f31-bc041fb07b3c-etc-machine-id\") pod \"manila-api-0\" (UID: \"9f40e5f0-7b57-47c4-8f31-bc041fb07b3c\") " pod="openstack/manila-api-0" Nov 22 08:53:55 crc kubenswrapper[4789]: I1122 08:53:55.116297 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9f40e5f0-7b57-47c4-8f31-bc041fb07b3c-config-data-custom\") pod \"manila-api-0\" (UID: \"9f40e5f0-7b57-47c4-8f31-bc041fb07b3c\") " pod="openstack/manila-api-0" Nov 22 08:53:55 crc kubenswrapper[4789]: I1122 08:53:55.119127 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f40e5f0-7b57-47c4-8f31-bc041fb07b3c-config-data\") pod \"manila-api-0\" (UID: \"9f40e5f0-7b57-47c4-8f31-bc041fb07b3c\") " pod="openstack/manila-api-0" Nov 22 08:53:55 crc kubenswrapper[4789]: I1122 08:53:55.121230 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f40e5f0-7b57-47c4-8f31-bc041fb07b3c-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"9f40e5f0-7b57-47c4-8f31-bc041fb07b3c\") " pod="openstack/manila-api-0" Nov 22 08:53:55 crc kubenswrapper[4789]: I1122 08:53:55.121333 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9f40e5f0-7b57-47c4-8f31-bc041fb07b3c-scripts\") pod \"manila-api-0\" (UID: \"9f40e5f0-7b57-47c4-8f31-bc041fb07b3c\") " pod="openstack/manila-api-0" Nov 22 08:53:55 crc kubenswrapper[4789]: I1122 08:53:55.129264 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pdnbw\" (UniqueName: \"kubernetes.io/projected/9f40e5f0-7b57-47c4-8f31-bc041fb07b3c-kube-api-access-pdnbw\") pod \"manila-api-0\" (UID: \"9f40e5f0-7b57-47c4-8f31-bc041fb07b3c\") " pod="openstack/manila-api-0" Nov 22 08:53:57 crc kubenswrapper[4789]: I1122 08:53:55.377212 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Nov 22 08:53:57 crc kubenswrapper[4789]: I1122 08:53:55.541696 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Nov 22 08:53:57 crc kubenswrapper[4789]: I1122 08:53:56.209984 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"3b679ee8-b6ac-4313-ac79-abd465d1b0c4","Type":"ContainerStarted","Data":"f41e7690841252fa485a3354fe7c66876a448965126fb4c0dadf03376c875ff6"} Nov 22 08:53:57 crc kubenswrapper[4789]: I1122 08:53:57.545369 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-api-0"] Nov 22 08:53:57 crc kubenswrapper[4789]: I1122 08:53:57.989101 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Nov 22 08:53:58 crc kubenswrapper[4789]: I1122 08:53:58.104342 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6885d49d55-m8c6x"] Nov 22 08:53:58 crc kubenswrapper[4789]: I1122 08:53:58.211642 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-api-0"] Nov 22 08:53:58 crc kubenswrapper[4789]: I1122 08:53:58.242353 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"9f40e5f0-7b57-47c4-8f31-bc041fb07b3c","Type":"ContainerStarted","Data":"2fe01025cd09f339780f5cbb351b5c9cc91ca4fe2cd75fead928dfa1c4e5538f"} Nov 22 08:53:58 crc kubenswrapper[4789]: I1122 08:53:58.247634 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"46af0a4f-fe66-47b9-b9d9-892c3eb3d97c","Type":"ContainerStarted","Data":"4a76c66c85baf528e18015f5fcf7ecc3cdf42023660ab95698cc768f618784ad"} Nov 22 08:53:58 crc kubenswrapper[4789]: I1122 08:53:58.249314 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6885d49d55-m8c6x" event={"ID":"5e1d2ec8-1172-4bcc-a5e7-5cbbea574845","Type":"ContainerStarted","Data":"b3db3920c71078b39ddd7461df8c5eabb226ccd934a16a55457aefda33ec9a7b"} Nov 22 08:53:59 crc kubenswrapper[4789]: I1122 08:53:59.260615 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"9f40e5f0-7b57-47c4-8f31-bc041fb07b3c","Type":"ContainerStarted","Data":"3844f5bf126c60293b5597ae121bee3569ae3b40cee283783c61b0e1ca44917a"} Nov 22 08:53:59 crc kubenswrapper[4789]: I1122 08:53:59.263449 4789 generic.go:334] "Generic (PLEG): container finished" podID="5e1d2ec8-1172-4bcc-a5e7-5cbbea574845" containerID="4e266a0fc10de0cc9436fc0b3d5e3f3a23fed7a16e0380e5853178f118be91c7" exitCode=0 Nov 22 08:53:59 crc kubenswrapper[4789]: I1122 08:53:59.263485 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6885d49d55-m8c6x" event={"ID":"5e1d2ec8-1172-4bcc-a5e7-5cbbea574845","Type":"ContainerDied","Data":"4e266a0fc10de0cc9436fc0b3d5e3f3a23fed7a16e0380e5853178f118be91c7"} Nov 22 08:53:59 crc kubenswrapper[4789]: I1122 08:53:59.335224 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:53:59 crc kubenswrapper[4789]: I1122 08:53:59.335666 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b3e0f7a4-cde7-4da6-8339-615054807f46" containerName="ceilometer-notification-agent" containerID="cri-o://0ba876bc146a633673594ffc7a55b1ef15f6d31b6f68e177c9d9b989cfe96170" gracePeriod=30 Nov 22 08:53:59 crc kubenswrapper[4789]: I1122 08:53:59.335897 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b3e0f7a4-cde7-4da6-8339-615054807f46" containerName="ceilometer-central-agent" containerID="cri-o://57bc1e114749de45fd56a9c528a09c54ebfb4ff2b2cda822e7b1411b31acac80" gracePeriod=30 Nov 22 08:53:59 crc kubenswrapper[4789]: I1122 08:53:59.335993 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b3e0f7a4-cde7-4da6-8339-615054807f46" containerName="proxy-httpd" containerID="cri-o://2fb144347d1d1aad510a6ea7031cee3e906598e8fd06b81b6777431331eafb09" gracePeriod=30 Nov 22 08:53:59 crc kubenswrapper[4789]: I1122 08:53:59.336077 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b3e0f7a4-cde7-4da6-8339-615054807f46" containerName="sg-core" containerID="cri-o://e7e7196b9d8de4d32b0220fbdd5f43d6bd1d61ff63e6f25250d131b43d76827d" gracePeriod=30 Nov 22 08:54:00 crc kubenswrapper[4789]: I1122 08:54:00.273647 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"9f40e5f0-7b57-47c4-8f31-bc041fb07b3c","Type":"ContainerStarted","Data":"36db728ef80e6e234991810c3f372fbf4cf4b3b5106eb7812562616dd262075f"} Nov 22 08:54:00 crc kubenswrapper[4789]: I1122 08:54:00.273895 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-api-0" podUID="9f40e5f0-7b57-47c4-8f31-bc041fb07b3c" containerName="manila-api-log" containerID="cri-o://3844f5bf126c60293b5597ae121bee3569ae3b40cee283783c61b0e1ca44917a" gracePeriod=30 Nov 22 08:54:00 crc kubenswrapper[4789]: I1122 08:54:00.273911 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-api-0" podUID="9f40e5f0-7b57-47c4-8f31-bc041fb07b3c" containerName="manila-api" containerID="cri-o://36db728ef80e6e234991810c3f372fbf4cf4b3b5106eb7812562616dd262075f" gracePeriod=30 Nov 22 08:54:00 crc kubenswrapper[4789]: I1122 08:54:00.277978 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/manila-api-0" Nov 22 08:54:00 crc kubenswrapper[4789]: I1122 08:54:00.278017 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6885d49d55-m8c6x" event={"ID":"5e1d2ec8-1172-4bcc-a5e7-5cbbea574845","Type":"ContainerStarted","Data":"52701131b28b6af95e09dc5735d81cb0eca8b1d7e8da128155cee1e41f957f14"} Nov 22 08:54:00 crc kubenswrapper[4789]: I1122 08:54:00.278802 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6885d49d55-m8c6x" Nov 22 08:54:00 crc kubenswrapper[4789]: I1122 08:54:00.286251 4789 generic.go:334] "Generic (PLEG): container finished" podID="b3e0f7a4-cde7-4da6-8339-615054807f46" containerID="57bc1e114749de45fd56a9c528a09c54ebfb4ff2b2cda822e7b1411b31acac80" exitCode=0 Nov 22 08:54:00 crc kubenswrapper[4789]: I1122 08:54:00.286286 4789 generic.go:334] "Generic (PLEG): container finished" podID="b3e0f7a4-cde7-4da6-8339-615054807f46" containerID="2fb144347d1d1aad510a6ea7031cee3e906598e8fd06b81b6777431331eafb09" exitCode=0 Nov 22 08:54:00 crc kubenswrapper[4789]: I1122 08:54:00.286298 4789 generic.go:334] "Generic (PLEG): container finished" podID="b3e0f7a4-cde7-4da6-8339-615054807f46" containerID="e7e7196b9d8de4d32b0220fbdd5f43d6bd1d61ff63e6f25250d131b43d76827d" exitCode=2 Nov 22 08:54:00 crc kubenswrapper[4789]: I1122 08:54:00.286323 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b3e0f7a4-cde7-4da6-8339-615054807f46","Type":"ContainerDied","Data":"57bc1e114749de45fd56a9c528a09c54ebfb4ff2b2cda822e7b1411b31acac80"} Nov 22 08:54:00 crc kubenswrapper[4789]: I1122 08:54:00.286355 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b3e0f7a4-cde7-4da6-8339-615054807f46","Type":"ContainerDied","Data":"2fb144347d1d1aad510a6ea7031cee3e906598e8fd06b81b6777431331eafb09"} Nov 22 08:54:00 crc kubenswrapper[4789]: I1122 08:54:00.286370 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b3e0f7a4-cde7-4da6-8339-615054807f46","Type":"ContainerDied","Data":"e7e7196b9d8de4d32b0220fbdd5f43d6bd1d61ff63e6f25250d131b43d76827d"} Nov 22 08:54:00 crc kubenswrapper[4789]: I1122 08:54:00.286386 4789 scope.go:117] "RemoveContainer" containerID="a3569e8f73bbb94022be853ca4eca0f31e32f0be828470dce1bedafa7097f19c" Nov 22 08:54:00 crc kubenswrapper[4789]: I1122 08:54:00.308925 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-api-0" podStartSLOduration=6.308902259 podStartE2EDuration="6.308902259s" podCreationTimestamp="2025-11-22 08:53:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:54:00.299091585 +0000 UTC m=+3454.533491858" watchObservedRunningTime="2025-11-22 08:54:00.308902259 +0000 UTC m=+3454.543302532" Nov 22 08:54:00 crc kubenswrapper[4789]: I1122 08:54:00.341594 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6885d49d55-m8c6x" podStartSLOduration=6.341561178 podStartE2EDuration="6.341561178s" podCreationTimestamp="2025-11-22 08:53:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:54:00.323377509 +0000 UTC m=+3454.557777792" watchObservedRunningTime="2025-11-22 08:54:00.341561178 +0000 UTC m=+3454.575961481" Nov 22 08:54:01 crc kubenswrapper[4789]: I1122 08:54:01.299245 4789 generic.go:334] "Generic (PLEG): container finished" podID="9f40e5f0-7b57-47c4-8f31-bc041fb07b3c" containerID="36db728ef80e6e234991810c3f372fbf4cf4b3b5106eb7812562616dd262075f" exitCode=0 Nov 22 08:54:01 crc kubenswrapper[4789]: I1122 08:54:01.299548 4789 generic.go:334] "Generic (PLEG): container finished" podID="9f40e5f0-7b57-47c4-8f31-bc041fb07b3c" containerID="3844f5bf126c60293b5597ae121bee3569ae3b40cee283783c61b0e1ca44917a" exitCode=143 Nov 22 08:54:01 crc kubenswrapper[4789]: I1122 08:54:01.299290 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"9f40e5f0-7b57-47c4-8f31-bc041fb07b3c","Type":"ContainerDied","Data":"36db728ef80e6e234991810c3f372fbf4cf4b3b5106eb7812562616dd262075f"} Nov 22 08:54:01 crc kubenswrapper[4789]: I1122 08:54:01.299634 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"9f40e5f0-7b57-47c4-8f31-bc041fb07b3c","Type":"ContainerDied","Data":"3844f5bf126c60293b5597ae121bee3569ae3b40cee283783c61b0e1ca44917a"} Nov 22 08:54:04 crc kubenswrapper[4789]: I1122 08:54:04.130907 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Nov 22 08:54:04 crc kubenswrapper[4789]: I1122 08:54:04.185931 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9f40e5f0-7b57-47c4-8f31-bc041fb07b3c-etc-machine-id\") pod \"9f40e5f0-7b57-47c4-8f31-bc041fb07b3c\" (UID: \"9f40e5f0-7b57-47c4-8f31-bc041fb07b3c\") " Nov 22 08:54:04 crc kubenswrapper[4789]: I1122 08:54:04.186058 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9f40e5f0-7b57-47c4-8f31-bc041fb07b3c-config-data-custom\") pod \"9f40e5f0-7b57-47c4-8f31-bc041fb07b3c\" (UID: \"9f40e5f0-7b57-47c4-8f31-bc041fb07b3c\") " Nov 22 08:54:04 crc kubenswrapper[4789]: I1122 08:54:04.186160 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pdnbw\" (UniqueName: \"kubernetes.io/projected/9f40e5f0-7b57-47c4-8f31-bc041fb07b3c-kube-api-access-pdnbw\") pod \"9f40e5f0-7b57-47c4-8f31-bc041fb07b3c\" (UID: \"9f40e5f0-7b57-47c4-8f31-bc041fb07b3c\") " Nov 22 08:54:04 crc kubenswrapper[4789]: I1122 08:54:04.186221 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9f40e5f0-7b57-47c4-8f31-bc041fb07b3c-scripts\") pod \"9f40e5f0-7b57-47c4-8f31-bc041fb07b3c\" (UID: \"9f40e5f0-7b57-47c4-8f31-bc041fb07b3c\") " Nov 22 08:54:04 crc kubenswrapper[4789]: I1122 08:54:04.186278 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9f40e5f0-7b57-47c4-8f31-bc041fb07b3c-logs\") pod \"9f40e5f0-7b57-47c4-8f31-bc041fb07b3c\" (UID: \"9f40e5f0-7b57-47c4-8f31-bc041fb07b3c\") " Nov 22 08:54:04 crc kubenswrapper[4789]: I1122 08:54:04.186309 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f40e5f0-7b57-47c4-8f31-bc041fb07b3c-config-data\") pod \"9f40e5f0-7b57-47c4-8f31-bc041fb07b3c\" (UID: \"9f40e5f0-7b57-47c4-8f31-bc041fb07b3c\") " Nov 22 08:54:04 crc kubenswrapper[4789]: I1122 08:54:04.186876 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9f40e5f0-7b57-47c4-8f31-bc041fb07b3c-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "9f40e5f0-7b57-47c4-8f31-bc041fb07b3c" (UID: "9f40e5f0-7b57-47c4-8f31-bc041fb07b3c"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 08:54:04 crc kubenswrapper[4789]: I1122 08:54:04.187263 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f40e5f0-7b57-47c4-8f31-bc041fb07b3c-combined-ca-bundle\") pod \"9f40e5f0-7b57-47c4-8f31-bc041fb07b3c\" (UID: \"9f40e5f0-7b57-47c4-8f31-bc041fb07b3c\") " Nov 22 08:54:04 crc kubenswrapper[4789]: I1122 08:54:04.187533 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9f40e5f0-7b57-47c4-8f31-bc041fb07b3c-logs" (OuterVolumeSpecName: "logs") pod "9f40e5f0-7b57-47c4-8f31-bc041fb07b3c" (UID: "9f40e5f0-7b57-47c4-8f31-bc041fb07b3c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:54:04 crc kubenswrapper[4789]: I1122 08:54:04.188080 4789 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9f40e5f0-7b57-47c4-8f31-bc041fb07b3c-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 22 08:54:04 crc kubenswrapper[4789]: I1122 08:54:04.188107 4789 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9f40e5f0-7b57-47c4-8f31-bc041fb07b3c-logs\") on node \"crc\" DevicePath \"\"" Nov 22 08:54:04 crc kubenswrapper[4789]: I1122 08:54:04.191905 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f40e5f0-7b57-47c4-8f31-bc041fb07b3c-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "9f40e5f0-7b57-47c4-8f31-bc041fb07b3c" (UID: "9f40e5f0-7b57-47c4-8f31-bc041fb07b3c"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:54:04 crc kubenswrapper[4789]: I1122 08:54:04.192047 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f40e5f0-7b57-47c4-8f31-bc041fb07b3c-kube-api-access-pdnbw" (OuterVolumeSpecName: "kube-api-access-pdnbw") pod "9f40e5f0-7b57-47c4-8f31-bc041fb07b3c" (UID: "9f40e5f0-7b57-47c4-8f31-bc041fb07b3c"). InnerVolumeSpecName "kube-api-access-pdnbw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:54:04 crc kubenswrapper[4789]: I1122 08:54:04.193290 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f40e5f0-7b57-47c4-8f31-bc041fb07b3c-scripts" (OuterVolumeSpecName: "scripts") pod "9f40e5f0-7b57-47c4-8f31-bc041fb07b3c" (UID: "9f40e5f0-7b57-47c4-8f31-bc041fb07b3c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:54:04 crc kubenswrapper[4789]: I1122 08:54:04.222272 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f40e5f0-7b57-47c4-8f31-bc041fb07b3c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9f40e5f0-7b57-47c4-8f31-bc041fb07b3c" (UID: "9f40e5f0-7b57-47c4-8f31-bc041fb07b3c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:54:04 crc kubenswrapper[4789]: I1122 08:54:04.259236 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f40e5f0-7b57-47c4-8f31-bc041fb07b3c-config-data" (OuterVolumeSpecName: "config-data") pod "9f40e5f0-7b57-47c4-8f31-bc041fb07b3c" (UID: "9f40e5f0-7b57-47c4-8f31-bc041fb07b3c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:54:04 crc kubenswrapper[4789]: I1122 08:54:04.289012 4789 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f40e5f0-7b57-47c4-8f31-bc041fb07b3c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:54:04 crc kubenswrapper[4789]: I1122 08:54:04.289046 4789 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9f40e5f0-7b57-47c4-8f31-bc041fb07b3c-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 22 08:54:04 crc kubenswrapper[4789]: I1122 08:54:04.289058 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pdnbw\" (UniqueName: \"kubernetes.io/projected/9f40e5f0-7b57-47c4-8f31-bc041fb07b3c-kube-api-access-pdnbw\") on node \"crc\" DevicePath \"\"" Nov 22 08:54:04 crc kubenswrapper[4789]: I1122 08:54:04.289067 4789 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9f40e5f0-7b57-47c4-8f31-bc041fb07b3c-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:54:04 crc kubenswrapper[4789]: I1122 08:54:04.289075 4789 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f40e5f0-7b57-47c4-8f31-bc041fb07b3c-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:54:04 crc kubenswrapper[4789]: I1122 08:54:04.333481 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"9f40e5f0-7b57-47c4-8f31-bc041fb07b3c","Type":"ContainerDied","Data":"2fe01025cd09f339780f5cbb351b5c9cc91ca4fe2cd75fead928dfa1c4e5538f"} Nov 22 08:54:04 crc kubenswrapper[4789]: I1122 08:54:04.333523 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Nov 22 08:54:04 crc kubenswrapper[4789]: I1122 08:54:04.333548 4789 scope.go:117] "RemoveContainer" containerID="36db728ef80e6e234991810c3f372fbf4cf4b3b5106eb7812562616dd262075f" Nov 22 08:54:04 crc kubenswrapper[4789]: I1122 08:54:04.335873 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"46af0a4f-fe66-47b9-b9d9-892c3eb3d97c","Type":"ContainerStarted","Data":"49d5d401ae263afccfb4c1717363fd6219980a592b2a9754a395aa281e5faabd"} Nov 22 08:54:04 crc kubenswrapper[4789]: I1122 08:54:04.369492 4789 scope.go:117] "RemoveContainer" containerID="3844f5bf126c60293b5597ae121bee3569ae3b40cee283783c61b0e1ca44917a" Nov 22 08:54:04 crc kubenswrapper[4789]: I1122 08:54:04.371091 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-api-0"] Nov 22 08:54:04 crc kubenswrapper[4789]: I1122 08:54:04.391332 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-api-0"] Nov 22 08:54:04 crc kubenswrapper[4789]: I1122 08:54:04.398149 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-api-0"] Nov 22 08:54:04 crc kubenswrapper[4789]: E1122 08:54:04.398701 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f40e5f0-7b57-47c4-8f31-bc041fb07b3c" containerName="manila-api" Nov 22 08:54:04 crc kubenswrapper[4789]: I1122 08:54:04.398731 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f40e5f0-7b57-47c4-8f31-bc041fb07b3c" containerName="manila-api" Nov 22 08:54:04 crc kubenswrapper[4789]: E1122 08:54:04.398775 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f40e5f0-7b57-47c4-8f31-bc041fb07b3c" containerName="manila-api-log" Nov 22 08:54:04 crc kubenswrapper[4789]: I1122 08:54:04.398785 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f40e5f0-7b57-47c4-8f31-bc041fb07b3c" containerName="manila-api-log" Nov 22 08:54:04 crc kubenswrapper[4789]: I1122 08:54:04.398991 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f40e5f0-7b57-47c4-8f31-bc041fb07b3c" containerName="manila-api-log" Nov 22 08:54:04 crc kubenswrapper[4789]: I1122 08:54:04.399021 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f40e5f0-7b57-47c4-8f31-bc041fb07b3c" containerName="manila-api" Nov 22 08:54:04 crc kubenswrapper[4789]: I1122 08:54:04.400249 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Nov 22 08:54:04 crc kubenswrapper[4789]: I1122 08:54:04.402503 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-manila-public-svc" Nov 22 08:54:04 crc kubenswrapper[4789]: I1122 08:54:04.402911 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-manila-internal-svc" Nov 22 08:54:04 crc kubenswrapper[4789]: I1122 08:54:04.415039 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-api-config-data" Nov 22 08:54:04 crc kubenswrapper[4789]: I1122 08:54:04.417240 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Nov 22 08:54:04 crc kubenswrapper[4789]: I1122 08:54:04.492416 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e21ea1ba-4437-4d76-be37-0b285909d631-public-tls-certs\") pod \"manila-api-0\" (UID: \"e21ea1ba-4437-4d76-be37-0b285909d631\") " pod="openstack/manila-api-0" Nov 22 08:54:04 crc kubenswrapper[4789]: I1122 08:54:04.492480 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e21ea1ba-4437-4d76-be37-0b285909d631-scripts\") pod \"manila-api-0\" (UID: \"e21ea1ba-4437-4d76-be37-0b285909d631\") " pod="openstack/manila-api-0" Nov 22 08:54:04 crc kubenswrapper[4789]: I1122 08:54:04.492570 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cg65z\" (UniqueName: \"kubernetes.io/projected/e21ea1ba-4437-4d76-be37-0b285909d631-kube-api-access-cg65z\") pod \"manila-api-0\" (UID: \"e21ea1ba-4437-4d76-be37-0b285909d631\") " pod="openstack/manila-api-0" Nov 22 08:54:04 crc kubenswrapper[4789]: I1122 08:54:04.492609 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e21ea1ba-4437-4d76-be37-0b285909d631-logs\") pod \"manila-api-0\" (UID: \"e21ea1ba-4437-4d76-be37-0b285909d631\") " pod="openstack/manila-api-0" Nov 22 08:54:04 crc kubenswrapper[4789]: I1122 08:54:04.492632 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e21ea1ba-4437-4d76-be37-0b285909d631-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"e21ea1ba-4437-4d76-be37-0b285909d631\") " pod="openstack/manila-api-0" Nov 22 08:54:04 crc kubenswrapper[4789]: I1122 08:54:04.492650 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e21ea1ba-4437-4d76-be37-0b285909d631-config-data\") pod \"manila-api-0\" (UID: \"e21ea1ba-4437-4d76-be37-0b285909d631\") " pod="openstack/manila-api-0" Nov 22 08:54:04 crc kubenswrapper[4789]: I1122 08:54:04.492670 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e21ea1ba-4437-4d76-be37-0b285909d631-config-data-custom\") pod \"manila-api-0\" (UID: \"e21ea1ba-4437-4d76-be37-0b285909d631\") " pod="openstack/manila-api-0" Nov 22 08:54:04 crc kubenswrapper[4789]: I1122 08:54:04.492689 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e21ea1ba-4437-4d76-be37-0b285909d631-internal-tls-certs\") pod \"manila-api-0\" (UID: \"e21ea1ba-4437-4d76-be37-0b285909d631\") " pod="openstack/manila-api-0" Nov 22 08:54:04 crc kubenswrapper[4789]: I1122 08:54:04.492724 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e21ea1ba-4437-4d76-be37-0b285909d631-etc-machine-id\") pod \"manila-api-0\" (UID: \"e21ea1ba-4437-4d76-be37-0b285909d631\") " pod="openstack/manila-api-0" Nov 22 08:54:04 crc kubenswrapper[4789]: I1122 08:54:04.594189 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e21ea1ba-4437-4d76-be37-0b285909d631-public-tls-certs\") pod \"manila-api-0\" (UID: \"e21ea1ba-4437-4d76-be37-0b285909d631\") " pod="openstack/manila-api-0" Nov 22 08:54:04 crc kubenswrapper[4789]: I1122 08:54:04.594273 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e21ea1ba-4437-4d76-be37-0b285909d631-scripts\") pod \"manila-api-0\" (UID: \"e21ea1ba-4437-4d76-be37-0b285909d631\") " pod="openstack/manila-api-0" Nov 22 08:54:04 crc kubenswrapper[4789]: I1122 08:54:04.594336 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cg65z\" (UniqueName: \"kubernetes.io/projected/e21ea1ba-4437-4d76-be37-0b285909d631-kube-api-access-cg65z\") pod \"manila-api-0\" (UID: \"e21ea1ba-4437-4d76-be37-0b285909d631\") " pod="openstack/manila-api-0" Nov 22 08:54:04 crc kubenswrapper[4789]: I1122 08:54:04.594377 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e21ea1ba-4437-4d76-be37-0b285909d631-logs\") pod \"manila-api-0\" (UID: \"e21ea1ba-4437-4d76-be37-0b285909d631\") " pod="openstack/manila-api-0" Nov 22 08:54:04 crc kubenswrapper[4789]: I1122 08:54:04.594399 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e21ea1ba-4437-4d76-be37-0b285909d631-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"e21ea1ba-4437-4d76-be37-0b285909d631\") " pod="openstack/manila-api-0" Nov 22 08:54:04 crc kubenswrapper[4789]: I1122 08:54:04.594416 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e21ea1ba-4437-4d76-be37-0b285909d631-config-data\") pod \"manila-api-0\" (UID: \"e21ea1ba-4437-4d76-be37-0b285909d631\") " pod="openstack/manila-api-0" Nov 22 08:54:04 crc kubenswrapper[4789]: I1122 08:54:04.594437 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e21ea1ba-4437-4d76-be37-0b285909d631-config-data-custom\") pod \"manila-api-0\" (UID: \"e21ea1ba-4437-4d76-be37-0b285909d631\") " pod="openstack/manila-api-0" Nov 22 08:54:04 crc kubenswrapper[4789]: I1122 08:54:04.594453 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e21ea1ba-4437-4d76-be37-0b285909d631-internal-tls-certs\") pod \"manila-api-0\" (UID: \"e21ea1ba-4437-4d76-be37-0b285909d631\") " pod="openstack/manila-api-0" Nov 22 08:54:04 crc kubenswrapper[4789]: I1122 08:54:04.594488 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e21ea1ba-4437-4d76-be37-0b285909d631-etc-machine-id\") pod \"manila-api-0\" (UID: \"e21ea1ba-4437-4d76-be37-0b285909d631\") " pod="openstack/manila-api-0" Nov 22 08:54:04 crc kubenswrapper[4789]: I1122 08:54:04.594587 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e21ea1ba-4437-4d76-be37-0b285909d631-etc-machine-id\") pod \"manila-api-0\" (UID: \"e21ea1ba-4437-4d76-be37-0b285909d631\") " pod="openstack/manila-api-0" Nov 22 08:54:04 crc kubenswrapper[4789]: I1122 08:54:04.597289 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e21ea1ba-4437-4d76-be37-0b285909d631-logs\") pod \"manila-api-0\" (UID: \"e21ea1ba-4437-4d76-be37-0b285909d631\") " pod="openstack/manila-api-0" Nov 22 08:54:04 crc kubenswrapper[4789]: I1122 08:54:04.599389 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e21ea1ba-4437-4d76-be37-0b285909d631-public-tls-certs\") pod \"manila-api-0\" (UID: \"e21ea1ba-4437-4d76-be37-0b285909d631\") " pod="openstack/manila-api-0" Nov 22 08:54:04 crc kubenswrapper[4789]: I1122 08:54:04.600381 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e21ea1ba-4437-4d76-be37-0b285909d631-config-data\") pod \"manila-api-0\" (UID: \"e21ea1ba-4437-4d76-be37-0b285909d631\") " pod="openstack/manila-api-0" Nov 22 08:54:04 crc kubenswrapper[4789]: I1122 08:54:04.601284 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e21ea1ba-4437-4d76-be37-0b285909d631-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"e21ea1ba-4437-4d76-be37-0b285909d631\") " pod="openstack/manila-api-0" Nov 22 08:54:04 crc kubenswrapper[4789]: I1122 08:54:04.602194 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e21ea1ba-4437-4d76-be37-0b285909d631-config-data-custom\") pod \"manila-api-0\" (UID: \"e21ea1ba-4437-4d76-be37-0b285909d631\") " pod="openstack/manila-api-0" Nov 22 08:54:04 crc kubenswrapper[4789]: I1122 08:54:04.603421 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e21ea1ba-4437-4d76-be37-0b285909d631-internal-tls-certs\") pod \"manila-api-0\" (UID: \"e21ea1ba-4437-4d76-be37-0b285909d631\") " pod="openstack/manila-api-0" Nov 22 08:54:04 crc kubenswrapper[4789]: I1122 08:54:04.603546 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e21ea1ba-4437-4d76-be37-0b285909d631-scripts\") pod \"manila-api-0\" (UID: \"e21ea1ba-4437-4d76-be37-0b285909d631\") " pod="openstack/manila-api-0" Nov 22 08:54:04 crc kubenswrapper[4789]: I1122 08:54:04.621101 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cg65z\" (UniqueName: \"kubernetes.io/projected/e21ea1ba-4437-4d76-be37-0b285909d631-kube-api-access-cg65z\") pod \"manila-api-0\" (UID: \"e21ea1ba-4437-4d76-be37-0b285909d631\") " pod="openstack/manila-api-0" Nov 22 08:54:04 crc kubenswrapper[4789]: I1122 08:54:04.746844 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Nov 22 08:54:05 crc kubenswrapper[4789]: I1122 08:54:05.023118 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6885d49d55-m8c6x" Nov 22 08:54:05 crc kubenswrapper[4789]: I1122 08:54:05.097523 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6dc44c56c-bclr4"] Nov 22 08:54:05 crc kubenswrapper[4789]: I1122 08:54:05.097939 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6dc44c56c-bclr4" podUID="7785105e-179b-4b0b-9457-cd513d0b762d" containerName="dnsmasq-dns" containerID="cri-o://ec54bb2c6c2fd3993a4998368dd12ab72c475858fe45b965cb565d3a6b79af1f" gracePeriod=10 Nov 22 08:54:05 crc kubenswrapper[4789]: I1122 08:54:05.352687 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"46af0a4f-fe66-47b9-b9d9-892c3eb3d97c","Type":"ContainerStarted","Data":"b6d59c915dfda3979f0214cccd7333388cee6ab9e25416acb131775f14556d36"} Nov 22 08:54:05 crc kubenswrapper[4789]: I1122 08:54:05.359453 4789 generic.go:334] "Generic (PLEG): container finished" podID="b3e0f7a4-cde7-4da6-8339-615054807f46" containerID="0ba876bc146a633673594ffc7a55b1ef15f6d31b6f68e177c9d9b989cfe96170" exitCode=0 Nov 22 08:54:05 crc kubenswrapper[4789]: I1122 08:54:05.359516 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b3e0f7a4-cde7-4da6-8339-615054807f46","Type":"ContainerDied","Data":"0ba876bc146a633673594ffc7a55b1ef15f6d31b6f68e177c9d9b989cfe96170"} Nov 22 08:54:05 crc kubenswrapper[4789]: I1122 08:54:05.361007 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"3b679ee8-b6ac-4313-ac79-abd465d1b0c4","Type":"ContainerStarted","Data":"18527ce10ae16f3927ffa1f263ae903cc74829ed9695fa52e8ea021020d91dbc"} Nov 22 08:54:05 crc kubenswrapper[4789]: I1122 08:54:05.361033 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"3b679ee8-b6ac-4313-ac79-abd465d1b0c4","Type":"ContainerStarted","Data":"f79b8f35c30c6903423dfa326fc73e1e524773a5acfa2c6d1ab55bb96502e49b"} Nov 22 08:54:05 crc kubenswrapper[4789]: I1122 08:54:05.364999 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Nov 22 08:54:05 crc kubenswrapper[4789]: I1122 08:54:05.368378 4789 generic.go:334] "Generic (PLEG): container finished" podID="7785105e-179b-4b0b-9457-cd513d0b762d" containerID="ec54bb2c6c2fd3993a4998368dd12ab72c475858fe45b965cb565d3a6b79af1f" exitCode=0 Nov 22 08:54:05 crc kubenswrapper[4789]: I1122 08:54:05.368426 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6dc44c56c-bclr4" event={"ID":"7785105e-179b-4b0b-9457-cd513d0b762d","Type":"ContainerDied","Data":"ec54bb2c6c2fd3993a4998368dd12ab72c475858fe45b965cb565d3a6b79af1f"} Nov 22 08:54:05 crc kubenswrapper[4789]: I1122 08:54:05.371926 4789 patch_prober.go:28] interesting pod/machine-config-daemon-pxvrm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 08:54:05 crc kubenswrapper[4789]: I1122 08:54:05.371994 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 08:54:05 crc kubenswrapper[4789]: I1122 08:54:05.372041 4789 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" Nov 22 08:54:05 crc kubenswrapper[4789]: I1122 08:54:05.372935 4789 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2e2387a756e8b093bdd9ccbfb31650bf2886cf3ee1392229517961fc3889046d"} pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 22 08:54:05 crc kubenswrapper[4789]: I1122 08:54:05.372998 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" containerID="cri-o://2e2387a756e8b093bdd9ccbfb31650bf2886cf3ee1392229517961fc3889046d" gracePeriod=600 Nov 22 08:54:05 crc kubenswrapper[4789]: W1122 08:54:05.387566 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode21ea1ba_4437_4d76_be37_0b285909d631.slice/crio-80340d999bda30d3932c894b7f14021356148146cd3e0505080fdf3f53413b9f WatchSource:0}: Error finding container 80340d999bda30d3932c894b7f14021356148146cd3e0505080fdf3f53413b9f: Status 404 returned error can't find the container with id 80340d999bda30d3932c894b7f14021356148146cd3e0505080fdf3f53413b9f Nov 22 08:54:05 crc kubenswrapper[4789]: I1122 08:54:05.409182 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-share-share1-0" podStartSLOduration=3.079551866 podStartE2EDuration="11.409164432s" podCreationTimestamp="2025-11-22 08:53:54 +0000 UTC" firstStartedPulling="2025-11-22 08:53:55.557236501 +0000 UTC m=+3449.791636774" lastFinishedPulling="2025-11-22 08:54:03.886849057 +0000 UTC m=+3458.121249340" observedRunningTime="2025-11-22 08:54:05.408150875 +0000 UTC m=+3459.642551158" watchObservedRunningTime="2025-11-22 08:54:05.409164432 +0000 UTC m=+3459.643564705" Nov 22 08:54:05 crc kubenswrapper[4789]: I1122 08:54:05.421808 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-scheduler-0" podStartSLOduration=5.66485337 podStartE2EDuration="11.421781711s" podCreationTimestamp="2025-11-22 08:53:54 +0000 UTC" firstStartedPulling="2025-11-22 08:53:58.030659253 +0000 UTC m=+3452.265059536" lastFinishedPulling="2025-11-22 08:54:03.787587604 +0000 UTC m=+3458.021987877" observedRunningTime="2025-11-22 08:54:05.377123619 +0000 UTC m=+3459.611523892" watchObservedRunningTime="2025-11-22 08:54:05.421781711 +0000 UTC m=+3459.656181984" Nov 22 08:54:05 crc kubenswrapper[4789]: I1122 08:54:05.696917 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 08:54:05 crc kubenswrapper[4789]: I1122 08:54:05.819346 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b3e0f7a4-cde7-4da6-8339-615054807f46-log-httpd\") pod \"b3e0f7a4-cde7-4da6-8339-615054807f46\" (UID: \"b3e0f7a4-cde7-4da6-8339-615054807f46\") " Nov 22 08:54:05 crc kubenswrapper[4789]: I1122 08:54:05.819635 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3e0f7a4-cde7-4da6-8339-615054807f46-combined-ca-bundle\") pod \"b3e0f7a4-cde7-4da6-8339-615054807f46\" (UID: \"b3e0f7a4-cde7-4da6-8339-615054807f46\") " Nov 22 08:54:05 crc kubenswrapper[4789]: I1122 08:54:05.819682 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8ctgf\" (UniqueName: \"kubernetes.io/projected/b3e0f7a4-cde7-4da6-8339-615054807f46-kube-api-access-8ctgf\") pod \"b3e0f7a4-cde7-4da6-8339-615054807f46\" (UID: \"b3e0f7a4-cde7-4da6-8339-615054807f46\") " Nov 22 08:54:05 crc kubenswrapper[4789]: I1122 08:54:05.819714 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3e0f7a4-cde7-4da6-8339-615054807f46-config-data\") pod \"b3e0f7a4-cde7-4da6-8339-615054807f46\" (UID: \"b3e0f7a4-cde7-4da6-8339-615054807f46\") " Nov 22 08:54:05 crc kubenswrapper[4789]: I1122 08:54:05.819828 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b3e0f7a4-cde7-4da6-8339-615054807f46-sg-core-conf-yaml\") pod \"b3e0f7a4-cde7-4da6-8339-615054807f46\" (UID: \"b3e0f7a4-cde7-4da6-8339-615054807f46\") " Nov 22 08:54:05 crc kubenswrapper[4789]: I1122 08:54:05.819929 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b3e0f7a4-cde7-4da6-8339-615054807f46-scripts\") pod \"b3e0f7a4-cde7-4da6-8339-615054807f46\" (UID: \"b3e0f7a4-cde7-4da6-8339-615054807f46\") " Nov 22 08:54:05 crc kubenswrapper[4789]: I1122 08:54:05.819966 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b3e0f7a4-cde7-4da6-8339-615054807f46-run-httpd\") pod \"b3e0f7a4-cde7-4da6-8339-615054807f46\" (UID: \"b3e0f7a4-cde7-4da6-8339-615054807f46\") " Nov 22 08:54:05 crc kubenswrapper[4789]: I1122 08:54:05.819968 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b3e0f7a4-cde7-4da6-8339-615054807f46-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "b3e0f7a4-cde7-4da6-8339-615054807f46" (UID: "b3e0f7a4-cde7-4da6-8339-615054807f46"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:54:05 crc kubenswrapper[4789]: I1122 08:54:05.819994 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/b3e0f7a4-cde7-4da6-8339-615054807f46-ceilometer-tls-certs\") pod \"b3e0f7a4-cde7-4da6-8339-615054807f46\" (UID: \"b3e0f7a4-cde7-4da6-8339-615054807f46\") " Nov 22 08:54:05 crc kubenswrapper[4789]: I1122 08:54:05.821005 4789 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b3e0f7a4-cde7-4da6-8339-615054807f46-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 22 08:54:05 crc kubenswrapper[4789]: I1122 08:54:05.823108 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b3e0f7a4-cde7-4da6-8339-615054807f46-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "b3e0f7a4-cde7-4da6-8339-615054807f46" (UID: "b3e0f7a4-cde7-4da6-8339-615054807f46"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:54:05 crc kubenswrapper[4789]: I1122 08:54:05.828947 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3e0f7a4-cde7-4da6-8339-615054807f46-kube-api-access-8ctgf" (OuterVolumeSpecName: "kube-api-access-8ctgf") pod "b3e0f7a4-cde7-4da6-8339-615054807f46" (UID: "b3e0f7a4-cde7-4da6-8339-615054807f46"). InnerVolumeSpecName "kube-api-access-8ctgf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:54:05 crc kubenswrapper[4789]: I1122 08:54:05.855617 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3e0f7a4-cde7-4da6-8339-615054807f46-scripts" (OuterVolumeSpecName: "scripts") pod "b3e0f7a4-cde7-4da6-8339-615054807f46" (UID: "b3e0f7a4-cde7-4da6-8339-615054807f46"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:54:05 crc kubenswrapper[4789]: I1122 08:54:05.909777 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3e0f7a4-cde7-4da6-8339-615054807f46-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "b3e0f7a4-cde7-4da6-8339-615054807f46" (UID: "b3e0f7a4-cde7-4da6-8339-615054807f46"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:54:05 crc kubenswrapper[4789]: I1122 08:54:05.923386 4789 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b3e0f7a4-cde7-4da6-8339-615054807f46-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 22 08:54:05 crc kubenswrapper[4789]: I1122 08:54:05.923426 4789 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b3e0f7a4-cde7-4da6-8339-615054807f46-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:54:05 crc kubenswrapper[4789]: I1122 08:54:05.923439 4789 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b3e0f7a4-cde7-4da6-8339-615054807f46-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 22 08:54:05 crc kubenswrapper[4789]: I1122 08:54:05.923474 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8ctgf\" (UniqueName: \"kubernetes.io/projected/b3e0f7a4-cde7-4da6-8339-615054807f46-kube-api-access-8ctgf\") on node \"crc\" DevicePath \"\"" Nov 22 08:54:05 crc kubenswrapper[4789]: I1122 08:54:05.980094 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9f40e5f0-7b57-47c4-8f31-bc041fb07b3c" path="/var/lib/kubelet/pods/9f40e5f0-7b57-47c4-8f31-bc041fb07b3c/volumes" Nov 22 08:54:05 crc kubenswrapper[4789]: I1122 08:54:05.982721 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3e0f7a4-cde7-4da6-8339-615054807f46-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "b3e0f7a4-cde7-4da6-8339-615054807f46" (UID: "b3e0f7a4-cde7-4da6-8339-615054807f46"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:54:05 crc kubenswrapper[4789]: I1122 08:54:05.996859 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3e0f7a4-cde7-4da6-8339-615054807f46-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b3e0f7a4-cde7-4da6-8339-615054807f46" (UID: "b3e0f7a4-cde7-4da6-8339-615054807f46"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:54:05 crc kubenswrapper[4789]: I1122 08:54:05.996939 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3e0f7a4-cde7-4da6-8339-615054807f46-config-data" (OuterVolumeSpecName: "config-data") pod "b3e0f7a4-cde7-4da6-8339-615054807f46" (UID: "b3e0f7a4-cde7-4da6-8339-615054807f46"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:54:06 crc kubenswrapper[4789]: I1122 08:54:06.024235 4789 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3e0f7a4-cde7-4da6-8339-615054807f46-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:54:06 crc kubenswrapper[4789]: I1122 08:54:06.024274 4789 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3e0f7a4-cde7-4da6-8339-615054807f46-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:54:06 crc kubenswrapper[4789]: I1122 08:54:06.024285 4789 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/b3e0f7a4-cde7-4da6-8339-615054807f46-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 22 08:54:06 crc kubenswrapper[4789]: I1122 08:54:06.053361 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6dc44c56c-bclr4" Nov 22 08:54:06 crc kubenswrapper[4789]: I1122 08:54:06.125608 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qgzxc\" (UniqueName: \"kubernetes.io/projected/7785105e-179b-4b0b-9457-cd513d0b762d-kube-api-access-qgzxc\") pod \"7785105e-179b-4b0b-9457-cd513d0b762d\" (UID: \"7785105e-179b-4b0b-9457-cd513d0b762d\") " Nov 22 08:54:06 crc kubenswrapper[4789]: I1122 08:54:06.126012 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7785105e-179b-4b0b-9457-cd513d0b762d-ovsdbserver-nb\") pod \"7785105e-179b-4b0b-9457-cd513d0b762d\" (UID: \"7785105e-179b-4b0b-9457-cd513d0b762d\") " Nov 22 08:54:06 crc kubenswrapper[4789]: I1122 08:54:06.126162 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7785105e-179b-4b0b-9457-cd513d0b762d-dns-svc\") pod \"7785105e-179b-4b0b-9457-cd513d0b762d\" (UID: \"7785105e-179b-4b0b-9457-cd513d0b762d\") " Nov 22 08:54:06 crc kubenswrapper[4789]: I1122 08:54:06.126206 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7785105e-179b-4b0b-9457-cd513d0b762d-ovsdbserver-sb\") pod \"7785105e-179b-4b0b-9457-cd513d0b762d\" (UID: \"7785105e-179b-4b0b-9457-cd513d0b762d\") " Nov 22 08:54:06 crc kubenswrapper[4789]: I1122 08:54:06.126277 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/7785105e-179b-4b0b-9457-cd513d0b762d-openstack-edpm-ipam\") pod \"7785105e-179b-4b0b-9457-cd513d0b762d\" (UID: \"7785105e-179b-4b0b-9457-cd513d0b762d\") " Nov 22 08:54:06 crc kubenswrapper[4789]: I1122 08:54:06.126320 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7785105e-179b-4b0b-9457-cd513d0b762d-config\") pod \"7785105e-179b-4b0b-9457-cd513d0b762d\" (UID: \"7785105e-179b-4b0b-9457-cd513d0b762d\") " Nov 22 08:54:06 crc kubenswrapper[4789]: I1122 08:54:06.130544 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7785105e-179b-4b0b-9457-cd513d0b762d-kube-api-access-qgzxc" (OuterVolumeSpecName: "kube-api-access-qgzxc") pod "7785105e-179b-4b0b-9457-cd513d0b762d" (UID: "7785105e-179b-4b0b-9457-cd513d0b762d"). InnerVolumeSpecName "kube-api-access-qgzxc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:54:06 crc kubenswrapper[4789]: I1122 08:54:06.191024 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7785105e-179b-4b0b-9457-cd513d0b762d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7785105e-179b-4b0b-9457-cd513d0b762d" (UID: "7785105e-179b-4b0b-9457-cd513d0b762d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:54:06 crc kubenswrapper[4789]: I1122 08:54:06.196562 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7785105e-179b-4b0b-9457-cd513d0b762d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7785105e-179b-4b0b-9457-cd513d0b762d" (UID: "7785105e-179b-4b0b-9457-cd513d0b762d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:54:06 crc kubenswrapper[4789]: I1122 08:54:06.199124 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7785105e-179b-4b0b-9457-cd513d0b762d-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "7785105e-179b-4b0b-9457-cd513d0b762d" (UID: "7785105e-179b-4b0b-9457-cd513d0b762d"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:54:06 crc kubenswrapper[4789]: I1122 08:54:06.203340 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7785105e-179b-4b0b-9457-cd513d0b762d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7785105e-179b-4b0b-9457-cd513d0b762d" (UID: "7785105e-179b-4b0b-9457-cd513d0b762d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:54:06 crc kubenswrapper[4789]: I1122 08:54:06.206474 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7785105e-179b-4b0b-9457-cd513d0b762d-config" (OuterVolumeSpecName: "config") pod "7785105e-179b-4b0b-9457-cd513d0b762d" (UID: "7785105e-179b-4b0b-9457-cd513d0b762d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 08:54:06 crc kubenswrapper[4789]: I1122 08:54:06.230653 4789 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7785105e-179b-4b0b-9457-cd513d0b762d-config\") on node \"crc\" DevicePath \"\"" Nov 22 08:54:06 crc kubenswrapper[4789]: I1122 08:54:06.230684 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qgzxc\" (UniqueName: \"kubernetes.io/projected/7785105e-179b-4b0b-9457-cd513d0b762d-kube-api-access-qgzxc\") on node \"crc\" DevicePath \"\"" Nov 22 08:54:06 crc kubenswrapper[4789]: I1122 08:54:06.230693 4789 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7785105e-179b-4b0b-9457-cd513d0b762d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 22 08:54:06 crc kubenswrapper[4789]: I1122 08:54:06.230702 4789 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7785105e-179b-4b0b-9457-cd513d0b762d-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 22 08:54:06 crc kubenswrapper[4789]: I1122 08:54:06.230711 4789 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7785105e-179b-4b0b-9457-cd513d0b762d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 22 08:54:06 crc kubenswrapper[4789]: I1122 08:54:06.230719 4789 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/7785105e-179b-4b0b-9457-cd513d0b762d-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Nov 22 08:54:06 crc kubenswrapper[4789]: I1122 08:54:06.383653 4789 generic.go:334] "Generic (PLEG): container finished" podID="58aa8071-7ecc-4692-8789-2db04ec70647" containerID="2e2387a756e8b093bdd9ccbfb31650bf2886cf3ee1392229517961fc3889046d" exitCode=0 Nov 22 08:54:06 crc kubenswrapper[4789]: I1122 08:54:06.383726 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" event={"ID":"58aa8071-7ecc-4692-8789-2db04ec70647","Type":"ContainerDied","Data":"2e2387a756e8b093bdd9ccbfb31650bf2886cf3ee1392229517961fc3889046d"} Nov 22 08:54:06 crc kubenswrapper[4789]: I1122 08:54:06.383773 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" event={"ID":"58aa8071-7ecc-4692-8789-2db04ec70647","Type":"ContainerStarted","Data":"9a037c1a2417030bb4790bc38e61b3464365daf9c9e9963ba880fe8c29455d3b"} Nov 22 08:54:06 crc kubenswrapper[4789]: I1122 08:54:06.383796 4789 scope.go:117] "RemoveContainer" containerID="e6fcfa0e8fd700efa0433d598c23ce626ca7a007b2b45b45ee839e17923bad02" Nov 22 08:54:06 crc kubenswrapper[4789]: I1122 08:54:06.387415 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6dc44c56c-bclr4" Nov 22 08:54:06 crc kubenswrapper[4789]: I1122 08:54:06.387517 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6dc44c56c-bclr4" event={"ID":"7785105e-179b-4b0b-9457-cd513d0b762d","Type":"ContainerDied","Data":"c2e34834666a28e734d9e59dfa0e4cf2b249b55fab28d0800cc4646e97c30629"} Nov 22 08:54:06 crc kubenswrapper[4789]: I1122 08:54:06.392253 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 08:54:06 crc kubenswrapper[4789]: I1122 08:54:06.392500 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b3e0f7a4-cde7-4da6-8339-615054807f46","Type":"ContainerDied","Data":"4c51d2ca1d082219d1fe4a73d0ff639d303d7e5c7e73c3a9868682aade936de1"} Nov 22 08:54:06 crc kubenswrapper[4789]: I1122 08:54:06.400703 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"e21ea1ba-4437-4d76-be37-0b285909d631","Type":"ContainerStarted","Data":"9dfaa018a894a09e4663d842878d2a4949501dc5b3e5f524036c35d1b65e318a"} Nov 22 08:54:06 crc kubenswrapper[4789]: I1122 08:54:06.400744 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"e21ea1ba-4437-4d76-be37-0b285909d631","Type":"ContainerStarted","Data":"80340d999bda30d3932c894b7f14021356148146cd3e0505080fdf3f53413b9f"} Nov 22 08:54:06 crc kubenswrapper[4789]: I1122 08:54:06.438763 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6dc44c56c-bclr4"] Nov 22 08:54:06 crc kubenswrapper[4789]: I1122 08:54:06.455953 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6dc44c56c-bclr4"] Nov 22 08:54:06 crc kubenswrapper[4789]: I1122 08:54:06.475580 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:54:06 crc kubenswrapper[4789]: I1122 08:54:06.490521 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:54:06 crc kubenswrapper[4789]: I1122 08:54:06.513081 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:54:06 crc kubenswrapper[4789]: E1122 08:54:06.513528 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3e0f7a4-cde7-4da6-8339-615054807f46" containerName="ceilometer-central-agent" Nov 22 08:54:06 crc kubenswrapper[4789]: I1122 08:54:06.513545 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3e0f7a4-cde7-4da6-8339-615054807f46" containerName="ceilometer-central-agent" Nov 22 08:54:06 crc kubenswrapper[4789]: E1122 08:54:06.513563 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3e0f7a4-cde7-4da6-8339-615054807f46" containerName="ceilometer-central-agent" Nov 22 08:54:06 crc kubenswrapper[4789]: I1122 08:54:06.513570 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3e0f7a4-cde7-4da6-8339-615054807f46" containerName="ceilometer-central-agent" Nov 22 08:54:06 crc kubenswrapper[4789]: E1122 08:54:06.513589 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3e0f7a4-cde7-4da6-8339-615054807f46" containerName="ceilometer-notification-agent" Nov 22 08:54:06 crc kubenswrapper[4789]: I1122 08:54:06.513595 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3e0f7a4-cde7-4da6-8339-615054807f46" containerName="ceilometer-notification-agent" Nov 22 08:54:06 crc kubenswrapper[4789]: E1122 08:54:06.513612 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3e0f7a4-cde7-4da6-8339-615054807f46" containerName="sg-core" Nov 22 08:54:06 crc kubenswrapper[4789]: I1122 08:54:06.513639 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3e0f7a4-cde7-4da6-8339-615054807f46" containerName="sg-core" Nov 22 08:54:06 crc kubenswrapper[4789]: E1122 08:54:06.513650 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7785105e-179b-4b0b-9457-cd513d0b762d" containerName="dnsmasq-dns" Nov 22 08:54:06 crc kubenswrapper[4789]: I1122 08:54:06.513655 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="7785105e-179b-4b0b-9457-cd513d0b762d" containerName="dnsmasq-dns" Nov 22 08:54:06 crc kubenswrapper[4789]: E1122 08:54:06.513666 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7785105e-179b-4b0b-9457-cd513d0b762d" containerName="init" Nov 22 08:54:06 crc kubenswrapper[4789]: I1122 08:54:06.513672 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="7785105e-179b-4b0b-9457-cd513d0b762d" containerName="init" Nov 22 08:54:06 crc kubenswrapper[4789]: E1122 08:54:06.513680 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3e0f7a4-cde7-4da6-8339-615054807f46" containerName="proxy-httpd" Nov 22 08:54:06 crc kubenswrapper[4789]: I1122 08:54:06.513687 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3e0f7a4-cde7-4da6-8339-615054807f46" containerName="proxy-httpd" Nov 22 08:54:06 crc kubenswrapper[4789]: I1122 08:54:06.513897 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3e0f7a4-cde7-4da6-8339-615054807f46" containerName="proxy-httpd" Nov 22 08:54:06 crc kubenswrapper[4789]: I1122 08:54:06.513917 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3e0f7a4-cde7-4da6-8339-615054807f46" containerName="ceilometer-central-agent" Nov 22 08:54:06 crc kubenswrapper[4789]: I1122 08:54:06.513925 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3e0f7a4-cde7-4da6-8339-615054807f46" containerName="sg-core" Nov 22 08:54:06 crc kubenswrapper[4789]: I1122 08:54:06.513936 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="7785105e-179b-4b0b-9457-cd513d0b762d" containerName="dnsmasq-dns" Nov 22 08:54:06 crc kubenswrapper[4789]: I1122 08:54:06.513947 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3e0f7a4-cde7-4da6-8339-615054807f46" containerName="ceilometer-notification-agent" Nov 22 08:54:06 crc kubenswrapper[4789]: I1122 08:54:06.514717 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3e0f7a4-cde7-4da6-8339-615054807f46" containerName="ceilometer-central-agent" Nov 22 08:54:06 crc kubenswrapper[4789]: I1122 08:54:06.516074 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 08:54:06 crc kubenswrapper[4789]: I1122 08:54:06.521566 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 22 08:54:06 crc kubenswrapper[4789]: I1122 08:54:06.521702 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 22 08:54:06 crc kubenswrapper[4789]: I1122 08:54:06.522025 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 22 08:54:06 crc kubenswrapper[4789]: I1122 08:54:06.525132 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:54:06 crc kubenswrapper[4789]: I1122 08:54:06.547849 4789 scope.go:117] "RemoveContainer" containerID="ec54bb2c6c2fd3993a4998368dd12ab72c475858fe45b965cb565d3a6b79af1f" Nov 22 08:54:06 crc kubenswrapper[4789]: I1122 08:54:06.574585 4789 scope.go:117] "RemoveContainer" containerID="7102520ed46ba5732ae374650631275c786a0e7dab1acfd1c3ee85a2dc51b901" Nov 22 08:54:06 crc kubenswrapper[4789]: I1122 08:54:06.619375 4789 scope.go:117] "RemoveContainer" containerID="57bc1e114749de45fd56a9c528a09c54ebfb4ff2b2cda822e7b1411b31acac80" Nov 22 08:54:06 crc kubenswrapper[4789]: I1122 08:54:06.649943 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/db7c20fc-ffc4-4425-9184-5f78bb01c3b3-log-httpd\") pod \"ceilometer-0\" (UID: \"db7c20fc-ffc4-4425-9184-5f78bb01c3b3\") " pod="openstack/ceilometer-0" Nov 22 08:54:06 crc kubenswrapper[4789]: I1122 08:54:06.650039 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/db7c20fc-ffc4-4425-9184-5f78bb01c3b3-run-httpd\") pod \"ceilometer-0\" (UID: \"db7c20fc-ffc4-4425-9184-5f78bb01c3b3\") " pod="openstack/ceilometer-0" Nov 22 08:54:06 crc kubenswrapper[4789]: I1122 08:54:06.650059 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/db7c20fc-ffc4-4425-9184-5f78bb01c3b3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"db7c20fc-ffc4-4425-9184-5f78bb01c3b3\") " pod="openstack/ceilometer-0" Nov 22 08:54:06 crc kubenswrapper[4789]: I1122 08:54:06.650257 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/db7c20fc-ffc4-4425-9184-5f78bb01c3b3-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"db7c20fc-ffc4-4425-9184-5f78bb01c3b3\") " pod="openstack/ceilometer-0" Nov 22 08:54:06 crc kubenswrapper[4789]: I1122 08:54:06.650292 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db7c20fc-ffc4-4425-9184-5f78bb01c3b3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"db7c20fc-ffc4-4425-9184-5f78bb01c3b3\") " pod="openstack/ceilometer-0" Nov 22 08:54:06 crc kubenswrapper[4789]: I1122 08:54:06.650437 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db7c20fc-ffc4-4425-9184-5f78bb01c3b3-config-data\") pod \"ceilometer-0\" (UID: \"db7c20fc-ffc4-4425-9184-5f78bb01c3b3\") " pod="openstack/ceilometer-0" Nov 22 08:54:06 crc kubenswrapper[4789]: I1122 08:54:06.650516 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db7c20fc-ffc4-4425-9184-5f78bb01c3b3-scripts\") pod \"ceilometer-0\" (UID: \"db7c20fc-ffc4-4425-9184-5f78bb01c3b3\") " pod="openstack/ceilometer-0" Nov 22 08:54:06 crc kubenswrapper[4789]: I1122 08:54:06.650569 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wtz7b\" (UniqueName: \"kubernetes.io/projected/db7c20fc-ffc4-4425-9184-5f78bb01c3b3-kube-api-access-wtz7b\") pod \"ceilometer-0\" (UID: \"db7c20fc-ffc4-4425-9184-5f78bb01c3b3\") " pod="openstack/ceilometer-0" Nov 22 08:54:06 crc kubenswrapper[4789]: I1122 08:54:06.657051 4789 scope.go:117] "RemoveContainer" containerID="2fb144347d1d1aad510a6ea7031cee3e906598e8fd06b81b6777431331eafb09" Nov 22 08:54:06 crc kubenswrapper[4789]: I1122 08:54:06.696634 4789 scope.go:117] "RemoveContainer" containerID="e7e7196b9d8de4d32b0220fbdd5f43d6bd1d61ff63e6f25250d131b43d76827d" Nov 22 08:54:06 crc kubenswrapper[4789]: I1122 08:54:06.725419 4789 scope.go:117] "RemoveContainer" containerID="0ba876bc146a633673594ffc7a55b1ef15f6d31b6f68e177c9d9b989cfe96170" Nov 22 08:54:06 crc kubenswrapper[4789]: I1122 08:54:06.752577 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/db7c20fc-ffc4-4425-9184-5f78bb01c3b3-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"db7c20fc-ffc4-4425-9184-5f78bb01c3b3\") " pod="openstack/ceilometer-0" Nov 22 08:54:06 crc kubenswrapper[4789]: I1122 08:54:06.757413 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db7c20fc-ffc4-4425-9184-5f78bb01c3b3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"db7c20fc-ffc4-4425-9184-5f78bb01c3b3\") " pod="openstack/ceilometer-0" Nov 22 08:54:06 crc kubenswrapper[4789]: I1122 08:54:06.757499 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db7c20fc-ffc4-4425-9184-5f78bb01c3b3-config-data\") pod \"ceilometer-0\" (UID: \"db7c20fc-ffc4-4425-9184-5f78bb01c3b3\") " pod="openstack/ceilometer-0" Nov 22 08:54:06 crc kubenswrapper[4789]: I1122 08:54:06.757580 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db7c20fc-ffc4-4425-9184-5f78bb01c3b3-scripts\") pod \"ceilometer-0\" (UID: \"db7c20fc-ffc4-4425-9184-5f78bb01c3b3\") " pod="openstack/ceilometer-0" Nov 22 08:54:06 crc kubenswrapper[4789]: I1122 08:54:06.757622 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wtz7b\" (UniqueName: \"kubernetes.io/projected/db7c20fc-ffc4-4425-9184-5f78bb01c3b3-kube-api-access-wtz7b\") pod \"ceilometer-0\" (UID: \"db7c20fc-ffc4-4425-9184-5f78bb01c3b3\") " pod="openstack/ceilometer-0" Nov 22 08:54:06 crc kubenswrapper[4789]: I1122 08:54:06.761317 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/db7c20fc-ffc4-4425-9184-5f78bb01c3b3-log-httpd\") pod \"ceilometer-0\" (UID: \"db7c20fc-ffc4-4425-9184-5f78bb01c3b3\") " pod="openstack/ceilometer-0" Nov 22 08:54:06 crc kubenswrapper[4789]: I1122 08:54:06.761451 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/db7c20fc-ffc4-4425-9184-5f78bb01c3b3-run-httpd\") pod \"ceilometer-0\" (UID: \"db7c20fc-ffc4-4425-9184-5f78bb01c3b3\") " pod="openstack/ceilometer-0" Nov 22 08:54:06 crc kubenswrapper[4789]: I1122 08:54:06.761337 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/db7c20fc-ffc4-4425-9184-5f78bb01c3b3-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"db7c20fc-ffc4-4425-9184-5f78bb01c3b3\") " pod="openstack/ceilometer-0" Nov 22 08:54:06 crc kubenswrapper[4789]: I1122 08:54:06.761855 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/db7c20fc-ffc4-4425-9184-5f78bb01c3b3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"db7c20fc-ffc4-4425-9184-5f78bb01c3b3\") " pod="openstack/ceilometer-0" Nov 22 08:54:06 crc kubenswrapper[4789]: I1122 08:54:06.761897 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/db7c20fc-ffc4-4425-9184-5f78bb01c3b3-log-httpd\") pod \"ceilometer-0\" (UID: \"db7c20fc-ffc4-4425-9184-5f78bb01c3b3\") " pod="openstack/ceilometer-0" Nov 22 08:54:06 crc kubenswrapper[4789]: I1122 08:54:06.762885 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/db7c20fc-ffc4-4425-9184-5f78bb01c3b3-run-httpd\") pod \"ceilometer-0\" (UID: \"db7c20fc-ffc4-4425-9184-5f78bb01c3b3\") " pod="openstack/ceilometer-0" Nov 22 08:54:06 crc kubenswrapper[4789]: I1122 08:54:06.765605 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db7c20fc-ffc4-4425-9184-5f78bb01c3b3-scripts\") pod \"ceilometer-0\" (UID: \"db7c20fc-ffc4-4425-9184-5f78bb01c3b3\") " pod="openstack/ceilometer-0" Nov 22 08:54:06 crc kubenswrapper[4789]: I1122 08:54:06.770485 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db7c20fc-ffc4-4425-9184-5f78bb01c3b3-config-data\") pod \"ceilometer-0\" (UID: \"db7c20fc-ffc4-4425-9184-5f78bb01c3b3\") " pod="openstack/ceilometer-0" Nov 22 08:54:06 crc kubenswrapper[4789]: I1122 08:54:06.771130 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/db7c20fc-ffc4-4425-9184-5f78bb01c3b3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"db7c20fc-ffc4-4425-9184-5f78bb01c3b3\") " pod="openstack/ceilometer-0" Nov 22 08:54:06 crc kubenswrapper[4789]: I1122 08:54:06.777256 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db7c20fc-ffc4-4425-9184-5f78bb01c3b3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"db7c20fc-ffc4-4425-9184-5f78bb01c3b3\") " pod="openstack/ceilometer-0" Nov 22 08:54:06 crc kubenswrapper[4789]: I1122 08:54:06.784714 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wtz7b\" (UniqueName: \"kubernetes.io/projected/db7c20fc-ffc4-4425-9184-5f78bb01c3b3-kube-api-access-wtz7b\") pod \"ceilometer-0\" (UID: \"db7c20fc-ffc4-4425-9184-5f78bb01c3b3\") " pod="openstack/ceilometer-0" Nov 22 08:54:06 crc kubenswrapper[4789]: I1122 08:54:06.846890 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 08:54:07 crc kubenswrapper[4789]: I1122 08:54:07.320192 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:54:07 crc kubenswrapper[4789]: I1122 08:54:07.410252 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"e21ea1ba-4437-4d76-be37-0b285909d631","Type":"ContainerStarted","Data":"b3d7c2d9fba16e75a0420bd439bdba7972e8e87814951756e8719b1fc1036bc9"} Nov 22 08:54:07 crc kubenswrapper[4789]: I1122 08:54:07.410667 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/manila-api-0" Nov 22 08:54:07 crc kubenswrapper[4789]: I1122 08:54:07.419563 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"db7c20fc-ffc4-4425-9184-5f78bb01c3b3","Type":"ContainerStarted","Data":"80a58533d0b84568f90079b945e9e9333b7340197735b4ae971445bb3f101591"} Nov 22 08:54:07 crc kubenswrapper[4789]: I1122 08:54:07.438132 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-api-0" podStartSLOduration=3.438104666 podStartE2EDuration="3.438104666s" podCreationTimestamp="2025-11-22 08:54:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:54:07.430506042 +0000 UTC m=+3461.664906315" watchObservedRunningTime="2025-11-22 08:54:07.438104666 +0000 UTC m=+3461.672504939" Nov 22 08:54:07 crc kubenswrapper[4789]: I1122 08:54:07.977524 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7785105e-179b-4b0b-9457-cd513d0b762d" path="/var/lib/kubelet/pods/7785105e-179b-4b0b-9457-cd513d0b762d/volumes" Nov 22 08:54:07 crc kubenswrapper[4789]: I1122 08:54:07.978628 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b3e0f7a4-cde7-4da6-8339-615054807f46" path="/var/lib/kubelet/pods/b3e0f7a4-cde7-4da6-8339-615054807f46/volumes" Nov 22 08:54:08 crc kubenswrapper[4789]: I1122 08:54:08.193355 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:54:09 crc kubenswrapper[4789]: I1122 08:54:09.443460 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"db7c20fc-ffc4-4425-9184-5f78bb01c3b3","Type":"ContainerStarted","Data":"6e60a0f0105608918d36f6891b8ff825867899a8a6e9c4fc3c99083ad978976b"} Nov 22 08:54:10 crc kubenswrapper[4789]: I1122 08:54:10.464742 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"db7c20fc-ffc4-4425-9184-5f78bb01c3b3","Type":"ContainerStarted","Data":"49eccd7f899b7b0468fbadaf86d3806940ba10d15f30a4bf44eb92680d8ac4a7"} Nov 22 08:54:10 crc kubenswrapper[4789]: I1122 08:54:10.465337 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"db7c20fc-ffc4-4425-9184-5f78bb01c3b3","Type":"ContainerStarted","Data":"79a6fc0183a496b5cd22a6f0fc019521e93e74fc55d0cf141e373378bd8e9bb4"} Nov 22 08:54:12 crc kubenswrapper[4789]: I1122 08:54:12.484093 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"db7c20fc-ffc4-4425-9184-5f78bb01c3b3","Type":"ContainerStarted","Data":"1d5eb7748b63ff62b9d874d14701987b832023cd7a54cb4fc6589eadce68687c"} Nov 22 08:54:12 crc kubenswrapper[4789]: I1122 08:54:12.484690 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 22 08:54:12 crc kubenswrapper[4789]: I1122 08:54:12.484308 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="db7c20fc-ffc4-4425-9184-5f78bb01c3b3" containerName="proxy-httpd" containerID="cri-o://1d5eb7748b63ff62b9d874d14701987b832023cd7a54cb4fc6589eadce68687c" gracePeriod=30 Nov 22 08:54:12 crc kubenswrapper[4789]: I1122 08:54:12.484391 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="db7c20fc-ffc4-4425-9184-5f78bb01c3b3" containerName="ceilometer-notification-agent" containerID="cri-o://79a6fc0183a496b5cd22a6f0fc019521e93e74fc55d0cf141e373378bd8e9bb4" gracePeriod=30 Nov 22 08:54:12 crc kubenswrapper[4789]: I1122 08:54:12.484409 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="db7c20fc-ffc4-4425-9184-5f78bb01c3b3" containerName="sg-core" containerID="cri-o://49eccd7f899b7b0468fbadaf86d3806940ba10d15f30a4bf44eb92680d8ac4a7" gracePeriod=30 Nov 22 08:54:12 crc kubenswrapper[4789]: I1122 08:54:12.484272 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="db7c20fc-ffc4-4425-9184-5f78bb01c3b3" containerName="ceilometer-central-agent" containerID="cri-o://6e60a0f0105608918d36f6891b8ff825867899a8a6e9c4fc3c99083ad978976b" gracePeriod=30 Nov 22 08:54:12 crc kubenswrapper[4789]: I1122 08:54:12.514494 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.123632482 podStartE2EDuration="6.514470366s" podCreationTimestamp="2025-11-22 08:54:06 +0000 UTC" firstStartedPulling="2025-11-22 08:54:07.323645055 +0000 UTC m=+3461.558045328" lastFinishedPulling="2025-11-22 08:54:11.714482939 +0000 UTC m=+3465.948883212" observedRunningTime="2025-11-22 08:54:12.511676851 +0000 UTC m=+3466.746077124" watchObservedRunningTime="2025-11-22 08:54:12.514470366 +0000 UTC m=+3466.748870639" Nov 22 08:54:13 crc kubenswrapper[4789]: I1122 08:54:13.496608 4789 generic.go:334] "Generic (PLEG): container finished" podID="db7c20fc-ffc4-4425-9184-5f78bb01c3b3" containerID="1d5eb7748b63ff62b9d874d14701987b832023cd7a54cb4fc6589eadce68687c" exitCode=0 Nov 22 08:54:13 crc kubenswrapper[4789]: I1122 08:54:13.496646 4789 generic.go:334] "Generic (PLEG): container finished" podID="db7c20fc-ffc4-4425-9184-5f78bb01c3b3" containerID="49eccd7f899b7b0468fbadaf86d3806940ba10d15f30a4bf44eb92680d8ac4a7" exitCode=2 Nov 22 08:54:13 crc kubenswrapper[4789]: I1122 08:54:13.496653 4789 generic.go:334] "Generic (PLEG): container finished" podID="db7c20fc-ffc4-4425-9184-5f78bb01c3b3" containerID="79a6fc0183a496b5cd22a6f0fc019521e93e74fc55d0cf141e373378bd8e9bb4" exitCode=0 Nov 22 08:54:13 crc kubenswrapper[4789]: I1122 08:54:13.496692 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"db7c20fc-ffc4-4425-9184-5f78bb01c3b3","Type":"ContainerDied","Data":"1d5eb7748b63ff62b9d874d14701987b832023cd7a54cb4fc6589eadce68687c"} Nov 22 08:54:13 crc kubenswrapper[4789]: I1122 08:54:13.496789 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"db7c20fc-ffc4-4425-9184-5f78bb01c3b3","Type":"ContainerDied","Data":"49eccd7f899b7b0468fbadaf86d3806940ba10d15f30a4bf44eb92680d8ac4a7"} Nov 22 08:54:13 crc kubenswrapper[4789]: I1122 08:54:13.496804 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"db7c20fc-ffc4-4425-9184-5f78bb01c3b3","Type":"ContainerDied","Data":"79a6fc0183a496b5cd22a6f0fc019521e93e74fc55d0cf141e373378bd8e9bb4"} Nov 22 08:54:14 crc kubenswrapper[4789]: I1122 08:54:14.149294 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 08:54:14 crc kubenswrapper[4789]: I1122 08:54:14.205267 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/db7c20fc-ffc4-4425-9184-5f78bb01c3b3-sg-core-conf-yaml\") pod \"db7c20fc-ffc4-4425-9184-5f78bb01c3b3\" (UID: \"db7c20fc-ffc4-4425-9184-5f78bb01c3b3\") " Nov 22 08:54:14 crc kubenswrapper[4789]: I1122 08:54:14.206002 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/db7c20fc-ffc4-4425-9184-5f78bb01c3b3-log-httpd\") pod \"db7c20fc-ffc4-4425-9184-5f78bb01c3b3\" (UID: \"db7c20fc-ffc4-4425-9184-5f78bb01c3b3\") " Nov 22 08:54:14 crc kubenswrapper[4789]: I1122 08:54:14.206049 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db7c20fc-ffc4-4425-9184-5f78bb01c3b3-scripts\") pod \"db7c20fc-ffc4-4425-9184-5f78bb01c3b3\" (UID: \"db7c20fc-ffc4-4425-9184-5f78bb01c3b3\") " Nov 22 08:54:14 crc kubenswrapper[4789]: I1122 08:54:14.206089 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/db7c20fc-ffc4-4425-9184-5f78bb01c3b3-run-httpd\") pod \"db7c20fc-ffc4-4425-9184-5f78bb01c3b3\" (UID: \"db7c20fc-ffc4-4425-9184-5f78bb01c3b3\") " Nov 22 08:54:14 crc kubenswrapper[4789]: I1122 08:54:14.206213 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db7c20fc-ffc4-4425-9184-5f78bb01c3b3-config-data\") pod \"db7c20fc-ffc4-4425-9184-5f78bb01c3b3\" (UID: \"db7c20fc-ffc4-4425-9184-5f78bb01c3b3\") " Nov 22 08:54:14 crc kubenswrapper[4789]: I1122 08:54:14.206267 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/db7c20fc-ffc4-4425-9184-5f78bb01c3b3-ceilometer-tls-certs\") pod \"db7c20fc-ffc4-4425-9184-5f78bb01c3b3\" (UID: \"db7c20fc-ffc4-4425-9184-5f78bb01c3b3\") " Nov 22 08:54:14 crc kubenswrapper[4789]: I1122 08:54:14.206436 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/db7c20fc-ffc4-4425-9184-5f78bb01c3b3-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "db7c20fc-ffc4-4425-9184-5f78bb01c3b3" (UID: "db7c20fc-ffc4-4425-9184-5f78bb01c3b3"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:54:14 crc kubenswrapper[4789]: I1122 08:54:14.206627 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/db7c20fc-ffc4-4425-9184-5f78bb01c3b3-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "db7c20fc-ffc4-4425-9184-5f78bb01c3b3" (UID: "db7c20fc-ffc4-4425-9184-5f78bb01c3b3"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:54:14 crc kubenswrapper[4789]: I1122 08:54:14.206706 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db7c20fc-ffc4-4425-9184-5f78bb01c3b3-combined-ca-bundle\") pod \"db7c20fc-ffc4-4425-9184-5f78bb01c3b3\" (UID: \"db7c20fc-ffc4-4425-9184-5f78bb01c3b3\") " Nov 22 08:54:14 crc kubenswrapper[4789]: I1122 08:54:14.206743 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wtz7b\" (UniqueName: \"kubernetes.io/projected/db7c20fc-ffc4-4425-9184-5f78bb01c3b3-kube-api-access-wtz7b\") pod \"db7c20fc-ffc4-4425-9184-5f78bb01c3b3\" (UID: \"db7c20fc-ffc4-4425-9184-5f78bb01c3b3\") " Nov 22 08:54:14 crc kubenswrapper[4789]: I1122 08:54:14.207189 4789 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/db7c20fc-ffc4-4425-9184-5f78bb01c3b3-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 22 08:54:14 crc kubenswrapper[4789]: I1122 08:54:14.207205 4789 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/db7c20fc-ffc4-4425-9184-5f78bb01c3b3-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 22 08:54:14 crc kubenswrapper[4789]: I1122 08:54:14.219445 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db7c20fc-ffc4-4425-9184-5f78bb01c3b3-scripts" (OuterVolumeSpecName: "scripts") pod "db7c20fc-ffc4-4425-9184-5f78bb01c3b3" (UID: "db7c20fc-ffc4-4425-9184-5f78bb01c3b3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:54:14 crc kubenswrapper[4789]: I1122 08:54:14.219953 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db7c20fc-ffc4-4425-9184-5f78bb01c3b3-kube-api-access-wtz7b" (OuterVolumeSpecName: "kube-api-access-wtz7b") pod "db7c20fc-ffc4-4425-9184-5f78bb01c3b3" (UID: "db7c20fc-ffc4-4425-9184-5f78bb01c3b3"). InnerVolumeSpecName "kube-api-access-wtz7b". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:54:14 crc kubenswrapper[4789]: I1122 08:54:14.241443 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db7c20fc-ffc4-4425-9184-5f78bb01c3b3-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "db7c20fc-ffc4-4425-9184-5f78bb01c3b3" (UID: "db7c20fc-ffc4-4425-9184-5f78bb01c3b3"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:54:14 crc kubenswrapper[4789]: I1122 08:54:14.266312 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db7c20fc-ffc4-4425-9184-5f78bb01c3b3-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "db7c20fc-ffc4-4425-9184-5f78bb01c3b3" (UID: "db7c20fc-ffc4-4425-9184-5f78bb01c3b3"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:54:14 crc kubenswrapper[4789]: I1122 08:54:14.285816 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db7c20fc-ffc4-4425-9184-5f78bb01c3b3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "db7c20fc-ffc4-4425-9184-5f78bb01c3b3" (UID: "db7c20fc-ffc4-4425-9184-5f78bb01c3b3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:54:14 crc kubenswrapper[4789]: I1122 08:54:14.309557 4789 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/db7c20fc-ffc4-4425-9184-5f78bb01c3b3-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 22 08:54:14 crc kubenswrapper[4789]: I1122 08:54:14.309599 4789 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db7c20fc-ffc4-4425-9184-5f78bb01c3b3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:54:14 crc kubenswrapper[4789]: I1122 08:54:14.309612 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wtz7b\" (UniqueName: \"kubernetes.io/projected/db7c20fc-ffc4-4425-9184-5f78bb01c3b3-kube-api-access-wtz7b\") on node \"crc\" DevicePath \"\"" Nov 22 08:54:14 crc kubenswrapper[4789]: I1122 08:54:14.309625 4789 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/db7c20fc-ffc4-4425-9184-5f78bb01c3b3-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 22 08:54:14 crc kubenswrapper[4789]: I1122 08:54:14.309636 4789 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db7c20fc-ffc4-4425-9184-5f78bb01c3b3-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:54:14 crc kubenswrapper[4789]: I1122 08:54:14.310165 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db7c20fc-ffc4-4425-9184-5f78bb01c3b3-config-data" (OuterVolumeSpecName: "config-data") pod "db7c20fc-ffc4-4425-9184-5f78bb01c3b3" (UID: "db7c20fc-ffc4-4425-9184-5f78bb01c3b3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:54:14 crc kubenswrapper[4789]: I1122 08:54:14.411204 4789 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db7c20fc-ffc4-4425-9184-5f78bb01c3b3-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:54:14 crc kubenswrapper[4789]: I1122 08:54:14.509781 4789 generic.go:334] "Generic (PLEG): container finished" podID="db7c20fc-ffc4-4425-9184-5f78bb01c3b3" containerID="6e60a0f0105608918d36f6891b8ff825867899a8a6e9c4fc3c99083ad978976b" exitCode=0 Nov 22 08:54:14 crc kubenswrapper[4789]: I1122 08:54:14.509860 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"db7c20fc-ffc4-4425-9184-5f78bb01c3b3","Type":"ContainerDied","Data":"6e60a0f0105608918d36f6891b8ff825867899a8a6e9c4fc3c99083ad978976b"} Nov 22 08:54:14 crc kubenswrapper[4789]: I1122 08:54:14.509895 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"db7c20fc-ffc4-4425-9184-5f78bb01c3b3","Type":"ContainerDied","Data":"80a58533d0b84568f90079b945e9e9333b7340197735b4ae971445bb3f101591"} Nov 22 08:54:14 crc kubenswrapper[4789]: I1122 08:54:14.509933 4789 scope.go:117] "RemoveContainer" containerID="1d5eb7748b63ff62b9d874d14701987b832023cd7a54cb4fc6589eadce68687c" Nov 22 08:54:14 crc kubenswrapper[4789]: I1122 08:54:14.510760 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 08:54:14 crc kubenswrapper[4789]: I1122 08:54:14.548211 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:54:14 crc kubenswrapper[4789]: I1122 08:54:14.549131 4789 scope.go:117] "RemoveContainer" containerID="49eccd7f899b7b0468fbadaf86d3806940ba10d15f30a4bf44eb92680d8ac4a7" Nov 22 08:54:14 crc kubenswrapper[4789]: I1122 08:54:14.563298 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:54:14 crc kubenswrapper[4789]: I1122 08:54:14.605480 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:54:14 crc kubenswrapper[4789]: I1122 08:54:14.608902 4789 scope.go:117] "RemoveContainer" containerID="79a6fc0183a496b5cd22a6f0fc019521e93e74fc55d0cf141e373378bd8e9bb4" Nov 22 08:54:14 crc kubenswrapper[4789]: E1122 08:54:14.609081 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db7c20fc-ffc4-4425-9184-5f78bb01c3b3" containerName="sg-core" Nov 22 08:54:14 crc kubenswrapper[4789]: I1122 08:54:14.609123 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="db7c20fc-ffc4-4425-9184-5f78bb01c3b3" containerName="sg-core" Nov 22 08:54:14 crc kubenswrapper[4789]: E1122 08:54:14.609786 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db7c20fc-ffc4-4425-9184-5f78bb01c3b3" containerName="ceilometer-notification-agent" Nov 22 08:54:14 crc kubenswrapper[4789]: I1122 08:54:14.609819 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="db7c20fc-ffc4-4425-9184-5f78bb01c3b3" containerName="ceilometer-notification-agent" Nov 22 08:54:14 crc kubenswrapper[4789]: E1122 08:54:14.609896 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db7c20fc-ffc4-4425-9184-5f78bb01c3b3" containerName="ceilometer-central-agent" Nov 22 08:54:14 crc kubenswrapper[4789]: I1122 08:54:14.609919 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="db7c20fc-ffc4-4425-9184-5f78bb01c3b3" containerName="ceilometer-central-agent" Nov 22 08:54:14 crc kubenswrapper[4789]: E1122 08:54:14.610010 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db7c20fc-ffc4-4425-9184-5f78bb01c3b3" containerName="proxy-httpd" Nov 22 08:54:14 crc kubenswrapper[4789]: I1122 08:54:14.610020 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="db7c20fc-ffc4-4425-9184-5f78bb01c3b3" containerName="proxy-httpd" Nov 22 08:54:14 crc kubenswrapper[4789]: I1122 08:54:14.612299 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="db7c20fc-ffc4-4425-9184-5f78bb01c3b3" containerName="ceilometer-central-agent" Nov 22 08:54:14 crc kubenswrapper[4789]: I1122 08:54:14.612395 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="db7c20fc-ffc4-4425-9184-5f78bb01c3b3" containerName="ceilometer-notification-agent" Nov 22 08:54:14 crc kubenswrapper[4789]: I1122 08:54:14.612414 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="db7c20fc-ffc4-4425-9184-5f78bb01c3b3" containerName="proxy-httpd" Nov 22 08:54:14 crc kubenswrapper[4789]: I1122 08:54:14.612443 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="db7c20fc-ffc4-4425-9184-5f78bb01c3b3" containerName="sg-core" Nov 22 08:54:14 crc kubenswrapper[4789]: I1122 08:54:14.616614 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 08:54:14 crc kubenswrapper[4789]: I1122 08:54:14.619214 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 22 08:54:14 crc kubenswrapper[4789]: I1122 08:54:14.619350 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 22 08:54:14 crc kubenswrapper[4789]: I1122 08:54:14.619924 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 22 08:54:14 crc kubenswrapper[4789]: I1122 08:54:14.624415 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:54:14 crc kubenswrapper[4789]: I1122 08:54:14.630666 4789 scope.go:117] "RemoveContainer" containerID="6e60a0f0105608918d36f6891b8ff825867899a8a6e9c4fc3c99083ad978976b" Nov 22 08:54:14 crc kubenswrapper[4789]: I1122 08:54:14.650917 4789 scope.go:117] "RemoveContainer" containerID="1d5eb7748b63ff62b9d874d14701987b832023cd7a54cb4fc6589eadce68687c" Nov 22 08:54:14 crc kubenswrapper[4789]: E1122 08:54:14.651314 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d5eb7748b63ff62b9d874d14701987b832023cd7a54cb4fc6589eadce68687c\": container with ID starting with 1d5eb7748b63ff62b9d874d14701987b832023cd7a54cb4fc6589eadce68687c not found: ID does not exist" containerID="1d5eb7748b63ff62b9d874d14701987b832023cd7a54cb4fc6589eadce68687c" Nov 22 08:54:14 crc kubenswrapper[4789]: I1122 08:54:14.651347 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d5eb7748b63ff62b9d874d14701987b832023cd7a54cb4fc6589eadce68687c"} err="failed to get container status \"1d5eb7748b63ff62b9d874d14701987b832023cd7a54cb4fc6589eadce68687c\": rpc error: code = NotFound desc = could not find container \"1d5eb7748b63ff62b9d874d14701987b832023cd7a54cb4fc6589eadce68687c\": container with ID starting with 1d5eb7748b63ff62b9d874d14701987b832023cd7a54cb4fc6589eadce68687c not found: ID does not exist" Nov 22 08:54:14 crc kubenswrapper[4789]: I1122 08:54:14.651367 4789 scope.go:117] "RemoveContainer" containerID="49eccd7f899b7b0468fbadaf86d3806940ba10d15f30a4bf44eb92680d8ac4a7" Nov 22 08:54:14 crc kubenswrapper[4789]: E1122 08:54:14.651661 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"49eccd7f899b7b0468fbadaf86d3806940ba10d15f30a4bf44eb92680d8ac4a7\": container with ID starting with 49eccd7f899b7b0468fbadaf86d3806940ba10d15f30a4bf44eb92680d8ac4a7 not found: ID does not exist" containerID="49eccd7f899b7b0468fbadaf86d3806940ba10d15f30a4bf44eb92680d8ac4a7" Nov 22 08:54:14 crc kubenswrapper[4789]: I1122 08:54:14.651771 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"49eccd7f899b7b0468fbadaf86d3806940ba10d15f30a4bf44eb92680d8ac4a7"} err="failed to get container status \"49eccd7f899b7b0468fbadaf86d3806940ba10d15f30a4bf44eb92680d8ac4a7\": rpc error: code = NotFound desc = could not find container \"49eccd7f899b7b0468fbadaf86d3806940ba10d15f30a4bf44eb92680d8ac4a7\": container with ID starting with 49eccd7f899b7b0468fbadaf86d3806940ba10d15f30a4bf44eb92680d8ac4a7 not found: ID does not exist" Nov 22 08:54:14 crc kubenswrapper[4789]: I1122 08:54:14.651874 4789 scope.go:117] "RemoveContainer" containerID="79a6fc0183a496b5cd22a6f0fc019521e93e74fc55d0cf141e373378bd8e9bb4" Nov 22 08:54:14 crc kubenswrapper[4789]: E1122 08:54:14.652248 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"79a6fc0183a496b5cd22a6f0fc019521e93e74fc55d0cf141e373378bd8e9bb4\": container with ID starting with 79a6fc0183a496b5cd22a6f0fc019521e93e74fc55d0cf141e373378bd8e9bb4 not found: ID does not exist" containerID="79a6fc0183a496b5cd22a6f0fc019521e93e74fc55d0cf141e373378bd8e9bb4" Nov 22 08:54:14 crc kubenswrapper[4789]: I1122 08:54:14.652272 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"79a6fc0183a496b5cd22a6f0fc019521e93e74fc55d0cf141e373378bd8e9bb4"} err="failed to get container status \"79a6fc0183a496b5cd22a6f0fc019521e93e74fc55d0cf141e373378bd8e9bb4\": rpc error: code = NotFound desc = could not find container \"79a6fc0183a496b5cd22a6f0fc019521e93e74fc55d0cf141e373378bd8e9bb4\": container with ID starting with 79a6fc0183a496b5cd22a6f0fc019521e93e74fc55d0cf141e373378bd8e9bb4 not found: ID does not exist" Nov 22 08:54:14 crc kubenswrapper[4789]: I1122 08:54:14.652291 4789 scope.go:117] "RemoveContainer" containerID="6e60a0f0105608918d36f6891b8ff825867899a8a6e9c4fc3c99083ad978976b" Nov 22 08:54:14 crc kubenswrapper[4789]: E1122 08:54:14.652654 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6e60a0f0105608918d36f6891b8ff825867899a8a6e9c4fc3c99083ad978976b\": container with ID starting with 6e60a0f0105608918d36f6891b8ff825867899a8a6e9c4fc3c99083ad978976b not found: ID does not exist" containerID="6e60a0f0105608918d36f6891b8ff825867899a8a6e9c4fc3c99083ad978976b" Nov 22 08:54:14 crc kubenswrapper[4789]: I1122 08:54:14.652781 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e60a0f0105608918d36f6891b8ff825867899a8a6e9c4fc3c99083ad978976b"} err="failed to get container status \"6e60a0f0105608918d36f6891b8ff825867899a8a6e9c4fc3c99083ad978976b\": rpc error: code = NotFound desc = could not find container \"6e60a0f0105608918d36f6891b8ff825867899a8a6e9c4fc3c99083ad978976b\": container with ID starting with 6e60a0f0105608918d36f6891b8ff825867899a8a6e9c4fc3c99083ad978976b not found: ID does not exist" Nov 22 08:54:14 crc kubenswrapper[4789]: I1122 08:54:14.717301 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/deee1bbe-0f19-4f57-bebc-de5b01ccac7f-scripts\") pod \"ceilometer-0\" (UID: \"deee1bbe-0f19-4f57-bebc-de5b01ccac7f\") " pod="openstack/ceilometer-0" Nov 22 08:54:14 crc kubenswrapper[4789]: I1122 08:54:14.717355 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/deee1bbe-0f19-4f57-bebc-de5b01ccac7f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"deee1bbe-0f19-4f57-bebc-de5b01ccac7f\") " pod="openstack/ceilometer-0" Nov 22 08:54:14 crc kubenswrapper[4789]: I1122 08:54:14.717375 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/deee1bbe-0f19-4f57-bebc-de5b01ccac7f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"deee1bbe-0f19-4f57-bebc-de5b01ccac7f\") " pod="openstack/ceilometer-0" Nov 22 08:54:14 crc kubenswrapper[4789]: I1122 08:54:14.717402 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/deee1bbe-0f19-4f57-bebc-de5b01ccac7f-config-data\") pod \"ceilometer-0\" (UID: \"deee1bbe-0f19-4f57-bebc-de5b01ccac7f\") " pod="openstack/ceilometer-0" Nov 22 08:54:14 crc kubenswrapper[4789]: I1122 08:54:14.717444 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d5fql\" (UniqueName: \"kubernetes.io/projected/deee1bbe-0f19-4f57-bebc-de5b01ccac7f-kube-api-access-d5fql\") pod \"ceilometer-0\" (UID: \"deee1bbe-0f19-4f57-bebc-de5b01ccac7f\") " pod="openstack/ceilometer-0" Nov 22 08:54:14 crc kubenswrapper[4789]: I1122 08:54:14.717479 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/deee1bbe-0f19-4f57-bebc-de5b01ccac7f-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"deee1bbe-0f19-4f57-bebc-de5b01ccac7f\") " pod="openstack/ceilometer-0" Nov 22 08:54:14 crc kubenswrapper[4789]: I1122 08:54:14.717550 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/deee1bbe-0f19-4f57-bebc-de5b01ccac7f-run-httpd\") pod \"ceilometer-0\" (UID: \"deee1bbe-0f19-4f57-bebc-de5b01ccac7f\") " pod="openstack/ceilometer-0" Nov 22 08:54:14 crc kubenswrapper[4789]: I1122 08:54:14.717567 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/deee1bbe-0f19-4f57-bebc-de5b01ccac7f-log-httpd\") pod \"ceilometer-0\" (UID: \"deee1bbe-0f19-4f57-bebc-de5b01ccac7f\") " pod="openstack/ceilometer-0" Nov 22 08:54:14 crc kubenswrapper[4789]: I1122 08:54:14.819343 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/deee1bbe-0f19-4f57-bebc-de5b01ccac7f-config-data\") pod \"ceilometer-0\" (UID: \"deee1bbe-0f19-4f57-bebc-de5b01ccac7f\") " pod="openstack/ceilometer-0" Nov 22 08:54:14 crc kubenswrapper[4789]: I1122 08:54:14.820401 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d5fql\" (UniqueName: \"kubernetes.io/projected/deee1bbe-0f19-4f57-bebc-de5b01ccac7f-kube-api-access-d5fql\") pod \"ceilometer-0\" (UID: \"deee1bbe-0f19-4f57-bebc-de5b01ccac7f\") " pod="openstack/ceilometer-0" Nov 22 08:54:14 crc kubenswrapper[4789]: I1122 08:54:14.820451 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/deee1bbe-0f19-4f57-bebc-de5b01ccac7f-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"deee1bbe-0f19-4f57-bebc-de5b01ccac7f\") " pod="openstack/ceilometer-0" Nov 22 08:54:14 crc kubenswrapper[4789]: I1122 08:54:14.820569 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/deee1bbe-0f19-4f57-bebc-de5b01ccac7f-run-httpd\") pod \"ceilometer-0\" (UID: \"deee1bbe-0f19-4f57-bebc-de5b01ccac7f\") " pod="openstack/ceilometer-0" Nov 22 08:54:14 crc kubenswrapper[4789]: I1122 08:54:14.820673 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/deee1bbe-0f19-4f57-bebc-de5b01ccac7f-log-httpd\") pod \"ceilometer-0\" (UID: \"deee1bbe-0f19-4f57-bebc-de5b01ccac7f\") " pod="openstack/ceilometer-0" Nov 22 08:54:14 crc kubenswrapper[4789]: I1122 08:54:14.820726 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/deee1bbe-0f19-4f57-bebc-de5b01ccac7f-scripts\") pod \"ceilometer-0\" (UID: \"deee1bbe-0f19-4f57-bebc-de5b01ccac7f\") " pod="openstack/ceilometer-0" Nov 22 08:54:14 crc kubenswrapper[4789]: I1122 08:54:14.820818 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/deee1bbe-0f19-4f57-bebc-de5b01ccac7f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"deee1bbe-0f19-4f57-bebc-de5b01ccac7f\") " pod="openstack/ceilometer-0" Nov 22 08:54:14 crc kubenswrapper[4789]: I1122 08:54:14.821067 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/deee1bbe-0f19-4f57-bebc-de5b01ccac7f-run-httpd\") pod \"ceilometer-0\" (UID: \"deee1bbe-0f19-4f57-bebc-de5b01ccac7f\") " pod="openstack/ceilometer-0" Nov 22 08:54:14 crc kubenswrapper[4789]: I1122 08:54:14.821239 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/deee1bbe-0f19-4f57-bebc-de5b01ccac7f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"deee1bbe-0f19-4f57-bebc-de5b01ccac7f\") " pod="openstack/ceilometer-0" Nov 22 08:54:14 crc kubenswrapper[4789]: I1122 08:54:14.821329 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/deee1bbe-0f19-4f57-bebc-de5b01ccac7f-log-httpd\") pod \"ceilometer-0\" (UID: \"deee1bbe-0f19-4f57-bebc-de5b01ccac7f\") " pod="openstack/ceilometer-0" Nov 22 08:54:14 crc kubenswrapper[4789]: I1122 08:54:14.825698 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/deee1bbe-0f19-4f57-bebc-de5b01ccac7f-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"deee1bbe-0f19-4f57-bebc-de5b01ccac7f\") " pod="openstack/ceilometer-0" Nov 22 08:54:14 crc kubenswrapper[4789]: I1122 08:54:14.826183 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/deee1bbe-0f19-4f57-bebc-de5b01ccac7f-config-data\") pod \"ceilometer-0\" (UID: \"deee1bbe-0f19-4f57-bebc-de5b01ccac7f\") " pod="openstack/ceilometer-0" Nov 22 08:54:14 crc kubenswrapper[4789]: I1122 08:54:14.826237 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/deee1bbe-0f19-4f57-bebc-de5b01ccac7f-scripts\") pod \"ceilometer-0\" (UID: \"deee1bbe-0f19-4f57-bebc-de5b01ccac7f\") " pod="openstack/ceilometer-0" Nov 22 08:54:14 crc kubenswrapper[4789]: I1122 08:54:14.826954 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/deee1bbe-0f19-4f57-bebc-de5b01ccac7f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"deee1bbe-0f19-4f57-bebc-de5b01ccac7f\") " pod="openstack/ceilometer-0" Nov 22 08:54:14 crc kubenswrapper[4789]: I1122 08:54:14.827445 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/deee1bbe-0f19-4f57-bebc-de5b01ccac7f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"deee1bbe-0f19-4f57-bebc-de5b01ccac7f\") " pod="openstack/ceilometer-0" Nov 22 08:54:14 crc kubenswrapper[4789]: I1122 08:54:14.839553 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d5fql\" (UniqueName: \"kubernetes.io/projected/deee1bbe-0f19-4f57-bebc-de5b01ccac7f-kube-api-access-d5fql\") pod \"ceilometer-0\" (UID: \"deee1bbe-0f19-4f57-bebc-de5b01ccac7f\") " pod="openstack/ceilometer-0" Nov 22 08:54:14 crc kubenswrapper[4789]: I1122 08:54:14.894864 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-share-share1-0" Nov 22 08:54:14 crc kubenswrapper[4789]: I1122 08:54:14.934889 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 22 08:54:14 crc kubenswrapper[4789]: I1122 08:54:14.935359 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-scheduler-0" Nov 22 08:54:15 crc kubenswrapper[4789]: I1122 08:54:15.381477 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 22 08:54:15 crc kubenswrapper[4789]: W1122 08:54:15.389204 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddeee1bbe_0f19_4f57_bebc_de5b01ccac7f.slice/crio-f8528560d0da9168179edcb623525e3ef40fe757601d044a298a26c25b23088f WatchSource:0}: Error finding container f8528560d0da9168179edcb623525e3ef40fe757601d044a298a26c25b23088f: Status 404 returned error can't find the container with id f8528560d0da9168179edcb623525e3ef40fe757601d044a298a26c25b23088f Nov 22 08:54:15 crc kubenswrapper[4789]: I1122 08:54:15.521835 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"deee1bbe-0f19-4f57-bebc-de5b01ccac7f","Type":"ContainerStarted","Data":"f8528560d0da9168179edcb623525e3ef40fe757601d044a298a26c25b23088f"} Nov 22 08:54:15 crc kubenswrapper[4789]: I1122 08:54:15.978067 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db7c20fc-ffc4-4425-9184-5f78bb01c3b3" path="/var/lib/kubelet/pods/db7c20fc-ffc4-4425-9184-5f78bb01c3b3/volumes" Nov 22 08:54:16 crc kubenswrapper[4789]: I1122 08:54:16.495032 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-share-share1-0" Nov 22 08:54:16 crc kubenswrapper[4789]: I1122 08:54:16.510652 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-scheduler-0" Nov 22 08:54:16 crc kubenswrapper[4789]: I1122 08:54:16.554715 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"deee1bbe-0f19-4f57-bebc-de5b01ccac7f","Type":"ContainerStarted","Data":"156c27da864eebfb12372068b839ebbd34f238ba528e1e5a41b839fc5378b2a2"} Nov 22 08:54:16 crc kubenswrapper[4789]: I1122 08:54:16.566597 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-share-share1-0"] Nov 22 08:54:16 crc kubenswrapper[4789]: I1122 08:54:16.567074 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-share-share1-0" podUID="3b679ee8-b6ac-4313-ac79-abd465d1b0c4" containerName="manila-share" containerID="cri-o://f79b8f35c30c6903423dfa326fc73e1e524773a5acfa2c6d1ab55bb96502e49b" gracePeriod=30 Nov 22 08:54:16 crc kubenswrapper[4789]: I1122 08:54:16.567333 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-share-share1-0" podUID="3b679ee8-b6ac-4313-ac79-abd465d1b0c4" containerName="probe" containerID="cri-o://18527ce10ae16f3927ffa1f263ae903cc74829ed9695fa52e8ea021020d91dbc" gracePeriod=30 Nov 22 08:54:16 crc kubenswrapper[4789]: I1122 08:54:16.591736 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-scheduler-0"] Nov 22 08:54:16 crc kubenswrapper[4789]: I1122 08:54:16.591985 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-scheduler-0" podUID="46af0a4f-fe66-47b9-b9d9-892c3eb3d97c" containerName="manila-scheduler" containerID="cri-o://49d5d401ae263afccfb4c1717363fd6219980a592b2a9754a395aa281e5faabd" gracePeriod=30 Nov 22 08:54:16 crc kubenswrapper[4789]: I1122 08:54:16.592456 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-scheduler-0" podUID="46af0a4f-fe66-47b9-b9d9-892c3eb3d97c" containerName="probe" containerID="cri-o://b6d59c915dfda3979f0214cccd7333388cee6ab9e25416acb131775f14556d36" gracePeriod=30 Nov 22 08:54:17 crc kubenswrapper[4789]: I1122 08:54:17.570434 4789 generic.go:334] "Generic (PLEG): container finished" podID="3b679ee8-b6ac-4313-ac79-abd465d1b0c4" containerID="18527ce10ae16f3927ffa1f263ae903cc74829ed9695fa52e8ea021020d91dbc" exitCode=0 Nov 22 08:54:17 crc kubenswrapper[4789]: I1122 08:54:17.571028 4789 generic.go:334] "Generic (PLEG): container finished" podID="3b679ee8-b6ac-4313-ac79-abd465d1b0c4" containerID="f79b8f35c30c6903423dfa326fc73e1e524773a5acfa2c6d1ab55bb96502e49b" exitCode=1 Nov 22 08:54:17 crc kubenswrapper[4789]: I1122 08:54:17.570551 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"3b679ee8-b6ac-4313-ac79-abd465d1b0c4","Type":"ContainerDied","Data":"18527ce10ae16f3927ffa1f263ae903cc74829ed9695fa52e8ea021020d91dbc"} Nov 22 08:54:17 crc kubenswrapper[4789]: I1122 08:54:17.571116 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"3b679ee8-b6ac-4313-ac79-abd465d1b0c4","Type":"ContainerDied","Data":"f79b8f35c30c6903423dfa326fc73e1e524773a5acfa2c6d1ab55bb96502e49b"} Nov 22 08:54:17 crc kubenswrapper[4789]: I1122 08:54:17.571136 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"3b679ee8-b6ac-4313-ac79-abd465d1b0c4","Type":"ContainerDied","Data":"f41e7690841252fa485a3354fe7c66876a448965126fb4c0dadf03376c875ff6"} Nov 22 08:54:17 crc kubenswrapper[4789]: I1122 08:54:17.571152 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f41e7690841252fa485a3354fe7c66876a448965126fb4c0dadf03376c875ff6" Nov 22 08:54:17 crc kubenswrapper[4789]: I1122 08:54:17.578568 4789 generic.go:334] "Generic (PLEG): container finished" podID="46af0a4f-fe66-47b9-b9d9-892c3eb3d97c" containerID="b6d59c915dfda3979f0214cccd7333388cee6ab9e25416acb131775f14556d36" exitCode=0 Nov 22 08:54:17 crc kubenswrapper[4789]: I1122 08:54:17.578662 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"46af0a4f-fe66-47b9-b9d9-892c3eb3d97c","Type":"ContainerDied","Data":"b6d59c915dfda3979f0214cccd7333388cee6ab9e25416acb131775f14556d36"} Nov 22 08:54:17 crc kubenswrapper[4789]: I1122 08:54:17.585532 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"deee1bbe-0f19-4f57-bebc-de5b01ccac7f","Type":"ContainerStarted","Data":"d91fed2d49aca22771581330c602d4b11f5478f94e48aced3f2d893caf129941"} Nov 22 08:54:17 crc kubenswrapper[4789]: I1122 08:54:17.651384 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Nov 22 08:54:17 crc kubenswrapper[4789]: I1122 08:54:17.707080 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b679ee8-b6ac-4313-ac79-abd465d1b0c4-combined-ca-bundle\") pod \"3b679ee8-b6ac-4313-ac79-abd465d1b0c4\" (UID: \"3b679ee8-b6ac-4313-ac79-abd465d1b0c4\") " Nov 22 08:54:17 crc kubenswrapper[4789]: I1122 08:54:17.707827 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3b679ee8-b6ac-4313-ac79-abd465d1b0c4-etc-machine-id\") pod \"3b679ee8-b6ac-4313-ac79-abd465d1b0c4\" (UID: \"3b679ee8-b6ac-4313-ac79-abd465d1b0c4\") " Nov 22 08:54:17 crc kubenswrapper[4789]: I1122 08:54:17.707887 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h647z\" (UniqueName: \"kubernetes.io/projected/3b679ee8-b6ac-4313-ac79-abd465d1b0c4-kube-api-access-h647z\") pod \"3b679ee8-b6ac-4313-ac79-abd465d1b0c4\" (UID: \"3b679ee8-b6ac-4313-ac79-abd465d1b0c4\") " Nov 22 08:54:17 crc kubenswrapper[4789]: I1122 08:54:17.707989 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3b679ee8-b6ac-4313-ac79-abd465d1b0c4-config-data-custom\") pod \"3b679ee8-b6ac-4313-ac79-abd465d1b0c4\" (UID: \"3b679ee8-b6ac-4313-ac79-abd465d1b0c4\") " Nov 22 08:54:17 crc kubenswrapper[4789]: I1122 08:54:17.708157 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b679ee8-b6ac-4313-ac79-abd465d1b0c4-config-data\") pod \"3b679ee8-b6ac-4313-ac79-abd465d1b0c4\" (UID: \"3b679ee8-b6ac-4313-ac79-abd465d1b0c4\") " Nov 22 08:54:17 crc kubenswrapper[4789]: I1122 08:54:17.708236 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/3b679ee8-b6ac-4313-ac79-abd465d1b0c4-ceph\") pod \"3b679ee8-b6ac-4313-ac79-abd465d1b0c4\" (UID: \"3b679ee8-b6ac-4313-ac79-abd465d1b0c4\") " Nov 22 08:54:17 crc kubenswrapper[4789]: I1122 08:54:17.708349 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3b679ee8-b6ac-4313-ac79-abd465d1b0c4-scripts\") pod \"3b679ee8-b6ac-4313-ac79-abd465d1b0c4\" (UID: \"3b679ee8-b6ac-4313-ac79-abd465d1b0c4\") " Nov 22 08:54:17 crc kubenswrapper[4789]: I1122 08:54:17.708402 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/3b679ee8-b6ac-4313-ac79-abd465d1b0c4-var-lib-manila\") pod \"3b679ee8-b6ac-4313-ac79-abd465d1b0c4\" (UID: \"3b679ee8-b6ac-4313-ac79-abd465d1b0c4\") " Nov 22 08:54:17 crc kubenswrapper[4789]: I1122 08:54:17.717011 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b679ee8-b6ac-4313-ac79-abd465d1b0c4-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "3b679ee8-b6ac-4313-ac79-abd465d1b0c4" (UID: "3b679ee8-b6ac-4313-ac79-abd465d1b0c4"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:54:17 crc kubenswrapper[4789]: I1122 08:54:17.717200 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3b679ee8-b6ac-4313-ac79-abd465d1b0c4-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "3b679ee8-b6ac-4313-ac79-abd465d1b0c4" (UID: "3b679ee8-b6ac-4313-ac79-abd465d1b0c4"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 08:54:17 crc kubenswrapper[4789]: I1122 08:54:17.721303 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b679ee8-b6ac-4313-ac79-abd465d1b0c4-kube-api-access-h647z" (OuterVolumeSpecName: "kube-api-access-h647z") pod "3b679ee8-b6ac-4313-ac79-abd465d1b0c4" (UID: "3b679ee8-b6ac-4313-ac79-abd465d1b0c4"). InnerVolumeSpecName "kube-api-access-h647z". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:54:17 crc kubenswrapper[4789]: I1122 08:54:17.721927 4789 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3b679ee8-b6ac-4313-ac79-abd465d1b0c4-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 22 08:54:17 crc kubenswrapper[4789]: I1122 08:54:17.721956 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h647z\" (UniqueName: \"kubernetes.io/projected/3b679ee8-b6ac-4313-ac79-abd465d1b0c4-kube-api-access-h647z\") on node \"crc\" DevicePath \"\"" Nov 22 08:54:17 crc kubenswrapper[4789]: I1122 08:54:17.721994 4789 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3b679ee8-b6ac-4313-ac79-abd465d1b0c4-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 22 08:54:17 crc kubenswrapper[4789]: I1122 08:54:17.723702 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3b679ee8-b6ac-4313-ac79-abd465d1b0c4-var-lib-manila" (OuterVolumeSpecName: "var-lib-manila") pod "3b679ee8-b6ac-4313-ac79-abd465d1b0c4" (UID: "3b679ee8-b6ac-4313-ac79-abd465d1b0c4"). InnerVolumeSpecName "var-lib-manila". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 08:54:17 crc kubenswrapper[4789]: I1122 08:54:17.728114 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b679ee8-b6ac-4313-ac79-abd465d1b0c4-scripts" (OuterVolumeSpecName: "scripts") pod "3b679ee8-b6ac-4313-ac79-abd465d1b0c4" (UID: "3b679ee8-b6ac-4313-ac79-abd465d1b0c4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:54:17 crc kubenswrapper[4789]: I1122 08:54:17.737717 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b679ee8-b6ac-4313-ac79-abd465d1b0c4-ceph" (OuterVolumeSpecName: "ceph") pod "3b679ee8-b6ac-4313-ac79-abd465d1b0c4" (UID: "3b679ee8-b6ac-4313-ac79-abd465d1b0c4"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:54:17 crc kubenswrapper[4789]: I1122 08:54:17.826914 4789 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3b679ee8-b6ac-4313-ac79-abd465d1b0c4-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:54:17 crc kubenswrapper[4789]: I1122 08:54:17.826962 4789 reconciler_common.go:293] "Volume detached for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/3b679ee8-b6ac-4313-ac79-abd465d1b0c4-var-lib-manila\") on node \"crc\" DevicePath \"\"" Nov 22 08:54:17 crc kubenswrapper[4789]: I1122 08:54:17.826978 4789 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/3b679ee8-b6ac-4313-ac79-abd465d1b0c4-ceph\") on node \"crc\" DevicePath \"\"" Nov 22 08:54:17 crc kubenswrapper[4789]: I1122 08:54:17.848264 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b679ee8-b6ac-4313-ac79-abd465d1b0c4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3b679ee8-b6ac-4313-ac79-abd465d1b0c4" (UID: "3b679ee8-b6ac-4313-ac79-abd465d1b0c4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:54:17 crc kubenswrapper[4789]: I1122 08:54:17.891464 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b679ee8-b6ac-4313-ac79-abd465d1b0c4-config-data" (OuterVolumeSpecName: "config-data") pod "3b679ee8-b6ac-4313-ac79-abd465d1b0c4" (UID: "3b679ee8-b6ac-4313-ac79-abd465d1b0c4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:54:17 crc kubenswrapper[4789]: I1122 08:54:17.928558 4789 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b679ee8-b6ac-4313-ac79-abd465d1b0c4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:54:17 crc kubenswrapper[4789]: I1122 08:54:17.928595 4789 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b679ee8-b6ac-4313-ac79-abd465d1b0c4-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:54:18 crc kubenswrapper[4789]: I1122 08:54:18.596135 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Nov 22 08:54:18 crc kubenswrapper[4789]: I1122 08:54:18.596336 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"deee1bbe-0f19-4f57-bebc-de5b01ccac7f","Type":"ContainerStarted","Data":"fff46a6cc091f5ebdfc85ba9becf49ef5f0b9032111bb756a3dccb2310862e7e"} Nov 22 08:54:18 crc kubenswrapper[4789]: I1122 08:54:18.618327 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-share-share1-0"] Nov 22 08:54:18 crc kubenswrapper[4789]: I1122 08:54:18.628225 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-share-share1-0"] Nov 22 08:54:18 crc kubenswrapper[4789]: I1122 08:54:18.637117 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-share-share1-0"] Nov 22 08:54:18 crc kubenswrapper[4789]: E1122 08:54:18.637552 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b679ee8-b6ac-4313-ac79-abd465d1b0c4" containerName="probe" Nov 22 08:54:18 crc kubenswrapper[4789]: I1122 08:54:18.637577 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b679ee8-b6ac-4313-ac79-abd465d1b0c4" containerName="probe" Nov 22 08:54:18 crc kubenswrapper[4789]: E1122 08:54:18.637613 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b679ee8-b6ac-4313-ac79-abd465d1b0c4" containerName="manila-share" Nov 22 08:54:18 crc kubenswrapper[4789]: I1122 08:54:18.637621 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b679ee8-b6ac-4313-ac79-abd465d1b0c4" containerName="manila-share" Nov 22 08:54:18 crc kubenswrapper[4789]: I1122 08:54:18.641039 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b679ee8-b6ac-4313-ac79-abd465d1b0c4" containerName="manila-share" Nov 22 08:54:18 crc kubenswrapper[4789]: I1122 08:54:18.641100 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b679ee8-b6ac-4313-ac79-abd465d1b0c4" containerName="probe" Nov 22 08:54:18 crc kubenswrapper[4789]: I1122 08:54:18.642313 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Nov 22 08:54:18 crc kubenswrapper[4789]: I1122 08:54:18.644290 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-share-share1-config-data" Nov 22 08:54:18 crc kubenswrapper[4789]: I1122 08:54:18.658378 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Nov 22 08:54:18 crc kubenswrapper[4789]: I1122 08:54:18.742489 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4l9v8\" (UniqueName: \"kubernetes.io/projected/f4f5702a-2376-40d4-9079-945c44d5466f-kube-api-access-4l9v8\") pod \"manila-share-share1-0\" (UID: \"f4f5702a-2376-40d4-9079-945c44d5466f\") " pod="openstack/manila-share-share1-0" Nov 22 08:54:18 crc kubenswrapper[4789]: I1122 08:54:18.742571 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4f5702a-2376-40d4-9079-945c44d5466f-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"f4f5702a-2376-40d4-9079-945c44d5466f\") " pod="openstack/manila-share-share1-0" Nov 22 08:54:18 crc kubenswrapper[4789]: I1122 08:54:18.742615 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4f5702a-2376-40d4-9079-945c44d5466f-config-data\") pod \"manila-share-share1-0\" (UID: \"f4f5702a-2376-40d4-9079-945c44d5466f\") " pod="openstack/manila-share-share1-0" Nov 22 08:54:18 crc kubenswrapper[4789]: I1122 08:54:18.742659 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f4f5702a-2376-40d4-9079-945c44d5466f-scripts\") pod \"manila-share-share1-0\" (UID: \"f4f5702a-2376-40d4-9079-945c44d5466f\") " pod="openstack/manila-share-share1-0" Nov 22 08:54:18 crc kubenswrapper[4789]: I1122 08:54:18.742677 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/f4f5702a-2376-40d4-9079-945c44d5466f-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"f4f5702a-2376-40d4-9079-945c44d5466f\") " pod="openstack/manila-share-share1-0" Nov 22 08:54:18 crc kubenswrapper[4789]: I1122 08:54:18.742722 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f4f5702a-2376-40d4-9079-945c44d5466f-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"f4f5702a-2376-40d4-9079-945c44d5466f\") " pod="openstack/manila-share-share1-0" Nov 22 08:54:18 crc kubenswrapper[4789]: I1122 08:54:18.742773 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f4f5702a-2376-40d4-9079-945c44d5466f-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"f4f5702a-2376-40d4-9079-945c44d5466f\") " pod="openstack/manila-share-share1-0" Nov 22 08:54:18 crc kubenswrapper[4789]: I1122 08:54:18.742831 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/f4f5702a-2376-40d4-9079-945c44d5466f-ceph\") pod \"manila-share-share1-0\" (UID: \"f4f5702a-2376-40d4-9079-945c44d5466f\") " pod="openstack/manila-share-share1-0" Nov 22 08:54:18 crc kubenswrapper[4789]: I1122 08:54:18.845771 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f4f5702a-2376-40d4-9079-945c44d5466f-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"f4f5702a-2376-40d4-9079-945c44d5466f\") " pod="openstack/manila-share-share1-0" Nov 22 08:54:18 crc kubenswrapper[4789]: I1122 08:54:18.846131 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f4f5702a-2376-40d4-9079-945c44d5466f-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"f4f5702a-2376-40d4-9079-945c44d5466f\") " pod="openstack/manila-share-share1-0" Nov 22 08:54:18 crc kubenswrapper[4789]: I1122 08:54:18.846256 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f4f5702a-2376-40d4-9079-945c44d5466f-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"f4f5702a-2376-40d4-9079-945c44d5466f\") " pod="openstack/manila-share-share1-0" Nov 22 08:54:18 crc kubenswrapper[4789]: I1122 08:54:18.846495 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/f4f5702a-2376-40d4-9079-945c44d5466f-ceph\") pod \"manila-share-share1-0\" (UID: \"f4f5702a-2376-40d4-9079-945c44d5466f\") " pod="openstack/manila-share-share1-0" Nov 22 08:54:18 crc kubenswrapper[4789]: I1122 08:54:18.846649 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4l9v8\" (UniqueName: \"kubernetes.io/projected/f4f5702a-2376-40d4-9079-945c44d5466f-kube-api-access-4l9v8\") pod \"manila-share-share1-0\" (UID: \"f4f5702a-2376-40d4-9079-945c44d5466f\") " pod="openstack/manila-share-share1-0" Nov 22 08:54:18 crc kubenswrapper[4789]: I1122 08:54:18.846832 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4f5702a-2376-40d4-9079-945c44d5466f-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"f4f5702a-2376-40d4-9079-945c44d5466f\") " pod="openstack/manila-share-share1-0" Nov 22 08:54:18 crc kubenswrapper[4789]: I1122 08:54:18.847070 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4f5702a-2376-40d4-9079-945c44d5466f-config-data\") pod \"manila-share-share1-0\" (UID: \"f4f5702a-2376-40d4-9079-945c44d5466f\") " pod="openstack/manila-share-share1-0" Nov 22 08:54:18 crc kubenswrapper[4789]: I1122 08:54:18.849698 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f4f5702a-2376-40d4-9079-945c44d5466f-scripts\") pod \"manila-share-share1-0\" (UID: \"f4f5702a-2376-40d4-9079-945c44d5466f\") " pod="openstack/manila-share-share1-0" Nov 22 08:54:18 crc kubenswrapper[4789]: I1122 08:54:18.850706 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/f4f5702a-2376-40d4-9079-945c44d5466f-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"f4f5702a-2376-40d4-9079-945c44d5466f\") " pod="openstack/manila-share-share1-0" Nov 22 08:54:18 crc kubenswrapper[4789]: I1122 08:54:18.850882 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/f4f5702a-2376-40d4-9079-945c44d5466f-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"f4f5702a-2376-40d4-9079-945c44d5466f\") " pod="openstack/manila-share-share1-0" Nov 22 08:54:18 crc kubenswrapper[4789]: I1122 08:54:18.850784 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f4f5702a-2376-40d4-9079-945c44d5466f-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"f4f5702a-2376-40d4-9079-945c44d5466f\") " pod="openstack/manila-share-share1-0" Nov 22 08:54:18 crc kubenswrapper[4789]: I1122 08:54:18.850237 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/f4f5702a-2376-40d4-9079-945c44d5466f-ceph\") pod \"manila-share-share1-0\" (UID: \"f4f5702a-2376-40d4-9079-945c44d5466f\") " pod="openstack/manila-share-share1-0" Nov 22 08:54:18 crc kubenswrapper[4789]: I1122 08:54:18.851663 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4f5702a-2376-40d4-9079-945c44d5466f-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"f4f5702a-2376-40d4-9079-945c44d5466f\") " pod="openstack/manila-share-share1-0" Nov 22 08:54:18 crc kubenswrapper[4789]: I1122 08:54:18.870311 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f4f5702a-2376-40d4-9079-945c44d5466f-scripts\") pod \"manila-share-share1-0\" (UID: \"f4f5702a-2376-40d4-9079-945c44d5466f\") " pod="openstack/manila-share-share1-0" Nov 22 08:54:18 crc kubenswrapper[4789]: I1122 08:54:18.870804 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4f5702a-2376-40d4-9079-945c44d5466f-config-data\") pod \"manila-share-share1-0\" (UID: \"f4f5702a-2376-40d4-9079-945c44d5466f\") " pod="openstack/manila-share-share1-0" Nov 22 08:54:18 crc kubenswrapper[4789]: I1122 08:54:18.873922 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4l9v8\" (UniqueName: \"kubernetes.io/projected/f4f5702a-2376-40d4-9079-945c44d5466f-kube-api-access-4l9v8\") pod \"manila-share-share1-0\" (UID: \"f4f5702a-2376-40d4-9079-945c44d5466f\") " pod="openstack/manila-share-share1-0" Nov 22 08:54:18 crc kubenswrapper[4789]: I1122 08:54:18.959431 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Nov 22 08:54:19 crc kubenswrapper[4789]: I1122 08:54:19.480164 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Nov 22 08:54:19 crc kubenswrapper[4789]: W1122 08:54:19.483158 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf4f5702a_2376_40d4_9079_945c44d5466f.slice/crio-8b5bab21b4c80b174ba96f365df62aa723d44620bf0509c495065da33ac9dd1c WatchSource:0}: Error finding container 8b5bab21b4c80b174ba96f365df62aa723d44620bf0509c495065da33ac9dd1c: Status 404 returned error can't find the container with id 8b5bab21b4c80b174ba96f365df62aa723d44620bf0509c495065da33ac9dd1c Nov 22 08:54:19 crc kubenswrapper[4789]: I1122 08:54:19.612984 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"f4f5702a-2376-40d4-9079-945c44d5466f","Type":"ContainerStarted","Data":"8b5bab21b4c80b174ba96f365df62aa723d44620bf0509c495065da33ac9dd1c"} Nov 22 08:54:19 crc kubenswrapper[4789]: I1122 08:54:19.618362 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"deee1bbe-0f19-4f57-bebc-de5b01ccac7f","Type":"ContainerStarted","Data":"ea0596baacae27357ff964ff1c1d866775d15756816fa095aa9a8c1c99c9341d"} Nov 22 08:54:19 crc kubenswrapper[4789]: I1122 08:54:19.619156 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 22 08:54:19 crc kubenswrapper[4789]: I1122 08:54:19.654579 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.216018372 podStartE2EDuration="5.654553757s" podCreationTimestamp="2025-11-22 08:54:14 +0000 UTC" firstStartedPulling="2025-11-22 08:54:15.391819072 +0000 UTC m=+3469.626219345" lastFinishedPulling="2025-11-22 08:54:18.830354457 +0000 UTC m=+3473.064754730" observedRunningTime="2025-11-22 08:54:19.640901089 +0000 UTC m=+3473.875301392" watchObservedRunningTime="2025-11-22 08:54:19.654553757 +0000 UTC m=+3473.888954030" Nov 22 08:54:19 crc kubenswrapper[4789]: I1122 08:54:19.980048 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3b679ee8-b6ac-4313-ac79-abd465d1b0c4" path="/var/lib/kubelet/pods/3b679ee8-b6ac-4313-ac79-abd465d1b0c4/volumes" Nov 22 08:54:20 crc kubenswrapper[4789]: I1122 08:54:20.628940 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"f4f5702a-2376-40d4-9079-945c44d5466f","Type":"ContainerStarted","Data":"38309336afe7c794c95728cabc6bc599bf78f4834dd7e9e62313c46ee7a67215"} Nov 22 08:54:20 crc kubenswrapper[4789]: I1122 08:54:20.630786 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"f4f5702a-2376-40d4-9079-945c44d5466f","Type":"ContainerStarted","Data":"4ccf0d7b720ba71909f641d70ae54a3512e5b23c646845c94761ebc4b3e735e6"} Nov 22 08:54:21 crc kubenswrapper[4789]: I1122 08:54:21.610592 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Nov 22 08:54:21 crc kubenswrapper[4789]: I1122 08:54:21.641137 4789 generic.go:334] "Generic (PLEG): container finished" podID="46af0a4f-fe66-47b9-b9d9-892c3eb3d97c" containerID="49d5d401ae263afccfb4c1717363fd6219980a592b2a9754a395aa281e5faabd" exitCode=0 Nov 22 08:54:21 crc kubenswrapper[4789]: I1122 08:54:21.641707 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Nov 22 08:54:21 crc kubenswrapper[4789]: I1122 08:54:21.642445 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"46af0a4f-fe66-47b9-b9d9-892c3eb3d97c","Type":"ContainerDied","Data":"49d5d401ae263afccfb4c1717363fd6219980a592b2a9754a395aa281e5faabd"} Nov 22 08:54:21 crc kubenswrapper[4789]: I1122 08:54:21.642491 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"46af0a4f-fe66-47b9-b9d9-892c3eb3d97c","Type":"ContainerDied","Data":"4a76c66c85baf528e18015f5fcf7ecc3cdf42023660ab95698cc768f618784ad"} Nov 22 08:54:21 crc kubenswrapper[4789]: I1122 08:54:21.642516 4789 scope.go:117] "RemoveContainer" containerID="b6d59c915dfda3979f0214cccd7333388cee6ab9e25416acb131775f14556d36" Nov 22 08:54:21 crc kubenswrapper[4789]: I1122 08:54:21.671426 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-share-share1-0" podStartSLOduration=3.671404175 podStartE2EDuration="3.671404175s" podCreationTimestamp="2025-11-22 08:54:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:54:21.66713432 +0000 UTC m=+3475.901534593" watchObservedRunningTime="2025-11-22 08:54:21.671404175 +0000 UTC m=+3475.905804448" Nov 22 08:54:21 crc kubenswrapper[4789]: I1122 08:54:21.692501 4789 scope.go:117] "RemoveContainer" containerID="49d5d401ae263afccfb4c1717363fd6219980a592b2a9754a395aa281e5faabd" Nov 22 08:54:21 crc kubenswrapper[4789]: I1122 08:54:21.717945 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r2mss\" (UniqueName: \"kubernetes.io/projected/46af0a4f-fe66-47b9-b9d9-892c3eb3d97c-kube-api-access-r2mss\") pod \"46af0a4f-fe66-47b9-b9d9-892c3eb3d97c\" (UID: \"46af0a4f-fe66-47b9-b9d9-892c3eb3d97c\") " Nov 22 08:54:21 crc kubenswrapper[4789]: I1122 08:54:21.718039 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/46af0a4f-fe66-47b9-b9d9-892c3eb3d97c-config-data-custom\") pod \"46af0a4f-fe66-47b9-b9d9-892c3eb3d97c\" (UID: \"46af0a4f-fe66-47b9-b9d9-892c3eb3d97c\") " Nov 22 08:54:21 crc kubenswrapper[4789]: I1122 08:54:21.718148 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/46af0a4f-fe66-47b9-b9d9-892c3eb3d97c-etc-machine-id\") pod \"46af0a4f-fe66-47b9-b9d9-892c3eb3d97c\" (UID: \"46af0a4f-fe66-47b9-b9d9-892c3eb3d97c\") " Nov 22 08:54:21 crc kubenswrapper[4789]: I1122 08:54:21.718176 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46af0a4f-fe66-47b9-b9d9-892c3eb3d97c-config-data\") pod \"46af0a4f-fe66-47b9-b9d9-892c3eb3d97c\" (UID: \"46af0a4f-fe66-47b9-b9d9-892c3eb3d97c\") " Nov 22 08:54:21 crc kubenswrapper[4789]: I1122 08:54:21.718196 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/46af0a4f-fe66-47b9-b9d9-892c3eb3d97c-scripts\") pod \"46af0a4f-fe66-47b9-b9d9-892c3eb3d97c\" (UID: \"46af0a4f-fe66-47b9-b9d9-892c3eb3d97c\") " Nov 22 08:54:21 crc kubenswrapper[4789]: I1122 08:54:21.718309 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46af0a4f-fe66-47b9-b9d9-892c3eb3d97c-combined-ca-bundle\") pod \"46af0a4f-fe66-47b9-b9d9-892c3eb3d97c\" (UID: \"46af0a4f-fe66-47b9-b9d9-892c3eb3d97c\") " Nov 22 08:54:21 crc kubenswrapper[4789]: I1122 08:54:21.718312 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/46af0a4f-fe66-47b9-b9d9-892c3eb3d97c-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "46af0a4f-fe66-47b9-b9d9-892c3eb3d97c" (UID: "46af0a4f-fe66-47b9-b9d9-892c3eb3d97c"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 08:54:21 crc kubenswrapper[4789]: I1122 08:54:21.720207 4789 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/46af0a4f-fe66-47b9-b9d9-892c3eb3d97c-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 22 08:54:21 crc kubenswrapper[4789]: I1122 08:54:21.723329 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46af0a4f-fe66-47b9-b9d9-892c3eb3d97c-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "46af0a4f-fe66-47b9-b9d9-892c3eb3d97c" (UID: "46af0a4f-fe66-47b9-b9d9-892c3eb3d97c"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:54:21 crc kubenswrapper[4789]: I1122 08:54:21.723979 4789 scope.go:117] "RemoveContainer" containerID="b6d59c915dfda3979f0214cccd7333388cee6ab9e25416acb131775f14556d36" Nov 22 08:54:21 crc kubenswrapper[4789]: I1122 08:54:21.724318 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46af0a4f-fe66-47b9-b9d9-892c3eb3d97c-scripts" (OuterVolumeSpecName: "scripts") pod "46af0a4f-fe66-47b9-b9d9-892c3eb3d97c" (UID: "46af0a4f-fe66-47b9-b9d9-892c3eb3d97c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:54:21 crc kubenswrapper[4789]: I1122 08:54:21.725549 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/46af0a4f-fe66-47b9-b9d9-892c3eb3d97c-kube-api-access-r2mss" (OuterVolumeSpecName: "kube-api-access-r2mss") pod "46af0a4f-fe66-47b9-b9d9-892c3eb3d97c" (UID: "46af0a4f-fe66-47b9-b9d9-892c3eb3d97c"). InnerVolumeSpecName "kube-api-access-r2mss". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:54:21 crc kubenswrapper[4789]: E1122 08:54:21.728044 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b6d59c915dfda3979f0214cccd7333388cee6ab9e25416acb131775f14556d36\": container with ID starting with b6d59c915dfda3979f0214cccd7333388cee6ab9e25416acb131775f14556d36 not found: ID does not exist" containerID="b6d59c915dfda3979f0214cccd7333388cee6ab9e25416acb131775f14556d36" Nov 22 08:54:21 crc kubenswrapper[4789]: I1122 08:54:21.728157 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b6d59c915dfda3979f0214cccd7333388cee6ab9e25416acb131775f14556d36"} err="failed to get container status \"b6d59c915dfda3979f0214cccd7333388cee6ab9e25416acb131775f14556d36\": rpc error: code = NotFound desc = could not find container \"b6d59c915dfda3979f0214cccd7333388cee6ab9e25416acb131775f14556d36\": container with ID starting with b6d59c915dfda3979f0214cccd7333388cee6ab9e25416acb131775f14556d36 not found: ID does not exist" Nov 22 08:54:21 crc kubenswrapper[4789]: I1122 08:54:21.728234 4789 scope.go:117] "RemoveContainer" containerID="49d5d401ae263afccfb4c1717363fd6219980a592b2a9754a395aa281e5faabd" Nov 22 08:54:21 crc kubenswrapper[4789]: E1122 08:54:21.728692 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"49d5d401ae263afccfb4c1717363fd6219980a592b2a9754a395aa281e5faabd\": container with ID starting with 49d5d401ae263afccfb4c1717363fd6219980a592b2a9754a395aa281e5faabd not found: ID does not exist" containerID="49d5d401ae263afccfb4c1717363fd6219980a592b2a9754a395aa281e5faabd" Nov 22 08:54:21 crc kubenswrapper[4789]: I1122 08:54:21.728792 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"49d5d401ae263afccfb4c1717363fd6219980a592b2a9754a395aa281e5faabd"} err="failed to get container status \"49d5d401ae263afccfb4c1717363fd6219980a592b2a9754a395aa281e5faabd\": rpc error: code = NotFound desc = could not find container \"49d5d401ae263afccfb4c1717363fd6219980a592b2a9754a395aa281e5faabd\": container with ID starting with 49d5d401ae263afccfb4c1717363fd6219980a592b2a9754a395aa281e5faabd not found: ID does not exist" Nov 22 08:54:21 crc kubenswrapper[4789]: I1122 08:54:21.770959 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46af0a4f-fe66-47b9-b9d9-892c3eb3d97c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "46af0a4f-fe66-47b9-b9d9-892c3eb3d97c" (UID: "46af0a4f-fe66-47b9-b9d9-892c3eb3d97c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:54:21 crc kubenswrapper[4789]: I1122 08:54:21.822210 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r2mss\" (UniqueName: \"kubernetes.io/projected/46af0a4f-fe66-47b9-b9d9-892c3eb3d97c-kube-api-access-r2mss\") on node \"crc\" DevicePath \"\"" Nov 22 08:54:21 crc kubenswrapper[4789]: I1122 08:54:21.822414 4789 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/46af0a4f-fe66-47b9-b9d9-892c3eb3d97c-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 22 08:54:21 crc kubenswrapper[4789]: I1122 08:54:21.822471 4789 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/46af0a4f-fe66-47b9-b9d9-892c3eb3d97c-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 08:54:21 crc kubenswrapper[4789]: I1122 08:54:21.822523 4789 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46af0a4f-fe66-47b9-b9d9-892c3eb3d97c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 08:54:21 crc kubenswrapper[4789]: I1122 08:54:21.837318 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46af0a4f-fe66-47b9-b9d9-892c3eb3d97c-config-data" (OuterVolumeSpecName: "config-data") pod "46af0a4f-fe66-47b9-b9d9-892c3eb3d97c" (UID: "46af0a4f-fe66-47b9-b9d9-892c3eb3d97c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 08:54:21 crc kubenswrapper[4789]: I1122 08:54:21.924587 4789 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46af0a4f-fe66-47b9-b9d9-892c3eb3d97c-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 08:54:21 crc kubenswrapper[4789]: I1122 08:54:21.986058 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-scheduler-0"] Nov 22 08:54:22 crc kubenswrapper[4789]: I1122 08:54:21.999739 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-scheduler-0"] Nov 22 08:54:22 crc kubenswrapper[4789]: I1122 08:54:22.008771 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-scheduler-0"] Nov 22 08:54:22 crc kubenswrapper[4789]: E1122 08:54:22.018919 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46af0a4f-fe66-47b9-b9d9-892c3eb3d97c" containerName="manila-scheduler" Nov 22 08:54:22 crc kubenswrapper[4789]: I1122 08:54:22.018960 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="46af0a4f-fe66-47b9-b9d9-892c3eb3d97c" containerName="manila-scheduler" Nov 22 08:54:22 crc kubenswrapper[4789]: E1122 08:54:22.018971 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46af0a4f-fe66-47b9-b9d9-892c3eb3d97c" containerName="probe" Nov 22 08:54:22 crc kubenswrapper[4789]: I1122 08:54:22.018978 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="46af0a4f-fe66-47b9-b9d9-892c3eb3d97c" containerName="probe" Nov 22 08:54:22 crc kubenswrapper[4789]: I1122 08:54:22.019249 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="46af0a4f-fe66-47b9-b9d9-892c3eb3d97c" containerName="manila-scheduler" Nov 22 08:54:22 crc kubenswrapper[4789]: I1122 08:54:22.019277 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="46af0a4f-fe66-47b9-b9d9-892c3eb3d97c" containerName="probe" Nov 22 08:54:22 crc kubenswrapper[4789]: I1122 08:54:22.020380 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Nov 22 08:54:22 crc kubenswrapper[4789]: I1122 08:54:22.020500 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Nov 22 08:54:22 crc kubenswrapper[4789]: I1122 08:54:22.023990 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scheduler-config-data" Nov 22 08:54:22 crc kubenswrapper[4789]: I1122 08:54:22.129580 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8a211302-5dc3-4ba1-959e-882c398a7f6c-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"8a211302-5dc3-4ba1-959e-882c398a7f6c\") " pod="openstack/manila-scheduler-0" Nov 22 08:54:22 crc kubenswrapper[4789]: I1122 08:54:22.129693 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a211302-5dc3-4ba1-959e-882c398a7f6c-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"8a211302-5dc3-4ba1-959e-882c398a7f6c\") " pod="openstack/manila-scheduler-0" Nov 22 08:54:22 crc kubenswrapper[4789]: I1122 08:54:22.129795 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8a211302-5dc3-4ba1-959e-882c398a7f6c-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"8a211302-5dc3-4ba1-959e-882c398a7f6c\") " pod="openstack/manila-scheduler-0" Nov 22 08:54:22 crc kubenswrapper[4789]: I1122 08:54:22.129906 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a211302-5dc3-4ba1-959e-882c398a7f6c-config-data\") pod \"manila-scheduler-0\" (UID: \"8a211302-5dc3-4ba1-959e-882c398a7f6c\") " pod="openstack/manila-scheduler-0" Nov 22 08:54:22 crc kubenswrapper[4789]: I1122 08:54:22.129960 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-92bcz\" (UniqueName: \"kubernetes.io/projected/8a211302-5dc3-4ba1-959e-882c398a7f6c-kube-api-access-92bcz\") pod \"manila-scheduler-0\" (UID: \"8a211302-5dc3-4ba1-959e-882c398a7f6c\") " pod="openstack/manila-scheduler-0" Nov 22 08:54:22 crc kubenswrapper[4789]: I1122 08:54:22.130191 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a211302-5dc3-4ba1-959e-882c398a7f6c-scripts\") pod \"manila-scheduler-0\" (UID: \"8a211302-5dc3-4ba1-959e-882c398a7f6c\") " pod="openstack/manila-scheduler-0" Nov 22 08:54:22 crc kubenswrapper[4789]: I1122 08:54:22.231935 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a211302-5dc3-4ba1-959e-882c398a7f6c-scripts\") pod \"manila-scheduler-0\" (UID: \"8a211302-5dc3-4ba1-959e-882c398a7f6c\") " pod="openstack/manila-scheduler-0" Nov 22 08:54:22 crc kubenswrapper[4789]: I1122 08:54:22.232000 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8a211302-5dc3-4ba1-959e-882c398a7f6c-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"8a211302-5dc3-4ba1-959e-882c398a7f6c\") " pod="openstack/manila-scheduler-0" Nov 22 08:54:22 crc kubenswrapper[4789]: I1122 08:54:22.232024 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a211302-5dc3-4ba1-959e-882c398a7f6c-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"8a211302-5dc3-4ba1-959e-882c398a7f6c\") " pod="openstack/manila-scheduler-0" Nov 22 08:54:22 crc kubenswrapper[4789]: I1122 08:54:22.232054 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8a211302-5dc3-4ba1-959e-882c398a7f6c-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"8a211302-5dc3-4ba1-959e-882c398a7f6c\") " pod="openstack/manila-scheduler-0" Nov 22 08:54:22 crc kubenswrapper[4789]: I1122 08:54:22.232096 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a211302-5dc3-4ba1-959e-882c398a7f6c-config-data\") pod \"manila-scheduler-0\" (UID: \"8a211302-5dc3-4ba1-959e-882c398a7f6c\") " pod="openstack/manila-scheduler-0" Nov 22 08:54:22 crc kubenswrapper[4789]: I1122 08:54:22.232118 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-92bcz\" (UniqueName: \"kubernetes.io/projected/8a211302-5dc3-4ba1-959e-882c398a7f6c-kube-api-access-92bcz\") pod \"manila-scheduler-0\" (UID: \"8a211302-5dc3-4ba1-959e-882c398a7f6c\") " pod="openstack/manila-scheduler-0" Nov 22 08:54:22 crc kubenswrapper[4789]: I1122 08:54:22.232604 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8a211302-5dc3-4ba1-959e-882c398a7f6c-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"8a211302-5dc3-4ba1-959e-882c398a7f6c\") " pod="openstack/manila-scheduler-0" Nov 22 08:54:22 crc kubenswrapper[4789]: I1122 08:54:22.235816 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a211302-5dc3-4ba1-959e-882c398a7f6c-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"8a211302-5dc3-4ba1-959e-882c398a7f6c\") " pod="openstack/manila-scheduler-0" Nov 22 08:54:22 crc kubenswrapper[4789]: I1122 08:54:22.236084 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8a211302-5dc3-4ba1-959e-882c398a7f6c-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"8a211302-5dc3-4ba1-959e-882c398a7f6c\") " pod="openstack/manila-scheduler-0" Nov 22 08:54:22 crc kubenswrapper[4789]: I1122 08:54:22.236083 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a211302-5dc3-4ba1-959e-882c398a7f6c-scripts\") pod \"manila-scheduler-0\" (UID: \"8a211302-5dc3-4ba1-959e-882c398a7f6c\") " pod="openstack/manila-scheduler-0" Nov 22 08:54:22 crc kubenswrapper[4789]: I1122 08:54:22.236728 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a211302-5dc3-4ba1-959e-882c398a7f6c-config-data\") pod \"manila-scheduler-0\" (UID: \"8a211302-5dc3-4ba1-959e-882c398a7f6c\") " pod="openstack/manila-scheduler-0" Nov 22 08:54:22 crc kubenswrapper[4789]: I1122 08:54:22.248712 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-92bcz\" (UniqueName: \"kubernetes.io/projected/8a211302-5dc3-4ba1-959e-882c398a7f6c-kube-api-access-92bcz\") pod \"manila-scheduler-0\" (UID: \"8a211302-5dc3-4ba1-959e-882c398a7f6c\") " pod="openstack/manila-scheduler-0" Nov 22 08:54:22 crc kubenswrapper[4789]: I1122 08:54:22.350477 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Nov 22 08:54:22 crc kubenswrapper[4789]: I1122 08:54:22.777234 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Nov 22 08:54:22 crc kubenswrapper[4789]: W1122 08:54:22.787366 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8a211302_5dc3_4ba1_959e_882c398a7f6c.slice/crio-78bcf427f02679bb8c00beb136676365652c0b1efcb27b3aae16fc9eeaf9be61 WatchSource:0}: Error finding container 78bcf427f02679bb8c00beb136676365652c0b1efcb27b3aae16fc9eeaf9be61: Status 404 returned error can't find the container with id 78bcf427f02679bb8c00beb136676365652c0b1efcb27b3aae16fc9eeaf9be61 Nov 22 08:54:23 crc kubenswrapper[4789]: I1122 08:54:23.679267 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"8a211302-5dc3-4ba1-959e-882c398a7f6c","Type":"ContainerStarted","Data":"b2fed9de5adf7ed419b23c0c73d53a9529607cfc37885e223901649e09d67535"} Nov 22 08:54:23 crc kubenswrapper[4789]: I1122 08:54:23.680069 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"8a211302-5dc3-4ba1-959e-882c398a7f6c","Type":"ContainerStarted","Data":"78bcf427f02679bb8c00beb136676365652c0b1efcb27b3aae16fc9eeaf9be61"} Nov 22 08:54:23 crc kubenswrapper[4789]: I1122 08:54:23.984863 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="46af0a4f-fe66-47b9-b9d9-892c3eb3d97c" path="/var/lib/kubelet/pods/46af0a4f-fe66-47b9-b9d9-892c3eb3d97c/volumes" Nov 22 08:54:24 crc kubenswrapper[4789]: I1122 08:54:24.690616 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"8a211302-5dc3-4ba1-959e-882c398a7f6c","Type":"ContainerStarted","Data":"dc9b6f9ff18046579de7512cd15a976b9db6500bc12a6a064bc28ddc55df5a8f"} Nov 22 08:54:24 crc kubenswrapper[4789]: I1122 08:54:24.712684 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-scheduler-0" podStartSLOduration=3.712664704 podStartE2EDuration="3.712664704s" podCreationTimestamp="2025-11-22 08:54:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:54:24.705988004 +0000 UTC m=+3478.940388297" watchObservedRunningTime="2025-11-22 08:54:24.712664704 +0000 UTC m=+3478.947064987" Nov 22 08:54:26 crc kubenswrapper[4789]: I1122 08:54:26.676226 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/manila-api-0" Nov 22 08:54:28 crc kubenswrapper[4789]: I1122 08:54:28.960380 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-share-share1-0" Nov 22 08:54:32 crc kubenswrapper[4789]: I1122 08:54:32.351175 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-scheduler-0" Nov 22 08:54:39 crc kubenswrapper[4789]: I1122 08:54:39.785685 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-kr24d"] Nov 22 08:54:39 crc kubenswrapper[4789]: I1122 08:54:39.788867 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kr24d" Nov 22 08:54:39 crc kubenswrapper[4789]: I1122 08:54:39.795378 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kr24d"] Nov 22 08:54:39 crc kubenswrapper[4789]: I1122 08:54:39.886072 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sxtzw\" (UniqueName: \"kubernetes.io/projected/8f44ffc1-1437-4e0a-b083-2f14ab9f89f1-kube-api-access-sxtzw\") pod \"redhat-operators-kr24d\" (UID: \"8f44ffc1-1437-4e0a-b083-2f14ab9f89f1\") " pod="openshift-marketplace/redhat-operators-kr24d" Nov 22 08:54:39 crc kubenswrapper[4789]: I1122 08:54:39.886130 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f44ffc1-1437-4e0a-b083-2f14ab9f89f1-utilities\") pod \"redhat-operators-kr24d\" (UID: \"8f44ffc1-1437-4e0a-b083-2f14ab9f89f1\") " pod="openshift-marketplace/redhat-operators-kr24d" Nov 22 08:54:39 crc kubenswrapper[4789]: I1122 08:54:39.886148 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f44ffc1-1437-4e0a-b083-2f14ab9f89f1-catalog-content\") pod \"redhat-operators-kr24d\" (UID: \"8f44ffc1-1437-4e0a-b083-2f14ab9f89f1\") " pod="openshift-marketplace/redhat-operators-kr24d" Nov 22 08:54:39 crc kubenswrapper[4789]: I1122 08:54:39.988293 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sxtzw\" (UniqueName: \"kubernetes.io/projected/8f44ffc1-1437-4e0a-b083-2f14ab9f89f1-kube-api-access-sxtzw\") pod \"redhat-operators-kr24d\" (UID: \"8f44ffc1-1437-4e0a-b083-2f14ab9f89f1\") " pod="openshift-marketplace/redhat-operators-kr24d" Nov 22 08:54:39 crc kubenswrapper[4789]: I1122 08:54:39.988351 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f44ffc1-1437-4e0a-b083-2f14ab9f89f1-utilities\") pod \"redhat-operators-kr24d\" (UID: \"8f44ffc1-1437-4e0a-b083-2f14ab9f89f1\") " pod="openshift-marketplace/redhat-operators-kr24d" Nov 22 08:54:39 crc kubenswrapper[4789]: I1122 08:54:39.988369 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f44ffc1-1437-4e0a-b083-2f14ab9f89f1-catalog-content\") pod \"redhat-operators-kr24d\" (UID: \"8f44ffc1-1437-4e0a-b083-2f14ab9f89f1\") " pod="openshift-marketplace/redhat-operators-kr24d" Nov 22 08:54:39 crc kubenswrapper[4789]: I1122 08:54:39.989017 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f44ffc1-1437-4e0a-b083-2f14ab9f89f1-catalog-content\") pod \"redhat-operators-kr24d\" (UID: \"8f44ffc1-1437-4e0a-b083-2f14ab9f89f1\") " pod="openshift-marketplace/redhat-operators-kr24d" Nov 22 08:54:39 crc kubenswrapper[4789]: I1122 08:54:39.989045 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f44ffc1-1437-4e0a-b083-2f14ab9f89f1-utilities\") pod \"redhat-operators-kr24d\" (UID: \"8f44ffc1-1437-4e0a-b083-2f14ab9f89f1\") " pod="openshift-marketplace/redhat-operators-kr24d" Nov 22 08:54:40 crc kubenswrapper[4789]: I1122 08:54:40.016112 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sxtzw\" (UniqueName: \"kubernetes.io/projected/8f44ffc1-1437-4e0a-b083-2f14ab9f89f1-kube-api-access-sxtzw\") pod \"redhat-operators-kr24d\" (UID: \"8f44ffc1-1437-4e0a-b083-2f14ab9f89f1\") " pod="openshift-marketplace/redhat-operators-kr24d" Nov 22 08:54:40 crc kubenswrapper[4789]: I1122 08:54:40.110903 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kr24d" Nov 22 08:54:40 crc kubenswrapper[4789]: I1122 08:54:40.714774 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kr24d"] Nov 22 08:54:40 crc kubenswrapper[4789]: I1122 08:54:40.775296 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-share-share1-0" Nov 22 08:54:40 crc kubenswrapper[4789]: I1122 08:54:40.838954 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kr24d" event={"ID":"8f44ffc1-1437-4e0a-b083-2f14ab9f89f1","Type":"ContainerStarted","Data":"864f12578898cc44ed92d4fa95c347d8044941de542f61f025c12d0247bf762f"} Nov 22 08:54:41 crc kubenswrapper[4789]: I1122 08:54:41.849827 4789 generic.go:334] "Generic (PLEG): container finished" podID="8f44ffc1-1437-4e0a-b083-2f14ab9f89f1" containerID="5878e41eb2b9dda2cbf8d65523f932bcca37839be0dd0a2a78a79c8e9cca90a9" exitCode=0 Nov 22 08:54:41 crc kubenswrapper[4789]: I1122 08:54:41.849897 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kr24d" event={"ID":"8f44ffc1-1437-4e0a-b083-2f14ab9f89f1","Type":"ContainerDied","Data":"5878e41eb2b9dda2cbf8d65523f932bcca37839be0dd0a2a78a79c8e9cca90a9"} Nov 22 08:54:41 crc kubenswrapper[4789]: I1122 08:54:41.852526 4789 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 22 08:54:43 crc kubenswrapper[4789]: I1122 08:54:43.877460 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kr24d" event={"ID":"8f44ffc1-1437-4e0a-b083-2f14ab9f89f1","Type":"ContainerStarted","Data":"7b9e2fb24cb3bdbe2c862c83a6e5622f9df816ff654ec2e5c97c0efa9c7d81fe"} Nov 22 08:54:43 crc kubenswrapper[4789]: I1122 08:54:43.932366 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-scheduler-0" Nov 22 08:54:44 crc kubenswrapper[4789]: I1122 08:54:44.943101 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 22 08:54:45 crc kubenswrapper[4789]: I1122 08:54:45.895831 4789 generic.go:334] "Generic (PLEG): container finished" podID="8f44ffc1-1437-4e0a-b083-2f14ab9f89f1" containerID="7b9e2fb24cb3bdbe2c862c83a6e5622f9df816ff654ec2e5c97c0efa9c7d81fe" exitCode=0 Nov 22 08:54:45 crc kubenswrapper[4789]: I1122 08:54:45.896059 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kr24d" event={"ID":"8f44ffc1-1437-4e0a-b083-2f14ab9f89f1","Type":"ContainerDied","Data":"7b9e2fb24cb3bdbe2c862c83a6e5622f9df816ff654ec2e5c97c0efa9c7d81fe"} Nov 22 08:54:47 crc kubenswrapper[4789]: I1122 08:54:47.915368 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kr24d" event={"ID":"8f44ffc1-1437-4e0a-b083-2f14ab9f89f1","Type":"ContainerStarted","Data":"f95f8ebc989eeb2a8f38242aa671bf0238b5577322da97486e429a7f3d816ba7"} Nov 22 08:54:47 crc kubenswrapper[4789]: I1122 08:54:47.938712 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-kr24d" podStartSLOduration=3.550890087 podStartE2EDuration="8.938687361s" podCreationTimestamp="2025-11-22 08:54:39 +0000 UTC" firstStartedPulling="2025-11-22 08:54:41.852242857 +0000 UTC m=+3496.086643130" lastFinishedPulling="2025-11-22 08:54:47.240040131 +0000 UTC m=+3501.474440404" observedRunningTime="2025-11-22 08:54:47.930676296 +0000 UTC m=+3502.165076569" watchObservedRunningTime="2025-11-22 08:54:47.938687361 +0000 UTC m=+3502.173087634" Nov 22 08:54:50 crc kubenswrapper[4789]: I1122 08:54:50.112094 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-kr24d" Nov 22 08:54:50 crc kubenswrapper[4789]: I1122 08:54:50.112600 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-kr24d" Nov 22 08:54:51 crc kubenswrapper[4789]: I1122 08:54:51.153007 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-kr24d" podUID="8f44ffc1-1437-4e0a-b083-2f14ab9f89f1" containerName="registry-server" probeResult="failure" output=< Nov 22 08:54:51 crc kubenswrapper[4789]: timeout: failed to connect service ":50051" within 1s Nov 22 08:54:51 crc kubenswrapper[4789]: > Nov 22 08:55:00 crc kubenswrapper[4789]: I1122 08:55:00.163497 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-kr24d" Nov 22 08:55:00 crc kubenswrapper[4789]: I1122 08:55:00.227935 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-kr24d" Nov 22 08:55:00 crc kubenswrapper[4789]: I1122 08:55:00.401565 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kr24d"] Nov 22 08:55:02 crc kubenswrapper[4789]: I1122 08:55:02.054028 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-kr24d" podUID="8f44ffc1-1437-4e0a-b083-2f14ab9f89f1" containerName="registry-server" containerID="cri-o://f95f8ebc989eeb2a8f38242aa671bf0238b5577322da97486e429a7f3d816ba7" gracePeriod=2 Nov 22 08:55:02 crc kubenswrapper[4789]: I1122 08:55:02.497379 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kr24d" Nov 22 08:55:02 crc kubenswrapper[4789]: I1122 08:55:02.544803 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f44ffc1-1437-4e0a-b083-2f14ab9f89f1-utilities\") pod \"8f44ffc1-1437-4e0a-b083-2f14ab9f89f1\" (UID: \"8f44ffc1-1437-4e0a-b083-2f14ab9f89f1\") " Nov 22 08:55:02 crc kubenswrapper[4789]: I1122 08:55:02.544869 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sxtzw\" (UniqueName: \"kubernetes.io/projected/8f44ffc1-1437-4e0a-b083-2f14ab9f89f1-kube-api-access-sxtzw\") pod \"8f44ffc1-1437-4e0a-b083-2f14ab9f89f1\" (UID: \"8f44ffc1-1437-4e0a-b083-2f14ab9f89f1\") " Nov 22 08:55:02 crc kubenswrapper[4789]: I1122 08:55:02.544989 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f44ffc1-1437-4e0a-b083-2f14ab9f89f1-catalog-content\") pod \"8f44ffc1-1437-4e0a-b083-2f14ab9f89f1\" (UID: \"8f44ffc1-1437-4e0a-b083-2f14ab9f89f1\") " Nov 22 08:55:02 crc kubenswrapper[4789]: I1122 08:55:02.549215 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f44ffc1-1437-4e0a-b083-2f14ab9f89f1-utilities" (OuterVolumeSpecName: "utilities") pod "8f44ffc1-1437-4e0a-b083-2f14ab9f89f1" (UID: "8f44ffc1-1437-4e0a-b083-2f14ab9f89f1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:55:02 crc kubenswrapper[4789]: I1122 08:55:02.566904 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f44ffc1-1437-4e0a-b083-2f14ab9f89f1-kube-api-access-sxtzw" (OuterVolumeSpecName: "kube-api-access-sxtzw") pod "8f44ffc1-1437-4e0a-b083-2f14ab9f89f1" (UID: "8f44ffc1-1437-4e0a-b083-2f14ab9f89f1"). InnerVolumeSpecName "kube-api-access-sxtzw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:55:02 crc kubenswrapper[4789]: I1122 08:55:02.632423 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f44ffc1-1437-4e0a-b083-2f14ab9f89f1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8f44ffc1-1437-4e0a-b083-2f14ab9f89f1" (UID: "8f44ffc1-1437-4e0a-b083-2f14ab9f89f1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:55:02 crc kubenswrapper[4789]: I1122 08:55:02.647960 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f44ffc1-1437-4e0a-b083-2f14ab9f89f1-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 08:55:02 crc kubenswrapper[4789]: I1122 08:55:02.647993 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f44ffc1-1437-4e0a-b083-2f14ab9f89f1-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 08:55:02 crc kubenswrapper[4789]: I1122 08:55:02.648022 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sxtzw\" (UniqueName: \"kubernetes.io/projected/8f44ffc1-1437-4e0a-b083-2f14ab9f89f1-kube-api-access-sxtzw\") on node \"crc\" DevicePath \"\"" Nov 22 08:55:03 crc kubenswrapper[4789]: I1122 08:55:03.064891 4789 generic.go:334] "Generic (PLEG): container finished" podID="8f44ffc1-1437-4e0a-b083-2f14ab9f89f1" containerID="f95f8ebc989eeb2a8f38242aa671bf0238b5577322da97486e429a7f3d816ba7" exitCode=0 Nov 22 08:55:03 crc kubenswrapper[4789]: I1122 08:55:03.064944 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kr24d" event={"ID":"8f44ffc1-1437-4e0a-b083-2f14ab9f89f1","Type":"ContainerDied","Data":"f95f8ebc989eeb2a8f38242aa671bf0238b5577322da97486e429a7f3d816ba7"} Nov 22 08:55:03 crc kubenswrapper[4789]: I1122 08:55:03.064972 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kr24d" event={"ID":"8f44ffc1-1437-4e0a-b083-2f14ab9f89f1","Type":"ContainerDied","Data":"864f12578898cc44ed92d4fa95c347d8044941de542f61f025c12d0247bf762f"} Nov 22 08:55:03 crc kubenswrapper[4789]: I1122 08:55:03.064991 4789 scope.go:117] "RemoveContainer" containerID="f95f8ebc989eeb2a8f38242aa671bf0238b5577322da97486e429a7f3d816ba7" Nov 22 08:55:03 crc kubenswrapper[4789]: I1122 08:55:03.065141 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kr24d" Nov 22 08:55:03 crc kubenswrapper[4789]: I1122 08:55:03.095826 4789 scope.go:117] "RemoveContainer" containerID="7b9e2fb24cb3bdbe2c862c83a6e5622f9df816ff654ec2e5c97c0efa9c7d81fe" Nov 22 08:55:03 crc kubenswrapper[4789]: I1122 08:55:03.105354 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kr24d"] Nov 22 08:55:03 crc kubenswrapper[4789]: I1122 08:55:03.141012 4789 scope.go:117] "RemoveContainer" containerID="5878e41eb2b9dda2cbf8d65523f932bcca37839be0dd0a2a78a79c8e9cca90a9" Nov 22 08:55:03 crc kubenswrapper[4789]: I1122 08:55:03.151627 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-kr24d"] Nov 22 08:55:03 crc kubenswrapper[4789]: I1122 08:55:03.177528 4789 scope.go:117] "RemoveContainer" containerID="f95f8ebc989eeb2a8f38242aa671bf0238b5577322da97486e429a7f3d816ba7" Nov 22 08:55:03 crc kubenswrapper[4789]: E1122 08:55:03.178212 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f95f8ebc989eeb2a8f38242aa671bf0238b5577322da97486e429a7f3d816ba7\": container with ID starting with f95f8ebc989eeb2a8f38242aa671bf0238b5577322da97486e429a7f3d816ba7 not found: ID does not exist" containerID="f95f8ebc989eeb2a8f38242aa671bf0238b5577322da97486e429a7f3d816ba7" Nov 22 08:55:03 crc kubenswrapper[4789]: I1122 08:55:03.178293 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f95f8ebc989eeb2a8f38242aa671bf0238b5577322da97486e429a7f3d816ba7"} err="failed to get container status \"f95f8ebc989eeb2a8f38242aa671bf0238b5577322da97486e429a7f3d816ba7\": rpc error: code = NotFound desc = could not find container \"f95f8ebc989eeb2a8f38242aa671bf0238b5577322da97486e429a7f3d816ba7\": container with ID starting with f95f8ebc989eeb2a8f38242aa671bf0238b5577322da97486e429a7f3d816ba7 not found: ID does not exist" Nov 22 08:55:03 crc kubenswrapper[4789]: I1122 08:55:03.178345 4789 scope.go:117] "RemoveContainer" containerID="7b9e2fb24cb3bdbe2c862c83a6e5622f9df816ff654ec2e5c97c0efa9c7d81fe" Nov 22 08:55:03 crc kubenswrapper[4789]: E1122 08:55:03.178893 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7b9e2fb24cb3bdbe2c862c83a6e5622f9df816ff654ec2e5c97c0efa9c7d81fe\": container with ID starting with 7b9e2fb24cb3bdbe2c862c83a6e5622f9df816ff654ec2e5c97c0efa9c7d81fe not found: ID does not exist" containerID="7b9e2fb24cb3bdbe2c862c83a6e5622f9df816ff654ec2e5c97c0efa9c7d81fe" Nov 22 08:55:03 crc kubenswrapper[4789]: I1122 08:55:03.178993 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b9e2fb24cb3bdbe2c862c83a6e5622f9df816ff654ec2e5c97c0efa9c7d81fe"} err="failed to get container status \"7b9e2fb24cb3bdbe2c862c83a6e5622f9df816ff654ec2e5c97c0efa9c7d81fe\": rpc error: code = NotFound desc = could not find container \"7b9e2fb24cb3bdbe2c862c83a6e5622f9df816ff654ec2e5c97c0efa9c7d81fe\": container with ID starting with 7b9e2fb24cb3bdbe2c862c83a6e5622f9df816ff654ec2e5c97c0efa9c7d81fe not found: ID does not exist" Nov 22 08:55:03 crc kubenswrapper[4789]: I1122 08:55:03.179034 4789 scope.go:117] "RemoveContainer" containerID="5878e41eb2b9dda2cbf8d65523f932bcca37839be0dd0a2a78a79c8e9cca90a9" Nov 22 08:55:03 crc kubenswrapper[4789]: E1122 08:55:03.179394 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5878e41eb2b9dda2cbf8d65523f932bcca37839be0dd0a2a78a79c8e9cca90a9\": container with ID starting with 5878e41eb2b9dda2cbf8d65523f932bcca37839be0dd0a2a78a79c8e9cca90a9 not found: ID does not exist" containerID="5878e41eb2b9dda2cbf8d65523f932bcca37839be0dd0a2a78a79c8e9cca90a9" Nov 22 08:55:03 crc kubenswrapper[4789]: I1122 08:55:03.179439 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5878e41eb2b9dda2cbf8d65523f932bcca37839be0dd0a2a78a79c8e9cca90a9"} err="failed to get container status \"5878e41eb2b9dda2cbf8d65523f932bcca37839be0dd0a2a78a79c8e9cca90a9\": rpc error: code = NotFound desc = could not find container \"5878e41eb2b9dda2cbf8d65523f932bcca37839be0dd0a2a78a79c8e9cca90a9\": container with ID starting with 5878e41eb2b9dda2cbf8d65523f932bcca37839be0dd0a2a78a79c8e9cca90a9 not found: ID does not exist" Nov 22 08:55:04 crc kubenswrapper[4789]: I1122 08:55:04.012388 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f44ffc1-1437-4e0a-b083-2f14ab9f89f1" path="/var/lib/kubelet/pods/8f44ffc1-1437-4e0a-b083-2f14ab9f89f1/volumes" Nov 22 08:55:37 crc kubenswrapper[4789]: I1122 08:55:37.273292 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-55c966b9c6-psmvq"] Nov 22 08:55:37 crc kubenswrapper[4789]: E1122 08:55:37.274178 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f44ffc1-1437-4e0a-b083-2f14ab9f89f1" containerName="registry-server" Nov 22 08:55:37 crc kubenswrapper[4789]: I1122 08:55:37.274191 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f44ffc1-1437-4e0a-b083-2f14ab9f89f1" containerName="registry-server" Nov 22 08:55:37 crc kubenswrapper[4789]: E1122 08:55:37.274206 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f44ffc1-1437-4e0a-b083-2f14ab9f89f1" containerName="extract-utilities" Nov 22 08:55:37 crc kubenswrapper[4789]: I1122 08:55:37.274212 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f44ffc1-1437-4e0a-b083-2f14ab9f89f1" containerName="extract-utilities" Nov 22 08:55:37 crc kubenswrapper[4789]: E1122 08:55:37.274222 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f44ffc1-1437-4e0a-b083-2f14ab9f89f1" containerName="extract-content" Nov 22 08:55:37 crc kubenswrapper[4789]: I1122 08:55:37.274231 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f44ffc1-1437-4e0a-b083-2f14ab9f89f1" containerName="extract-content" Nov 22 08:55:37 crc kubenswrapper[4789]: I1122 08:55:37.274414 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f44ffc1-1437-4e0a-b083-2f14ab9f89f1" containerName="registry-server" Nov 22 08:55:37 crc kubenswrapper[4789]: I1122 08:55:37.275689 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-55c966b9c6-psmvq" Nov 22 08:55:37 crc kubenswrapper[4789]: I1122 08:55:37.311925 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-55c966b9c6-psmvq"] Nov 22 08:55:37 crc kubenswrapper[4789]: I1122 08:55:37.348389 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cf88n\" (UniqueName: \"kubernetes.io/projected/5729bc29-a8c9-4e1c-93fe-4f1532823cb9-kube-api-access-cf88n\") pod \"openstack-operator-controller-operator-55c966b9c6-psmvq\" (UID: \"5729bc29-a8c9-4e1c-93fe-4f1532823cb9\") " pod="openstack-operators/openstack-operator-controller-operator-55c966b9c6-psmvq" Nov 22 08:55:37 crc kubenswrapper[4789]: I1122 08:55:37.450348 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cf88n\" (UniqueName: \"kubernetes.io/projected/5729bc29-a8c9-4e1c-93fe-4f1532823cb9-kube-api-access-cf88n\") pod \"openstack-operator-controller-operator-55c966b9c6-psmvq\" (UID: \"5729bc29-a8c9-4e1c-93fe-4f1532823cb9\") " pod="openstack-operators/openstack-operator-controller-operator-55c966b9c6-psmvq" Nov 22 08:55:37 crc kubenswrapper[4789]: I1122 08:55:37.468767 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cf88n\" (UniqueName: \"kubernetes.io/projected/5729bc29-a8c9-4e1c-93fe-4f1532823cb9-kube-api-access-cf88n\") pod \"openstack-operator-controller-operator-55c966b9c6-psmvq\" (UID: \"5729bc29-a8c9-4e1c-93fe-4f1532823cb9\") " pod="openstack-operators/openstack-operator-controller-operator-55c966b9c6-psmvq" Nov 22 08:55:37 crc kubenswrapper[4789]: I1122 08:55:37.596646 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-55c966b9c6-psmvq" Nov 22 08:55:38 crc kubenswrapper[4789]: I1122 08:55:38.047112 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-55c966b9c6-psmvq"] Nov 22 08:55:38 crc kubenswrapper[4789]: I1122 08:55:38.401471 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-55c966b9c6-psmvq" event={"ID":"5729bc29-a8c9-4e1c-93fe-4f1532823cb9","Type":"ContainerStarted","Data":"e8958db07ba396b56d29c9e6c5e60c8d9583be811f79bbb5882a6b9a2d20d39c"} Nov 22 08:55:38 crc kubenswrapper[4789]: I1122 08:55:38.401533 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-55c966b9c6-psmvq" event={"ID":"5729bc29-a8c9-4e1c-93fe-4f1532823cb9","Type":"ContainerStarted","Data":"fcb196d199e217ab38be2c6624f1cffcf6b0ce366fd96182bc5dfcd8f5a848dd"} Nov 22 08:55:39 crc kubenswrapper[4789]: I1122 08:55:39.414634 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-55c966b9c6-psmvq" event={"ID":"5729bc29-a8c9-4e1c-93fe-4f1532823cb9","Type":"ContainerStarted","Data":"ff1f63894abcc04f16a1f1857496df3fa186a500ff89728a482c232debae60a5"} Nov 22 08:55:39 crc kubenswrapper[4789]: I1122 08:55:39.416182 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-55c966b9c6-psmvq" Nov 22 08:55:39 crc kubenswrapper[4789]: I1122 08:55:39.455612 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-55c966b9c6-psmvq" podStartSLOduration=2.455588124 podStartE2EDuration="2.455588124s" podCreationTimestamp="2025-11-22 08:55:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 08:55:39.451284978 +0000 UTC m=+3553.685685261" watchObservedRunningTime="2025-11-22 08:55:39.455588124 +0000 UTC m=+3553.689988397" Nov 22 08:55:47 crc kubenswrapper[4789]: I1122 08:55:47.599097 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-55c966b9c6-psmvq" Nov 22 08:55:47 crc kubenswrapper[4789]: I1122 08:55:47.689112 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-8486c7f98b-7vhvn"] Nov 22 08:55:47 crc kubenswrapper[4789]: I1122 08:55:47.689436 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-controller-operator-8486c7f98b-7vhvn" podUID="7c594182-6941-4a46-aebc-0d9df403eec6" containerName="operator" containerID="cri-o://e3aefedff2f08c827ef8f118d9bba35b44fc182a50185b3711cc452702de7628" gracePeriod=10 Nov 22 08:55:47 crc kubenswrapper[4789]: I1122 08:55:47.689544 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-controller-operator-8486c7f98b-7vhvn" podUID="7c594182-6941-4a46-aebc-0d9df403eec6" containerName="kube-rbac-proxy" containerID="cri-o://47f0e289a486d69c522c755790ec9b35b33b2f34d5578e503a7442c5aa0ef0cf" gracePeriod=10 Nov 22 08:55:48 crc kubenswrapper[4789]: I1122 08:55:48.169602 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-8486c7f98b-7vhvn" Nov 22 08:55:48 crc kubenswrapper[4789]: I1122 08:55:48.274559 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2dqm\" (UniqueName: \"kubernetes.io/projected/7c594182-6941-4a46-aebc-0d9df403eec6-kube-api-access-x2dqm\") pod \"7c594182-6941-4a46-aebc-0d9df403eec6\" (UID: \"7c594182-6941-4a46-aebc-0d9df403eec6\") " Nov 22 08:55:48 crc kubenswrapper[4789]: I1122 08:55:48.282672 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c594182-6941-4a46-aebc-0d9df403eec6-kube-api-access-x2dqm" (OuterVolumeSpecName: "kube-api-access-x2dqm") pod "7c594182-6941-4a46-aebc-0d9df403eec6" (UID: "7c594182-6941-4a46-aebc-0d9df403eec6"). InnerVolumeSpecName "kube-api-access-x2dqm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:55:48 crc kubenswrapper[4789]: I1122 08:55:48.377297 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2dqm\" (UniqueName: \"kubernetes.io/projected/7c594182-6941-4a46-aebc-0d9df403eec6-kube-api-access-x2dqm\") on node \"crc\" DevicePath \"\"" Nov 22 08:55:48 crc kubenswrapper[4789]: I1122 08:55:48.496743 4789 generic.go:334] "Generic (PLEG): container finished" podID="7c594182-6941-4a46-aebc-0d9df403eec6" containerID="47f0e289a486d69c522c755790ec9b35b33b2f34d5578e503a7442c5aa0ef0cf" exitCode=0 Nov 22 08:55:48 crc kubenswrapper[4789]: I1122 08:55:48.496803 4789 generic.go:334] "Generic (PLEG): container finished" podID="7c594182-6941-4a46-aebc-0d9df403eec6" containerID="e3aefedff2f08c827ef8f118d9bba35b44fc182a50185b3711cc452702de7628" exitCode=0 Nov 22 08:55:48 crc kubenswrapper[4789]: I1122 08:55:48.496835 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-8486c7f98b-7vhvn" Nov 22 08:55:48 crc kubenswrapper[4789]: I1122 08:55:48.496826 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-8486c7f98b-7vhvn" event={"ID":"7c594182-6941-4a46-aebc-0d9df403eec6","Type":"ContainerDied","Data":"47f0e289a486d69c522c755790ec9b35b33b2f34d5578e503a7442c5aa0ef0cf"} Nov 22 08:55:48 crc kubenswrapper[4789]: I1122 08:55:48.496978 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-8486c7f98b-7vhvn" event={"ID":"7c594182-6941-4a46-aebc-0d9df403eec6","Type":"ContainerDied","Data":"e3aefedff2f08c827ef8f118d9bba35b44fc182a50185b3711cc452702de7628"} Nov 22 08:55:48 crc kubenswrapper[4789]: I1122 08:55:48.496999 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-8486c7f98b-7vhvn" event={"ID":"7c594182-6941-4a46-aebc-0d9df403eec6","Type":"ContainerDied","Data":"68124dd5ee4b58ee852f7e7890efdf3e3ba3a6e019ae7ad3dc9d4b46dbfc14b2"} Nov 22 08:55:48 crc kubenswrapper[4789]: I1122 08:55:48.497019 4789 scope.go:117] "RemoveContainer" containerID="47f0e289a486d69c522c755790ec9b35b33b2f34d5578e503a7442c5aa0ef0cf" Nov 22 08:55:48 crc kubenswrapper[4789]: I1122 08:55:48.536494 4789 scope.go:117] "RemoveContainer" containerID="e3aefedff2f08c827ef8f118d9bba35b44fc182a50185b3711cc452702de7628" Nov 22 08:55:48 crc kubenswrapper[4789]: I1122 08:55:48.539144 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-8486c7f98b-7vhvn"] Nov 22 08:55:48 crc kubenswrapper[4789]: I1122 08:55:48.547013 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-8486c7f98b-7vhvn"] Nov 22 08:55:48 crc kubenswrapper[4789]: I1122 08:55:48.559152 4789 scope.go:117] "RemoveContainer" containerID="47f0e289a486d69c522c755790ec9b35b33b2f34d5578e503a7442c5aa0ef0cf" Nov 22 08:55:48 crc kubenswrapper[4789]: E1122 08:55:48.559608 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"47f0e289a486d69c522c755790ec9b35b33b2f34d5578e503a7442c5aa0ef0cf\": container with ID starting with 47f0e289a486d69c522c755790ec9b35b33b2f34d5578e503a7442c5aa0ef0cf not found: ID does not exist" containerID="47f0e289a486d69c522c755790ec9b35b33b2f34d5578e503a7442c5aa0ef0cf" Nov 22 08:55:48 crc kubenswrapper[4789]: I1122 08:55:48.559650 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"47f0e289a486d69c522c755790ec9b35b33b2f34d5578e503a7442c5aa0ef0cf"} err="failed to get container status \"47f0e289a486d69c522c755790ec9b35b33b2f34d5578e503a7442c5aa0ef0cf\": rpc error: code = NotFound desc = could not find container \"47f0e289a486d69c522c755790ec9b35b33b2f34d5578e503a7442c5aa0ef0cf\": container with ID starting with 47f0e289a486d69c522c755790ec9b35b33b2f34d5578e503a7442c5aa0ef0cf not found: ID does not exist" Nov 22 08:55:48 crc kubenswrapper[4789]: I1122 08:55:48.559677 4789 scope.go:117] "RemoveContainer" containerID="e3aefedff2f08c827ef8f118d9bba35b44fc182a50185b3711cc452702de7628" Nov 22 08:55:48 crc kubenswrapper[4789]: E1122 08:55:48.559977 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e3aefedff2f08c827ef8f118d9bba35b44fc182a50185b3711cc452702de7628\": container with ID starting with e3aefedff2f08c827ef8f118d9bba35b44fc182a50185b3711cc452702de7628 not found: ID does not exist" containerID="e3aefedff2f08c827ef8f118d9bba35b44fc182a50185b3711cc452702de7628" Nov 22 08:55:48 crc kubenswrapper[4789]: I1122 08:55:48.560007 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e3aefedff2f08c827ef8f118d9bba35b44fc182a50185b3711cc452702de7628"} err="failed to get container status \"e3aefedff2f08c827ef8f118d9bba35b44fc182a50185b3711cc452702de7628\": rpc error: code = NotFound desc = could not find container \"e3aefedff2f08c827ef8f118d9bba35b44fc182a50185b3711cc452702de7628\": container with ID starting with e3aefedff2f08c827ef8f118d9bba35b44fc182a50185b3711cc452702de7628 not found: ID does not exist" Nov 22 08:55:48 crc kubenswrapper[4789]: I1122 08:55:48.560026 4789 scope.go:117] "RemoveContainer" containerID="47f0e289a486d69c522c755790ec9b35b33b2f34d5578e503a7442c5aa0ef0cf" Nov 22 08:55:48 crc kubenswrapper[4789]: I1122 08:55:48.560268 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"47f0e289a486d69c522c755790ec9b35b33b2f34d5578e503a7442c5aa0ef0cf"} err="failed to get container status \"47f0e289a486d69c522c755790ec9b35b33b2f34d5578e503a7442c5aa0ef0cf\": rpc error: code = NotFound desc = could not find container \"47f0e289a486d69c522c755790ec9b35b33b2f34d5578e503a7442c5aa0ef0cf\": container with ID starting with 47f0e289a486d69c522c755790ec9b35b33b2f34d5578e503a7442c5aa0ef0cf not found: ID does not exist" Nov 22 08:55:48 crc kubenswrapper[4789]: I1122 08:55:48.560287 4789 scope.go:117] "RemoveContainer" containerID="e3aefedff2f08c827ef8f118d9bba35b44fc182a50185b3711cc452702de7628" Nov 22 08:55:48 crc kubenswrapper[4789]: I1122 08:55:48.560602 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e3aefedff2f08c827ef8f118d9bba35b44fc182a50185b3711cc452702de7628"} err="failed to get container status \"e3aefedff2f08c827ef8f118d9bba35b44fc182a50185b3711cc452702de7628\": rpc error: code = NotFound desc = could not find container \"e3aefedff2f08c827ef8f118d9bba35b44fc182a50185b3711cc452702de7628\": container with ID starting with e3aefedff2f08c827ef8f118d9bba35b44fc182a50185b3711cc452702de7628 not found: ID does not exist" Nov 22 08:55:50 crc kubenswrapper[4789]: I1122 08:55:50.001981 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7c594182-6941-4a46-aebc-0d9df403eec6" path="/var/lib/kubelet/pods/7c594182-6941-4a46-aebc-0d9df403eec6/volumes" Nov 22 08:56:00 crc kubenswrapper[4789]: I1122 08:56:00.062648 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-kjt58"] Nov 22 08:56:00 crc kubenswrapper[4789]: E1122 08:56:00.068812 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c594182-6941-4a46-aebc-0d9df403eec6" containerName="kube-rbac-proxy" Nov 22 08:56:00 crc kubenswrapper[4789]: I1122 08:56:00.068893 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c594182-6941-4a46-aebc-0d9df403eec6" containerName="kube-rbac-proxy" Nov 22 08:56:00 crc kubenswrapper[4789]: E1122 08:56:00.068932 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c594182-6941-4a46-aebc-0d9df403eec6" containerName="operator" Nov 22 08:56:00 crc kubenswrapper[4789]: I1122 08:56:00.068985 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c594182-6941-4a46-aebc-0d9df403eec6" containerName="operator" Nov 22 08:56:00 crc kubenswrapper[4789]: I1122 08:56:00.071552 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c594182-6941-4a46-aebc-0d9df403eec6" containerName="operator" Nov 22 08:56:00 crc kubenswrapper[4789]: I1122 08:56:00.071644 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c594182-6941-4a46-aebc-0d9df403eec6" containerName="kube-rbac-proxy" Nov 22 08:56:00 crc kubenswrapper[4789]: I1122 08:56:00.076228 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kjt58" Nov 22 08:56:00 crc kubenswrapper[4789]: I1122 08:56:00.084047 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kjt58"] Nov 22 08:56:00 crc kubenswrapper[4789]: I1122 08:56:00.109314 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca6db9c4-016b-4f90-9c87-a18b290f5ed3-catalog-content\") pod \"certified-operators-kjt58\" (UID: \"ca6db9c4-016b-4f90-9c87-a18b290f5ed3\") " pod="openshift-marketplace/certified-operators-kjt58" Nov 22 08:56:00 crc kubenswrapper[4789]: I1122 08:56:00.109738 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca6db9c4-016b-4f90-9c87-a18b290f5ed3-utilities\") pod \"certified-operators-kjt58\" (UID: \"ca6db9c4-016b-4f90-9c87-a18b290f5ed3\") " pod="openshift-marketplace/certified-operators-kjt58" Nov 22 08:56:00 crc kubenswrapper[4789]: I1122 08:56:00.109915 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kwqbc\" (UniqueName: \"kubernetes.io/projected/ca6db9c4-016b-4f90-9c87-a18b290f5ed3-kube-api-access-kwqbc\") pod \"certified-operators-kjt58\" (UID: \"ca6db9c4-016b-4f90-9c87-a18b290f5ed3\") " pod="openshift-marketplace/certified-operators-kjt58" Nov 22 08:56:00 crc kubenswrapper[4789]: I1122 08:56:00.212014 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca6db9c4-016b-4f90-9c87-a18b290f5ed3-utilities\") pod \"certified-operators-kjt58\" (UID: \"ca6db9c4-016b-4f90-9c87-a18b290f5ed3\") " pod="openshift-marketplace/certified-operators-kjt58" Nov 22 08:56:00 crc kubenswrapper[4789]: I1122 08:56:00.212172 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kwqbc\" (UniqueName: \"kubernetes.io/projected/ca6db9c4-016b-4f90-9c87-a18b290f5ed3-kube-api-access-kwqbc\") pod \"certified-operators-kjt58\" (UID: \"ca6db9c4-016b-4f90-9c87-a18b290f5ed3\") " pod="openshift-marketplace/certified-operators-kjt58" Nov 22 08:56:00 crc kubenswrapper[4789]: I1122 08:56:00.212296 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca6db9c4-016b-4f90-9c87-a18b290f5ed3-catalog-content\") pod \"certified-operators-kjt58\" (UID: \"ca6db9c4-016b-4f90-9c87-a18b290f5ed3\") " pod="openshift-marketplace/certified-operators-kjt58" Nov 22 08:56:00 crc kubenswrapper[4789]: I1122 08:56:00.212623 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca6db9c4-016b-4f90-9c87-a18b290f5ed3-utilities\") pod \"certified-operators-kjt58\" (UID: \"ca6db9c4-016b-4f90-9c87-a18b290f5ed3\") " pod="openshift-marketplace/certified-operators-kjt58" Nov 22 08:56:00 crc kubenswrapper[4789]: I1122 08:56:00.212741 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca6db9c4-016b-4f90-9c87-a18b290f5ed3-catalog-content\") pod \"certified-operators-kjt58\" (UID: \"ca6db9c4-016b-4f90-9c87-a18b290f5ed3\") " pod="openshift-marketplace/certified-operators-kjt58" Nov 22 08:56:00 crc kubenswrapper[4789]: I1122 08:56:00.234454 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kwqbc\" (UniqueName: \"kubernetes.io/projected/ca6db9c4-016b-4f90-9c87-a18b290f5ed3-kube-api-access-kwqbc\") pod \"certified-operators-kjt58\" (UID: \"ca6db9c4-016b-4f90-9c87-a18b290f5ed3\") " pod="openshift-marketplace/certified-operators-kjt58" Nov 22 08:56:00 crc kubenswrapper[4789]: I1122 08:56:00.410402 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kjt58" Nov 22 08:56:00 crc kubenswrapper[4789]: W1122 08:56:00.942656 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podca6db9c4_016b_4f90_9c87_a18b290f5ed3.slice/crio-4eeb9996738ae2273432e4aa6bdd464fa65d429984ed6cf81ebf8625f7398d35 WatchSource:0}: Error finding container 4eeb9996738ae2273432e4aa6bdd464fa65d429984ed6cf81ebf8625f7398d35: Status 404 returned error can't find the container with id 4eeb9996738ae2273432e4aa6bdd464fa65d429984ed6cf81ebf8625f7398d35 Nov 22 08:56:00 crc kubenswrapper[4789]: I1122 08:56:00.944202 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kjt58"] Nov 22 08:56:01 crc kubenswrapper[4789]: I1122 08:56:01.618602 4789 generic.go:334] "Generic (PLEG): container finished" podID="ca6db9c4-016b-4f90-9c87-a18b290f5ed3" containerID="a81fc85e7a9ace7d383a7acfd0e53e41a17d45f413ee0bbc020cc9eb4ea14128" exitCode=0 Nov 22 08:56:01 crc kubenswrapper[4789]: I1122 08:56:01.618699 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kjt58" event={"ID":"ca6db9c4-016b-4f90-9c87-a18b290f5ed3","Type":"ContainerDied","Data":"a81fc85e7a9ace7d383a7acfd0e53e41a17d45f413ee0bbc020cc9eb4ea14128"} Nov 22 08:56:01 crc kubenswrapper[4789]: I1122 08:56:01.618892 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kjt58" event={"ID":"ca6db9c4-016b-4f90-9c87-a18b290f5ed3","Type":"ContainerStarted","Data":"4eeb9996738ae2273432e4aa6bdd464fa65d429984ed6cf81ebf8625f7398d35"} Nov 22 08:56:03 crc kubenswrapper[4789]: I1122 08:56:03.637195 4789 generic.go:334] "Generic (PLEG): container finished" podID="ca6db9c4-016b-4f90-9c87-a18b290f5ed3" containerID="e4a30edff2e5a606000160727108b984229aaba0aae696dc399b15184f47a237" exitCode=0 Nov 22 08:56:03 crc kubenswrapper[4789]: I1122 08:56:03.637259 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kjt58" event={"ID":"ca6db9c4-016b-4f90-9c87-a18b290f5ed3","Type":"ContainerDied","Data":"e4a30edff2e5a606000160727108b984229aaba0aae696dc399b15184f47a237"} Nov 22 08:56:04 crc kubenswrapper[4789]: I1122 08:56:04.649046 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kjt58" event={"ID":"ca6db9c4-016b-4f90-9c87-a18b290f5ed3","Type":"ContainerStarted","Data":"4f3c3860cf576299c53fea41e957c794029016071567ebf5017803cd17ae11f0"} Nov 22 08:56:04 crc kubenswrapper[4789]: I1122 08:56:04.675114 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-kjt58" podStartSLOduration=2.229601512 podStartE2EDuration="4.675092311s" podCreationTimestamp="2025-11-22 08:56:00 +0000 UTC" firstStartedPulling="2025-11-22 08:56:01.620866883 +0000 UTC m=+3575.855267156" lastFinishedPulling="2025-11-22 08:56:04.066357682 +0000 UTC m=+3578.300757955" observedRunningTime="2025-11-22 08:56:04.667423054 +0000 UTC m=+3578.901823347" watchObservedRunningTime="2025-11-22 08:56:04.675092311 +0000 UTC m=+3578.909492574" Nov 22 08:56:05 crc kubenswrapper[4789]: I1122 08:56:05.372190 4789 patch_prober.go:28] interesting pod/machine-config-daemon-pxvrm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 08:56:05 crc kubenswrapper[4789]: I1122 08:56:05.372282 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 08:56:10 crc kubenswrapper[4789]: I1122 08:56:10.410715 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-kjt58" Nov 22 08:56:10 crc kubenswrapper[4789]: I1122 08:56:10.411322 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-kjt58" Nov 22 08:56:10 crc kubenswrapper[4789]: I1122 08:56:10.458375 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-kjt58" Nov 22 08:56:10 crc kubenswrapper[4789]: I1122 08:56:10.740396 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-kjt58" Nov 22 08:56:10 crc kubenswrapper[4789]: I1122 08:56:10.787866 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kjt58"] Nov 22 08:56:12 crc kubenswrapper[4789]: I1122 08:56:12.712496 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-kjt58" podUID="ca6db9c4-016b-4f90-9c87-a18b290f5ed3" containerName="registry-server" containerID="cri-o://4f3c3860cf576299c53fea41e957c794029016071567ebf5017803cd17ae11f0" gracePeriod=2 Nov 22 08:56:13 crc kubenswrapper[4789]: I1122 08:56:13.137472 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kjt58" Nov 22 08:56:13 crc kubenswrapper[4789]: I1122 08:56:13.281401 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca6db9c4-016b-4f90-9c87-a18b290f5ed3-utilities\") pod \"ca6db9c4-016b-4f90-9c87-a18b290f5ed3\" (UID: \"ca6db9c4-016b-4f90-9c87-a18b290f5ed3\") " Nov 22 08:56:13 crc kubenswrapper[4789]: I1122 08:56:13.281568 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca6db9c4-016b-4f90-9c87-a18b290f5ed3-catalog-content\") pod \"ca6db9c4-016b-4f90-9c87-a18b290f5ed3\" (UID: \"ca6db9c4-016b-4f90-9c87-a18b290f5ed3\") " Nov 22 08:56:13 crc kubenswrapper[4789]: I1122 08:56:13.281660 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kwqbc\" (UniqueName: \"kubernetes.io/projected/ca6db9c4-016b-4f90-9c87-a18b290f5ed3-kube-api-access-kwqbc\") pod \"ca6db9c4-016b-4f90-9c87-a18b290f5ed3\" (UID: \"ca6db9c4-016b-4f90-9c87-a18b290f5ed3\") " Nov 22 08:56:13 crc kubenswrapper[4789]: I1122 08:56:13.282292 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ca6db9c4-016b-4f90-9c87-a18b290f5ed3-utilities" (OuterVolumeSpecName: "utilities") pod "ca6db9c4-016b-4f90-9c87-a18b290f5ed3" (UID: "ca6db9c4-016b-4f90-9c87-a18b290f5ed3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:56:13 crc kubenswrapper[4789]: I1122 08:56:13.288541 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca6db9c4-016b-4f90-9c87-a18b290f5ed3-kube-api-access-kwqbc" (OuterVolumeSpecName: "kube-api-access-kwqbc") pod "ca6db9c4-016b-4f90-9c87-a18b290f5ed3" (UID: "ca6db9c4-016b-4f90-9c87-a18b290f5ed3"). InnerVolumeSpecName "kube-api-access-kwqbc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:56:13 crc kubenswrapper[4789]: I1122 08:56:13.319364 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ca6db9c4-016b-4f90-9c87-a18b290f5ed3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ca6db9c4-016b-4f90-9c87-a18b290f5ed3" (UID: "ca6db9c4-016b-4f90-9c87-a18b290f5ed3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 08:56:13 crc kubenswrapper[4789]: I1122 08:56:13.383742 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca6db9c4-016b-4f90-9c87-a18b290f5ed3-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 08:56:13 crc kubenswrapper[4789]: I1122 08:56:13.383883 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kwqbc\" (UniqueName: \"kubernetes.io/projected/ca6db9c4-016b-4f90-9c87-a18b290f5ed3-kube-api-access-kwqbc\") on node \"crc\" DevicePath \"\"" Nov 22 08:56:13 crc kubenswrapper[4789]: I1122 08:56:13.383936 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca6db9c4-016b-4f90-9c87-a18b290f5ed3-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 08:56:13 crc kubenswrapper[4789]: I1122 08:56:13.725218 4789 generic.go:334] "Generic (PLEG): container finished" podID="ca6db9c4-016b-4f90-9c87-a18b290f5ed3" containerID="4f3c3860cf576299c53fea41e957c794029016071567ebf5017803cd17ae11f0" exitCode=0 Nov 22 08:56:13 crc kubenswrapper[4789]: I1122 08:56:13.725290 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kjt58" event={"ID":"ca6db9c4-016b-4f90-9c87-a18b290f5ed3","Type":"ContainerDied","Data":"4f3c3860cf576299c53fea41e957c794029016071567ebf5017803cd17ae11f0"} Nov 22 08:56:13 crc kubenswrapper[4789]: I1122 08:56:13.725308 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kjt58" Nov 22 08:56:13 crc kubenswrapper[4789]: I1122 08:56:13.725673 4789 scope.go:117] "RemoveContainer" containerID="4f3c3860cf576299c53fea41e957c794029016071567ebf5017803cd17ae11f0" Nov 22 08:56:13 crc kubenswrapper[4789]: I1122 08:56:13.725638 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kjt58" event={"ID":"ca6db9c4-016b-4f90-9c87-a18b290f5ed3","Type":"ContainerDied","Data":"4eeb9996738ae2273432e4aa6bdd464fa65d429984ed6cf81ebf8625f7398d35"} Nov 22 08:56:13 crc kubenswrapper[4789]: I1122 08:56:13.757236 4789 scope.go:117] "RemoveContainer" containerID="e4a30edff2e5a606000160727108b984229aaba0aae696dc399b15184f47a237" Nov 22 08:56:13 crc kubenswrapper[4789]: I1122 08:56:13.765552 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kjt58"] Nov 22 08:56:13 crc kubenswrapper[4789]: I1122 08:56:13.778311 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-kjt58"] Nov 22 08:56:13 crc kubenswrapper[4789]: I1122 08:56:13.793966 4789 scope.go:117] "RemoveContainer" containerID="a81fc85e7a9ace7d383a7acfd0e53e41a17d45f413ee0bbc020cc9eb4ea14128" Nov 22 08:56:13 crc kubenswrapper[4789]: I1122 08:56:13.848043 4789 scope.go:117] "RemoveContainer" containerID="4f3c3860cf576299c53fea41e957c794029016071567ebf5017803cd17ae11f0" Nov 22 08:56:13 crc kubenswrapper[4789]: E1122 08:56:13.848650 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4f3c3860cf576299c53fea41e957c794029016071567ebf5017803cd17ae11f0\": container with ID starting with 4f3c3860cf576299c53fea41e957c794029016071567ebf5017803cd17ae11f0 not found: ID does not exist" containerID="4f3c3860cf576299c53fea41e957c794029016071567ebf5017803cd17ae11f0" Nov 22 08:56:13 crc kubenswrapper[4789]: I1122 08:56:13.848688 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4f3c3860cf576299c53fea41e957c794029016071567ebf5017803cd17ae11f0"} err="failed to get container status \"4f3c3860cf576299c53fea41e957c794029016071567ebf5017803cd17ae11f0\": rpc error: code = NotFound desc = could not find container \"4f3c3860cf576299c53fea41e957c794029016071567ebf5017803cd17ae11f0\": container with ID starting with 4f3c3860cf576299c53fea41e957c794029016071567ebf5017803cd17ae11f0 not found: ID does not exist" Nov 22 08:56:13 crc kubenswrapper[4789]: I1122 08:56:13.848717 4789 scope.go:117] "RemoveContainer" containerID="e4a30edff2e5a606000160727108b984229aaba0aae696dc399b15184f47a237" Nov 22 08:56:13 crc kubenswrapper[4789]: E1122 08:56:13.849220 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e4a30edff2e5a606000160727108b984229aaba0aae696dc399b15184f47a237\": container with ID starting with e4a30edff2e5a606000160727108b984229aaba0aae696dc399b15184f47a237 not found: ID does not exist" containerID="e4a30edff2e5a606000160727108b984229aaba0aae696dc399b15184f47a237" Nov 22 08:56:13 crc kubenswrapper[4789]: I1122 08:56:13.849254 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e4a30edff2e5a606000160727108b984229aaba0aae696dc399b15184f47a237"} err="failed to get container status \"e4a30edff2e5a606000160727108b984229aaba0aae696dc399b15184f47a237\": rpc error: code = NotFound desc = could not find container \"e4a30edff2e5a606000160727108b984229aaba0aae696dc399b15184f47a237\": container with ID starting with e4a30edff2e5a606000160727108b984229aaba0aae696dc399b15184f47a237 not found: ID does not exist" Nov 22 08:56:13 crc kubenswrapper[4789]: I1122 08:56:13.849275 4789 scope.go:117] "RemoveContainer" containerID="a81fc85e7a9ace7d383a7acfd0e53e41a17d45f413ee0bbc020cc9eb4ea14128" Nov 22 08:56:13 crc kubenswrapper[4789]: E1122 08:56:13.849657 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a81fc85e7a9ace7d383a7acfd0e53e41a17d45f413ee0bbc020cc9eb4ea14128\": container with ID starting with a81fc85e7a9ace7d383a7acfd0e53e41a17d45f413ee0bbc020cc9eb4ea14128 not found: ID does not exist" containerID="a81fc85e7a9ace7d383a7acfd0e53e41a17d45f413ee0bbc020cc9eb4ea14128" Nov 22 08:56:13 crc kubenswrapper[4789]: I1122 08:56:13.849697 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a81fc85e7a9ace7d383a7acfd0e53e41a17d45f413ee0bbc020cc9eb4ea14128"} err="failed to get container status \"a81fc85e7a9ace7d383a7acfd0e53e41a17d45f413ee0bbc020cc9eb4ea14128\": rpc error: code = NotFound desc = could not find container \"a81fc85e7a9ace7d383a7acfd0e53e41a17d45f413ee0bbc020cc9eb4ea14128\": container with ID starting with a81fc85e7a9ace7d383a7acfd0e53e41a17d45f413ee0bbc020cc9eb4ea14128 not found: ID does not exist" Nov 22 08:56:13 crc kubenswrapper[4789]: I1122 08:56:13.980104 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ca6db9c4-016b-4f90-9c87-a18b290f5ed3" path="/var/lib/kubelet/pods/ca6db9c4-016b-4f90-9c87-a18b290f5ed3/volumes" Nov 22 08:56:20 crc kubenswrapper[4789]: I1122 08:56:20.914939 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-7b6cddbf79-2d282"] Nov 22 08:56:20 crc kubenswrapper[4789]: E1122 08:56:20.915670 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca6db9c4-016b-4f90-9c87-a18b290f5ed3" containerName="extract-content" Nov 22 08:56:20 crc kubenswrapper[4789]: I1122 08:56:20.915681 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca6db9c4-016b-4f90-9c87-a18b290f5ed3" containerName="extract-content" Nov 22 08:56:20 crc kubenswrapper[4789]: E1122 08:56:20.915696 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca6db9c4-016b-4f90-9c87-a18b290f5ed3" containerName="registry-server" Nov 22 08:56:20 crc kubenswrapper[4789]: I1122 08:56:20.915703 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca6db9c4-016b-4f90-9c87-a18b290f5ed3" containerName="registry-server" Nov 22 08:56:20 crc kubenswrapper[4789]: E1122 08:56:20.915722 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca6db9c4-016b-4f90-9c87-a18b290f5ed3" containerName="extract-utilities" Nov 22 08:56:20 crc kubenswrapper[4789]: I1122 08:56:20.915729 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca6db9c4-016b-4f90-9c87-a18b290f5ed3" containerName="extract-utilities" Nov 22 08:56:20 crc kubenswrapper[4789]: I1122 08:56:20.915906 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca6db9c4-016b-4f90-9c87-a18b290f5ed3" containerName="registry-server" Nov 22 08:56:20 crc kubenswrapper[4789]: I1122 08:56:20.916797 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-7b6cddbf79-2d282" Nov 22 08:56:20 crc kubenswrapper[4789]: I1122 08:56:20.937930 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-7b6cddbf79-2d282"] Nov 22 08:56:21 crc kubenswrapper[4789]: I1122 08:56:21.035390 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6q742\" (UniqueName: \"kubernetes.io/projected/f3873dc8-add1-47e4-9fee-e1023782e6da-kube-api-access-6q742\") pod \"test-operator-controller-manager-7b6cddbf79-2d282\" (UID: \"f3873dc8-add1-47e4-9fee-e1023782e6da\") " pod="openstack-operators/test-operator-controller-manager-7b6cddbf79-2d282" Nov 22 08:56:21 crc kubenswrapper[4789]: I1122 08:56:21.137870 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6q742\" (UniqueName: \"kubernetes.io/projected/f3873dc8-add1-47e4-9fee-e1023782e6da-kube-api-access-6q742\") pod \"test-operator-controller-manager-7b6cddbf79-2d282\" (UID: \"f3873dc8-add1-47e4-9fee-e1023782e6da\") " pod="openstack-operators/test-operator-controller-manager-7b6cddbf79-2d282" Nov 22 08:56:21 crc kubenswrapper[4789]: I1122 08:56:21.159328 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6q742\" (UniqueName: \"kubernetes.io/projected/f3873dc8-add1-47e4-9fee-e1023782e6da-kube-api-access-6q742\") pod \"test-operator-controller-manager-7b6cddbf79-2d282\" (UID: \"f3873dc8-add1-47e4-9fee-e1023782e6da\") " pod="openstack-operators/test-operator-controller-manager-7b6cddbf79-2d282" Nov 22 08:56:21 crc kubenswrapper[4789]: I1122 08:56:21.249033 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-7b6cddbf79-2d282" Nov 22 08:56:21 crc kubenswrapper[4789]: I1122 08:56:21.692857 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-7b6cddbf79-2d282"] Nov 22 08:56:21 crc kubenswrapper[4789]: I1122 08:56:21.797390 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-7b6cddbf79-2d282" event={"ID":"f3873dc8-add1-47e4-9fee-e1023782e6da","Type":"ContainerStarted","Data":"8809db1c18e49cd01a12bd094baea5cc74522888c82fb233c8e1ca83482f4941"} Nov 22 08:56:23 crc kubenswrapper[4789]: I1122 08:56:23.840546 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-7b6cddbf79-2d282" event={"ID":"f3873dc8-add1-47e4-9fee-e1023782e6da","Type":"ContainerStarted","Data":"48bb02e2a06bfb3bca7fe4d54ad856cbc67b115f387ab2f870cffcec08c29f98"} Nov 22 08:56:23 crc kubenswrapper[4789]: I1122 08:56:23.841260 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-7b6cddbf79-2d282" event={"ID":"f3873dc8-add1-47e4-9fee-e1023782e6da","Type":"ContainerStarted","Data":"a9925bd7f51d1ff145b890c657e61dd1a9f48bbd5522fe71558cf54674f1f3a3"} Nov 22 08:56:23 crc kubenswrapper[4789]: I1122 08:56:23.841284 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-7b6cddbf79-2d282" Nov 22 08:56:23 crc kubenswrapper[4789]: I1122 08:56:23.866999 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-7b6cddbf79-2d282" podStartSLOduration=2.943045841 podStartE2EDuration="3.866969366s" podCreationTimestamp="2025-11-22 08:56:20 +0000 UTC" firstStartedPulling="2025-11-22 08:56:21.695223596 +0000 UTC m=+3595.929623869" lastFinishedPulling="2025-11-22 08:56:22.619147121 +0000 UTC m=+3596.853547394" observedRunningTime="2025-11-22 08:56:23.862940958 +0000 UTC m=+3598.097341321" watchObservedRunningTime="2025-11-22 08:56:23.866969366 +0000 UTC m=+3598.101369679" Nov 22 08:56:31 crc kubenswrapper[4789]: I1122 08:56:31.255030 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-7b6cddbf79-2d282" Nov 22 08:56:31 crc kubenswrapper[4789]: I1122 08:56:31.324582 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/test-operator-controller-manager-8464cf66df-hc8lc"] Nov 22 08:56:31 crc kubenswrapper[4789]: I1122 08:56:31.324991 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/test-operator-controller-manager-8464cf66df-hc8lc" podUID="50a6fd25-a186-4125-99c1-427d6214b39a" containerName="kube-rbac-proxy" containerID="cri-o://0a32a6879db01bf9780e16610fa1359f977fc0147852bffb86708470546ce377" gracePeriod=10 Nov 22 08:56:31 crc kubenswrapper[4789]: I1122 08:56:31.325095 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/test-operator-controller-manager-8464cf66df-hc8lc" podUID="50a6fd25-a186-4125-99c1-427d6214b39a" containerName="manager" containerID="cri-o://91b0dd2990b97f4ac5d5a357acb0573fd640d2da6303f19ca22866fb509567c8" gracePeriod=10 Nov 22 08:56:31 crc kubenswrapper[4789]: I1122 08:56:31.804524 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-8464cf66df-hc8lc" Nov 22 08:56:31 crc kubenswrapper[4789]: I1122 08:56:31.913869 4789 generic.go:334] "Generic (PLEG): container finished" podID="50a6fd25-a186-4125-99c1-427d6214b39a" containerID="91b0dd2990b97f4ac5d5a357acb0573fd640d2da6303f19ca22866fb509567c8" exitCode=0 Nov 22 08:56:31 crc kubenswrapper[4789]: I1122 08:56:31.913904 4789 generic.go:334] "Generic (PLEG): container finished" podID="50a6fd25-a186-4125-99c1-427d6214b39a" containerID="0a32a6879db01bf9780e16610fa1359f977fc0147852bffb86708470546ce377" exitCode=0 Nov 22 08:56:31 crc kubenswrapper[4789]: I1122 08:56:31.913919 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-8464cf66df-hc8lc" Nov 22 08:56:31 crc kubenswrapper[4789]: I1122 08:56:31.913932 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-8464cf66df-hc8lc" event={"ID":"50a6fd25-a186-4125-99c1-427d6214b39a","Type":"ContainerDied","Data":"91b0dd2990b97f4ac5d5a357acb0573fd640d2da6303f19ca22866fb509567c8"} Nov 22 08:56:31 crc kubenswrapper[4789]: I1122 08:56:31.913974 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-8464cf66df-hc8lc" event={"ID":"50a6fd25-a186-4125-99c1-427d6214b39a","Type":"ContainerDied","Data":"0a32a6879db01bf9780e16610fa1359f977fc0147852bffb86708470546ce377"} Nov 22 08:56:31 crc kubenswrapper[4789]: I1122 08:56:31.913991 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-8464cf66df-hc8lc" event={"ID":"50a6fd25-a186-4125-99c1-427d6214b39a","Type":"ContainerDied","Data":"6c28791a936835e3dee64c24283ca9b86f7a7e771c346802cf472cf75057407f"} Nov 22 08:56:31 crc kubenswrapper[4789]: I1122 08:56:31.914009 4789 scope.go:117] "RemoveContainer" containerID="91b0dd2990b97f4ac5d5a357acb0573fd640d2da6303f19ca22866fb509567c8" Nov 22 08:56:31 crc kubenswrapper[4789]: I1122 08:56:31.954246 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pwkjl\" (UniqueName: \"kubernetes.io/projected/50a6fd25-a186-4125-99c1-427d6214b39a-kube-api-access-pwkjl\") pod \"50a6fd25-a186-4125-99c1-427d6214b39a\" (UID: \"50a6fd25-a186-4125-99c1-427d6214b39a\") " Nov 22 08:56:32 crc kubenswrapper[4789]: I1122 08:56:32.014065 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/50a6fd25-a186-4125-99c1-427d6214b39a-kube-api-access-pwkjl" (OuterVolumeSpecName: "kube-api-access-pwkjl") pod "50a6fd25-a186-4125-99c1-427d6214b39a" (UID: "50a6fd25-a186-4125-99c1-427d6214b39a"). InnerVolumeSpecName "kube-api-access-pwkjl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 08:56:32 crc kubenswrapper[4789]: I1122 08:56:32.026950 4789 scope.go:117] "RemoveContainer" containerID="0a32a6879db01bf9780e16610fa1359f977fc0147852bffb86708470546ce377" Nov 22 08:56:32 crc kubenswrapper[4789]: I1122 08:56:32.049510 4789 scope.go:117] "RemoveContainer" containerID="91b0dd2990b97f4ac5d5a357acb0573fd640d2da6303f19ca22866fb509567c8" Nov 22 08:56:32 crc kubenswrapper[4789]: E1122 08:56:32.050080 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"91b0dd2990b97f4ac5d5a357acb0573fd640d2da6303f19ca22866fb509567c8\": container with ID starting with 91b0dd2990b97f4ac5d5a357acb0573fd640d2da6303f19ca22866fb509567c8 not found: ID does not exist" containerID="91b0dd2990b97f4ac5d5a357acb0573fd640d2da6303f19ca22866fb509567c8" Nov 22 08:56:32 crc kubenswrapper[4789]: I1122 08:56:32.050123 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91b0dd2990b97f4ac5d5a357acb0573fd640d2da6303f19ca22866fb509567c8"} err="failed to get container status \"91b0dd2990b97f4ac5d5a357acb0573fd640d2da6303f19ca22866fb509567c8\": rpc error: code = NotFound desc = could not find container \"91b0dd2990b97f4ac5d5a357acb0573fd640d2da6303f19ca22866fb509567c8\": container with ID starting with 91b0dd2990b97f4ac5d5a357acb0573fd640d2da6303f19ca22866fb509567c8 not found: ID does not exist" Nov 22 08:56:32 crc kubenswrapper[4789]: I1122 08:56:32.050151 4789 scope.go:117] "RemoveContainer" containerID="0a32a6879db01bf9780e16610fa1359f977fc0147852bffb86708470546ce377" Nov 22 08:56:32 crc kubenswrapper[4789]: E1122 08:56:32.050444 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0a32a6879db01bf9780e16610fa1359f977fc0147852bffb86708470546ce377\": container with ID starting with 0a32a6879db01bf9780e16610fa1359f977fc0147852bffb86708470546ce377 not found: ID does not exist" containerID="0a32a6879db01bf9780e16610fa1359f977fc0147852bffb86708470546ce377" Nov 22 08:56:32 crc kubenswrapper[4789]: I1122 08:56:32.050472 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a32a6879db01bf9780e16610fa1359f977fc0147852bffb86708470546ce377"} err="failed to get container status \"0a32a6879db01bf9780e16610fa1359f977fc0147852bffb86708470546ce377\": rpc error: code = NotFound desc = could not find container \"0a32a6879db01bf9780e16610fa1359f977fc0147852bffb86708470546ce377\": container with ID starting with 0a32a6879db01bf9780e16610fa1359f977fc0147852bffb86708470546ce377 not found: ID does not exist" Nov 22 08:56:32 crc kubenswrapper[4789]: I1122 08:56:32.050489 4789 scope.go:117] "RemoveContainer" containerID="91b0dd2990b97f4ac5d5a357acb0573fd640d2da6303f19ca22866fb509567c8" Nov 22 08:56:32 crc kubenswrapper[4789]: I1122 08:56:32.051016 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91b0dd2990b97f4ac5d5a357acb0573fd640d2da6303f19ca22866fb509567c8"} err="failed to get container status \"91b0dd2990b97f4ac5d5a357acb0573fd640d2da6303f19ca22866fb509567c8\": rpc error: code = NotFound desc = could not find container \"91b0dd2990b97f4ac5d5a357acb0573fd640d2da6303f19ca22866fb509567c8\": container with ID starting with 91b0dd2990b97f4ac5d5a357acb0573fd640d2da6303f19ca22866fb509567c8 not found: ID does not exist" Nov 22 08:56:32 crc kubenswrapper[4789]: I1122 08:56:32.051064 4789 scope.go:117] "RemoveContainer" containerID="0a32a6879db01bf9780e16610fa1359f977fc0147852bffb86708470546ce377" Nov 22 08:56:32 crc kubenswrapper[4789]: I1122 08:56:32.051670 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a32a6879db01bf9780e16610fa1359f977fc0147852bffb86708470546ce377"} err="failed to get container status \"0a32a6879db01bf9780e16610fa1359f977fc0147852bffb86708470546ce377\": rpc error: code = NotFound desc = could not find container \"0a32a6879db01bf9780e16610fa1359f977fc0147852bffb86708470546ce377\": container with ID starting with 0a32a6879db01bf9780e16610fa1359f977fc0147852bffb86708470546ce377 not found: ID does not exist" Nov 22 08:56:32 crc kubenswrapper[4789]: I1122 08:56:32.057259 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pwkjl\" (UniqueName: \"kubernetes.io/projected/50a6fd25-a186-4125-99c1-427d6214b39a-kube-api-access-pwkjl\") on node \"crc\" DevicePath \"\"" Nov 22 08:56:32 crc kubenswrapper[4789]: I1122 08:56:32.245328 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/test-operator-controller-manager-8464cf66df-hc8lc"] Nov 22 08:56:32 crc kubenswrapper[4789]: I1122 08:56:32.253272 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/test-operator-controller-manager-8464cf66df-hc8lc"] Nov 22 08:56:33 crc kubenswrapper[4789]: I1122 08:56:33.976188 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="50a6fd25-a186-4125-99c1-427d6214b39a" path="/var/lib/kubelet/pods/50a6fd25-a186-4125-99c1-427d6214b39a/volumes" Nov 22 08:56:35 crc kubenswrapper[4789]: I1122 08:56:35.372115 4789 patch_prober.go:28] interesting pod/machine-config-daemon-pxvrm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 08:56:35 crc kubenswrapper[4789]: I1122 08:56:35.372511 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 08:57:05 crc kubenswrapper[4789]: I1122 08:57:05.372193 4789 patch_prober.go:28] interesting pod/machine-config-daemon-pxvrm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 08:57:05 crc kubenswrapper[4789]: I1122 08:57:05.372690 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 08:57:05 crc kubenswrapper[4789]: I1122 08:57:05.372739 4789 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" Nov 22 08:57:05 crc kubenswrapper[4789]: I1122 08:57:05.373606 4789 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9a037c1a2417030bb4790bc38e61b3464365daf9c9e9963ba880fe8c29455d3b"} pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 22 08:57:05 crc kubenswrapper[4789]: I1122 08:57:05.373672 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" containerID="cri-o://9a037c1a2417030bb4790bc38e61b3464365daf9c9e9963ba880fe8c29455d3b" gracePeriod=600 Nov 22 08:57:05 crc kubenswrapper[4789]: E1122 08:57:05.513789 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 08:57:06 crc kubenswrapper[4789]: I1122 08:57:06.199236 4789 generic.go:334] "Generic (PLEG): container finished" podID="58aa8071-7ecc-4692-8789-2db04ec70647" containerID="9a037c1a2417030bb4790bc38e61b3464365daf9c9e9963ba880fe8c29455d3b" exitCode=0 Nov 22 08:57:06 crc kubenswrapper[4789]: I1122 08:57:06.199343 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" event={"ID":"58aa8071-7ecc-4692-8789-2db04ec70647","Type":"ContainerDied","Data":"9a037c1a2417030bb4790bc38e61b3464365daf9c9e9963ba880fe8c29455d3b"} Nov 22 08:57:06 crc kubenswrapper[4789]: I1122 08:57:06.199520 4789 scope.go:117] "RemoveContainer" containerID="2e2387a756e8b093bdd9ccbfb31650bf2886cf3ee1392229517961fc3889046d" Nov 22 08:57:06 crc kubenswrapper[4789]: I1122 08:57:06.200203 4789 scope.go:117] "RemoveContainer" containerID="9a037c1a2417030bb4790bc38e61b3464365daf9c9e9963ba880fe8c29455d3b" Nov 22 08:57:06 crc kubenswrapper[4789]: E1122 08:57:06.200458 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 08:57:19 crc kubenswrapper[4789]: I1122 08:57:19.966315 4789 scope.go:117] "RemoveContainer" containerID="9a037c1a2417030bb4790bc38e61b3464365daf9c9e9963ba880fe8c29455d3b" Nov 22 08:57:19 crc kubenswrapper[4789]: E1122 08:57:19.967055 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 08:57:30 crc kubenswrapper[4789]: I1122 08:57:30.965286 4789 scope.go:117] "RemoveContainer" containerID="9a037c1a2417030bb4790bc38e61b3464365daf9c9e9963ba880fe8c29455d3b" Nov 22 08:57:30 crc kubenswrapper[4789]: E1122 08:57:30.966070 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 08:57:45 crc kubenswrapper[4789]: I1122 08:57:45.965671 4789 scope.go:117] "RemoveContainer" containerID="9a037c1a2417030bb4790bc38e61b3464365daf9c9e9963ba880fe8c29455d3b" Nov 22 08:57:45 crc kubenswrapper[4789]: E1122 08:57:45.967088 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 08:58:00 crc kubenswrapper[4789]: I1122 08:58:00.965621 4789 scope.go:117] "RemoveContainer" containerID="9a037c1a2417030bb4790bc38e61b3464365daf9c9e9963ba880fe8c29455d3b" Nov 22 08:58:00 crc kubenswrapper[4789]: E1122 08:58:00.966507 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 08:58:11 crc kubenswrapper[4789]: I1122 08:58:11.966635 4789 scope.go:117] "RemoveContainer" containerID="9a037c1a2417030bb4790bc38e61b3464365daf9c9e9963ba880fe8c29455d3b" Nov 22 08:58:11 crc kubenswrapper[4789]: E1122 08:58:11.968054 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 08:58:24 crc kubenswrapper[4789]: I1122 08:58:24.965392 4789 scope.go:117] "RemoveContainer" containerID="9a037c1a2417030bb4790bc38e61b3464365daf9c9e9963ba880fe8c29455d3b" Nov 22 08:58:24 crc kubenswrapper[4789]: E1122 08:58:24.966410 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 08:58:35 crc kubenswrapper[4789]: I1122 08:58:35.965871 4789 scope.go:117] "RemoveContainer" containerID="9a037c1a2417030bb4790bc38e61b3464365daf9c9e9963ba880fe8c29455d3b" Nov 22 08:58:35 crc kubenswrapper[4789]: E1122 08:58:35.966720 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 08:58:50 crc kubenswrapper[4789]: I1122 08:58:50.965394 4789 scope.go:117] "RemoveContainer" containerID="9a037c1a2417030bb4790bc38e61b3464365daf9c9e9963ba880fe8c29455d3b" Nov 22 08:58:50 crc kubenswrapper[4789]: E1122 08:58:50.966204 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 08:59:02 crc kubenswrapper[4789]: I1122 08:59:02.260101 4789 generic.go:334] "Generic (PLEG): container finished" podID="f3873dc8-add1-47e4-9fee-e1023782e6da" containerID="48bb02e2a06bfb3bca7fe4d54ad856cbc67b115f387ab2f870cffcec08c29f98" exitCode=1 Nov 22 08:59:02 crc kubenswrapper[4789]: I1122 08:59:02.260166 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-7b6cddbf79-2d282" event={"ID":"f3873dc8-add1-47e4-9fee-e1023782e6da","Type":"ContainerDied","Data":"48bb02e2a06bfb3bca7fe4d54ad856cbc67b115f387ab2f870cffcec08c29f98"} Nov 22 08:59:02 crc kubenswrapper[4789]: I1122 08:59:02.261455 4789 scope.go:117] "RemoveContainer" containerID="48bb02e2a06bfb3bca7fe4d54ad856cbc67b115f387ab2f870cffcec08c29f98" Nov 22 08:59:03 crc kubenswrapper[4789]: I1122 08:59:03.272334 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-7b6cddbf79-2d282" event={"ID":"f3873dc8-add1-47e4-9fee-e1023782e6da","Type":"ContainerStarted","Data":"302ce2fa221ecc22bb3b4c50c46861dd21375ac0c6c69441ac533d748612d08d"} Nov 22 08:59:03 crc kubenswrapper[4789]: I1122 08:59:03.273081 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-7b6cddbf79-2d282" Nov 22 08:59:03 crc kubenswrapper[4789]: I1122 08:59:03.965270 4789 scope.go:117] "RemoveContainer" containerID="9a037c1a2417030bb4790bc38e61b3464365daf9c9e9963ba880fe8c29455d3b" Nov 22 08:59:03 crc kubenswrapper[4789]: E1122 08:59:03.965624 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 08:59:11 crc kubenswrapper[4789]: I1122 08:59:11.250251 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-7b6cddbf79-2d282" Nov 22 08:59:17 crc kubenswrapper[4789]: I1122 08:59:17.972609 4789 scope.go:117] "RemoveContainer" containerID="9a037c1a2417030bb4790bc38e61b3464365daf9c9e9963ba880fe8c29455d3b" Nov 22 08:59:17 crc kubenswrapper[4789]: E1122 08:59:17.973514 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 08:59:28 crc kubenswrapper[4789]: I1122 08:59:28.965212 4789 scope.go:117] "RemoveContainer" containerID="9a037c1a2417030bb4790bc38e61b3464365daf9c9e9963ba880fe8c29455d3b" Nov 22 08:59:28 crc kubenswrapper[4789]: E1122 08:59:28.966059 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 08:59:37 crc kubenswrapper[4789]: I1122 08:59:37.475706 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest-s00-full"] Nov 22 08:59:37 crc kubenswrapper[4789]: E1122 08:59:37.476837 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50a6fd25-a186-4125-99c1-427d6214b39a" containerName="manager" Nov 22 08:59:37 crc kubenswrapper[4789]: I1122 08:59:37.476858 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="50a6fd25-a186-4125-99c1-427d6214b39a" containerName="manager" Nov 22 08:59:37 crc kubenswrapper[4789]: E1122 08:59:37.476920 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50a6fd25-a186-4125-99c1-427d6214b39a" containerName="kube-rbac-proxy" Nov 22 08:59:37 crc kubenswrapper[4789]: I1122 08:59:37.476931 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="50a6fd25-a186-4125-99c1-427d6214b39a" containerName="kube-rbac-proxy" Nov 22 08:59:37 crc kubenswrapper[4789]: I1122 08:59:37.477169 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="50a6fd25-a186-4125-99c1-427d6214b39a" containerName="kube-rbac-proxy" Nov 22 08:59:37 crc kubenswrapper[4789]: I1122 08:59:37.477201 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="50a6fd25-a186-4125-99c1-427d6214b39a" containerName="manager" Nov 22 08:59:37 crc kubenswrapper[4789]: I1122 08:59:37.478100 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest-s00-full" Nov 22 08:59:37 crc kubenswrapper[4789]: I1122 08:59:37.480361 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Nov 22 08:59:37 crc kubenswrapper[4789]: I1122 08:59:37.480733 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-49sx9" Nov 22 08:59:37 crc kubenswrapper[4789]: I1122 08:59:37.481102 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Nov 22 08:59:37 crc kubenswrapper[4789]: I1122 08:59:37.481859 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Nov 22 08:59:37 crc kubenswrapper[4789]: I1122 08:59:37.488044 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest-s00-full"] Nov 22 08:59:37 crc kubenswrapper[4789]: I1122 08:59:37.521681 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/18bbf966-a509-4faa-a8de-28045e5c01b4-openstack-config\") pod \"tempest-tests-tempest-s00-full\" (UID: \"18bbf966-a509-4faa-a8de-28045e5c01b4\") " pod="openstack/tempest-tests-tempest-s00-full" Nov 22 08:59:37 crc kubenswrapper[4789]: I1122 08:59:37.522073 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/18bbf966-a509-4faa-a8de-28045e5c01b4-config-data\") pod \"tempest-tests-tempest-s00-full\" (UID: \"18bbf966-a509-4faa-a8de-28045e5c01b4\") " pod="openstack/tempest-tests-tempest-s00-full" Nov 22 08:59:37 crc kubenswrapper[4789]: I1122 08:59:37.522624 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/18bbf966-a509-4faa-a8de-28045e5c01b4-openstack-config-secret\") pod \"tempest-tests-tempest-s00-full\" (UID: \"18bbf966-a509-4faa-a8de-28045e5c01b4\") " pod="openstack/tempest-tests-tempest-s00-full" Nov 22 08:59:37 crc kubenswrapper[4789]: I1122 08:59:37.624605 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/18bbf966-a509-4faa-a8de-28045e5c01b4-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest-s00-full\" (UID: \"18bbf966-a509-4faa-a8de-28045e5c01b4\") " pod="openstack/tempest-tests-tempest-s00-full" Nov 22 08:59:37 crc kubenswrapper[4789]: I1122 08:59:37.625035 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/18bbf966-a509-4faa-a8de-28045e5c01b4-openstack-config\") pod \"tempest-tests-tempest-s00-full\" (UID: \"18bbf966-a509-4faa-a8de-28045e5c01b4\") " pod="openstack/tempest-tests-tempest-s00-full" Nov 22 08:59:37 crc kubenswrapper[4789]: I1122 08:59:37.625074 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"tempest-tests-tempest-s00-full\" (UID: \"18bbf966-a509-4faa-a8de-28045e5c01b4\") " pod="openstack/tempest-tests-tempest-s00-full" Nov 22 08:59:37 crc kubenswrapper[4789]: I1122 08:59:37.625109 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/18bbf966-a509-4faa-a8de-28045e5c01b4-ssh-key\") pod \"tempest-tests-tempest-s00-full\" (UID: \"18bbf966-a509-4faa-a8de-28045e5c01b4\") " pod="openstack/tempest-tests-tempest-s00-full" Nov 22 08:59:37 crc kubenswrapper[4789]: I1122 08:59:37.625127 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-77qg6\" (UniqueName: \"kubernetes.io/projected/18bbf966-a509-4faa-a8de-28045e5c01b4-kube-api-access-77qg6\") pod \"tempest-tests-tempest-s00-full\" (UID: \"18bbf966-a509-4faa-a8de-28045e5c01b4\") " pod="openstack/tempest-tests-tempest-s00-full" Nov 22 08:59:37 crc kubenswrapper[4789]: I1122 08:59:37.625168 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/18bbf966-a509-4faa-a8de-28045e5c01b4-ca-certs\") pod \"tempest-tests-tempest-s00-full\" (UID: \"18bbf966-a509-4faa-a8de-28045e5c01b4\") " pod="openstack/tempest-tests-tempest-s00-full" Nov 22 08:59:37 crc kubenswrapper[4789]: I1122 08:59:37.625189 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/18bbf966-a509-4faa-a8de-28045e5c01b4-ceph\") pod \"tempest-tests-tempest-s00-full\" (UID: \"18bbf966-a509-4faa-a8de-28045e5c01b4\") " pod="openstack/tempest-tests-tempest-s00-full" Nov 22 08:59:37 crc kubenswrapper[4789]: I1122 08:59:37.625207 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/18bbf966-a509-4faa-a8de-28045e5c01b4-config-data\") pod \"tempest-tests-tempest-s00-full\" (UID: \"18bbf966-a509-4faa-a8de-28045e5c01b4\") " pod="openstack/tempest-tests-tempest-s00-full" Nov 22 08:59:37 crc kubenswrapper[4789]: I1122 08:59:37.625480 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/18bbf966-a509-4faa-a8de-28045e5c01b4-openstack-config-secret\") pod \"tempest-tests-tempest-s00-full\" (UID: \"18bbf966-a509-4faa-a8de-28045e5c01b4\") " pod="openstack/tempest-tests-tempest-s00-full" Nov 22 08:59:37 crc kubenswrapper[4789]: I1122 08:59:37.625593 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/18bbf966-a509-4faa-a8de-28045e5c01b4-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest-s00-full\" (UID: \"18bbf966-a509-4faa-a8de-28045e5c01b4\") " pod="openstack/tempest-tests-tempest-s00-full" Nov 22 08:59:37 crc kubenswrapper[4789]: I1122 08:59:37.626262 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/18bbf966-a509-4faa-a8de-28045e5c01b4-openstack-config\") pod \"tempest-tests-tempest-s00-full\" (UID: \"18bbf966-a509-4faa-a8de-28045e5c01b4\") " pod="openstack/tempest-tests-tempest-s00-full" Nov 22 08:59:37 crc kubenswrapper[4789]: I1122 08:59:37.626317 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/18bbf966-a509-4faa-a8de-28045e5c01b4-config-data\") pod \"tempest-tests-tempest-s00-full\" (UID: \"18bbf966-a509-4faa-a8de-28045e5c01b4\") " pod="openstack/tempest-tests-tempest-s00-full" Nov 22 08:59:37 crc kubenswrapper[4789]: I1122 08:59:37.632126 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/18bbf966-a509-4faa-a8de-28045e5c01b4-openstack-config-secret\") pod \"tempest-tests-tempest-s00-full\" (UID: \"18bbf966-a509-4faa-a8de-28045e5c01b4\") " pod="openstack/tempest-tests-tempest-s00-full" Nov 22 08:59:37 crc kubenswrapper[4789]: I1122 08:59:37.728049 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/18bbf966-a509-4faa-a8de-28045e5c01b4-ca-certs\") pod \"tempest-tests-tempest-s00-full\" (UID: \"18bbf966-a509-4faa-a8de-28045e5c01b4\") " pod="openstack/tempest-tests-tempest-s00-full" Nov 22 08:59:37 crc kubenswrapper[4789]: I1122 08:59:37.728096 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/18bbf966-a509-4faa-a8de-28045e5c01b4-ceph\") pod \"tempest-tests-tempest-s00-full\" (UID: \"18bbf966-a509-4faa-a8de-28045e5c01b4\") " pod="openstack/tempest-tests-tempest-s00-full" Nov 22 08:59:37 crc kubenswrapper[4789]: I1122 08:59:37.728137 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/18bbf966-a509-4faa-a8de-28045e5c01b4-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest-s00-full\" (UID: \"18bbf966-a509-4faa-a8de-28045e5c01b4\") " pod="openstack/tempest-tests-tempest-s00-full" Nov 22 08:59:37 crc kubenswrapper[4789]: I1122 08:59:37.728239 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/18bbf966-a509-4faa-a8de-28045e5c01b4-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest-s00-full\" (UID: \"18bbf966-a509-4faa-a8de-28045e5c01b4\") " pod="openstack/tempest-tests-tempest-s00-full" Nov 22 08:59:37 crc kubenswrapper[4789]: I1122 08:59:37.728321 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"tempest-tests-tempest-s00-full\" (UID: \"18bbf966-a509-4faa-a8de-28045e5c01b4\") " pod="openstack/tempest-tests-tempest-s00-full" Nov 22 08:59:37 crc kubenswrapper[4789]: I1122 08:59:37.728361 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/18bbf966-a509-4faa-a8de-28045e5c01b4-ssh-key\") pod \"tempest-tests-tempest-s00-full\" (UID: \"18bbf966-a509-4faa-a8de-28045e5c01b4\") " pod="openstack/tempest-tests-tempest-s00-full" Nov 22 08:59:37 crc kubenswrapper[4789]: I1122 08:59:37.728387 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-77qg6\" (UniqueName: \"kubernetes.io/projected/18bbf966-a509-4faa-a8de-28045e5c01b4-kube-api-access-77qg6\") pod \"tempest-tests-tempest-s00-full\" (UID: \"18bbf966-a509-4faa-a8de-28045e5c01b4\") " pod="openstack/tempest-tests-tempest-s00-full" Nov 22 08:59:37 crc kubenswrapper[4789]: I1122 08:59:37.728923 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/18bbf966-a509-4faa-a8de-28045e5c01b4-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest-s00-full\" (UID: \"18bbf966-a509-4faa-a8de-28045e5c01b4\") " pod="openstack/tempest-tests-tempest-s00-full" Nov 22 08:59:37 crc kubenswrapper[4789]: I1122 08:59:37.729051 4789 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"tempest-tests-tempest-s00-full\" (UID: \"18bbf966-a509-4faa-a8de-28045e5c01b4\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/tempest-tests-tempest-s00-full" Nov 22 08:59:37 crc kubenswrapper[4789]: I1122 08:59:37.729376 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/18bbf966-a509-4faa-a8de-28045e5c01b4-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest-s00-full\" (UID: \"18bbf966-a509-4faa-a8de-28045e5c01b4\") " pod="openstack/tempest-tests-tempest-s00-full" Nov 22 08:59:37 crc kubenswrapper[4789]: I1122 08:59:37.734246 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/18bbf966-a509-4faa-a8de-28045e5c01b4-ceph\") pod \"tempest-tests-tempest-s00-full\" (UID: \"18bbf966-a509-4faa-a8de-28045e5c01b4\") " pod="openstack/tempest-tests-tempest-s00-full" Nov 22 08:59:37 crc kubenswrapper[4789]: I1122 08:59:37.735416 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/18bbf966-a509-4faa-a8de-28045e5c01b4-ca-certs\") pod \"tempest-tests-tempest-s00-full\" (UID: \"18bbf966-a509-4faa-a8de-28045e5c01b4\") " pod="openstack/tempest-tests-tempest-s00-full" Nov 22 08:59:37 crc kubenswrapper[4789]: I1122 08:59:37.736039 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/18bbf966-a509-4faa-a8de-28045e5c01b4-ssh-key\") pod \"tempest-tests-tempest-s00-full\" (UID: \"18bbf966-a509-4faa-a8de-28045e5c01b4\") " pod="openstack/tempest-tests-tempest-s00-full" Nov 22 08:59:37 crc kubenswrapper[4789]: I1122 08:59:37.757537 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-77qg6\" (UniqueName: \"kubernetes.io/projected/18bbf966-a509-4faa-a8de-28045e5c01b4-kube-api-access-77qg6\") pod \"tempest-tests-tempest-s00-full\" (UID: \"18bbf966-a509-4faa-a8de-28045e5c01b4\") " pod="openstack/tempest-tests-tempest-s00-full" Nov 22 08:59:37 crc kubenswrapper[4789]: I1122 08:59:37.760777 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"tempest-tests-tempest-s00-full\" (UID: \"18bbf966-a509-4faa-a8de-28045e5c01b4\") " pod="openstack/tempest-tests-tempest-s00-full" Nov 22 08:59:37 crc kubenswrapper[4789]: I1122 08:59:37.802907 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest-s00-full" Nov 22 08:59:38 crc kubenswrapper[4789]: I1122 08:59:38.310288 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest-s00-full"] Nov 22 08:59:38 crc kubenswrapper[4789]: I1122 08:59:38.575643 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest-s00-full" event={"ID":"18bbf966-a509-4faa-a8de-28045e5c01b4","Type":"ContainerStarted","Data":"efa8a498577cb4705e8534f7b995927f554ec21e5098e6a05c0a810eb9d019e7"} Nov 22 08:59:42 crc kubenswrapper[4789]: I1122 08:59:42.965047 4789 scope.go:117] "RemoveContainer" containerID="9a037c1a2417030bb4790bc38e61b3464365daf9c9e9963ba880fe8c29455d3b" Nov 22 08:59:42 crc kubenswrapper[4789]: E1122 08:59:42.965842 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 08:59:54 crc kubenswrapper[4789]: I1122 08:59:54.965353 4789 scope.go:117] "RemoveContainer" containerID="9a037c1a2417030bb4790bc38e61b3464365daf9c9e9963ba880fe8c29455d3b" Nov 22 08:59:54 crc kubenswrapper[4789]: E1122 08:59:54.966062 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:00:00 crc kubenswrapper[4789]: I1122 09:00:00.145161 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396700-5m7wf"] Nov 22 09:00:00 crc kubenswrapper[4789]: I1122 09:00:00.148348 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396700-5m7wf" Nov 22 09:00:00 crc kubenswrapper[4789]: I1122 09:00:00.150423 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 22 09:00:00 crc kubenswrapper[4789]: I1122 09:00:00.150455 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 22 09:00:00 crc kubenswrapper[4789]: I1122 09:00:00.165575 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396700-5m7wf"] Nov 22 09:00:00 crc kubenswrapper[4789]: I1122 09:00:00.311121 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dkdwm\" (UniqueName: \"kubernetes.io/projected/b530e47e-9170-41a3-9135-19da688d2a1e-kube-api-access-dkdwm\") pod \"collect-profiles-29396700-5m7wf\" (UID: \"b530e47e-9170-41a3-9135-19da688d2a1e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396700-5m7wf" Nov 22 09:00:00 crc kubenswrapper[4789]: I1122 09:00:00.311277 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b530e47e-9170-41a3-9135-19da688d2a1e-config-volume\") pod \"collect-profiles-29396700-5m7wf\" (UID: \"b530e47e-9170-41a3-9135-19da688d2a1e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396700-5m7wf" Nov 22 09:00:00 crc kubenswrapper[4789]: I1122 09:00:00.311347 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b530e47e-9170-41a3-9135-19da688d2a1e-secret-volume\") pod \"collect-profiles-29396700-5m7wf\" (UID: \"b530e47e-9170-41a3-9135-19da688d2a1e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396700-5m7wf" Nov 22 09:00:00 crc kubenswrapper[4789]: I1122 09:00:00.413338 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dkdwm\" (UniqueName: \"kubernetes.io/projected/b530e47e-9170-41a3-9135-19da688d2a1e-kube-api-access-dkdwm\") pod \"collect-profiles-29396700-5m7wf\" (UID: \"b530e47e-9170-41a3-9135-19da688d2a1e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396700-5m7wf" Nov 22 09:00:00 crc kubenswrapper[4789]: I1122 09:00:00.413488 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b530e47e-9170-41a3-9135-19da688d2a1e-config-volume\") pod \"collect-profiles-29396700-5m7wf\" (UID: \"b530e47e-9170-41a3-9135-19da688d2a1e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396700-5m7wf" Nov 22 09:00:00 crc kubenswrapper[4789]: I1122 09:00:00.413544 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b530e47e-9170-41a3-9135-19da688d2a1e-secret-volume\") pod \"collect-profiles-29396700-5m7wf\" (UID: \"b530e47e-9170-41a3-9135-19da688d2a1e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396700-5m7wf" Nov 22 09:00:00 crc kubenswrapper[4789]: I1122 09:00:00.590899 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b530e47e-9170-41a3-9135-19da688d2a1e-config-volume\") pod \"collect-profiles-29396700-5m7wf\" (UID: \"b530e47e-9170-41a3-9135-19da688d2a1e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396700-5m7wf" Nov 22 09:00:00 crc kubenswrapper[4789]: I1122 09:00:00.696535 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dkdwm\" (UniqueName: \"kubernetes.io/projected/b530e47e-9170-41a3-9135-19da688d2a1e-kube-api-access-dkdwm\") pod \"collect-profiles-29396700-5m7wf\" (UID: \"b530e47e-9170-41a3-9135-19da688d2a1e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396700-5m7wf" Nov 22 09:00:00 crc kubenswrapper[4789]: I1122 09:00:00.713377 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b530e47e-9170-41a3-9135-19da688d2a1e-secret-volume\") pod \"collect-profiles-29396700-5m7wf\" (UID: \"b530e47e-9170-41a3-9135-19da688d2a1e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396700-5m7wf" Nov 22 09:00:00 crc kubenswrapper[4789]: I1122 09:00:00.890670 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396700-5m7wf" Nov 22 09:00:02 crc kubenswrapper[4789]: I1122 09:00:02.720487 4789 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.614558498s: [/var/lib/containers/storage/overlay/c1380741959ae1741034d4e6266953306024af0a74c8f3f841b554968c4b1b57/diff /var/log/pods/openshift-console_downloads-7954f5f757-tdpht_83ac1d78-cc4c-48b7-b975-2f88802e2d17/download-server/1.log]; will not log again for this container unless duration exceeds 2s Nov 22 09:00:08 crc kubenswrapper[4789]: I1122 09:00:08.965518 4789 scope.go:117] "RemoveContainer" containerID="9a037c1a2417030bb4790bc38e61b3464365daf9c9e9963ba880fe8c29455d3b" Nov 22 09:00:08 crc kubenswrapper[4789]: E1122 09:00:08.966434 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:00:19 crc kubenswrapper[4789]: I1122 09:00:19.811600 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396700-5m7wf"] Nov 22 09:00:19 crc kubenswrapper[4789]: E1122 09:00:19.957614 4789 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Nov 22 09:00:19 crc kubenswrapper[4789]: E1122 09:00:19.958466 4789 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ceph,ReadOnly:true,MountPath:/etc/ceph,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-77qg6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest-s00-full_openstack(18bbf966-a509-4faa-a8de-28045e5c01b4): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 22 09:00:19 crc kubenswrapper[4789]: E1122 09:00:19.959834 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest-s00-full" podUID="18bbf966-a509-4faa-a8de-28045e5c01b4" Nov 22 09:00:19 crc kubenswrapper[4789]: I1122 09:00:19.988792 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29396700-5m7wf" event={"ID":"b530e47e-9170-41a3-9135-19da688d2a1e","Type":"ContainerStarted","Data":"bc2436ee7e4faae5e90eac855b6206337d133e5fc40856a48e332257b74330d9"} Nov 22 09:00:19 crc kubenswrapper[4789]: E1122 09:00:19.991908 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest-s00-full" podUID="18bbf966-a509-4faa-a8de-28045e5c01b4" Nov 22 09:00:20 crc kubenswrapper[4789]: I1122 09:00:20.998460 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29396700-5m7wf" event={"ID":"b530e47e-9170-41a3-9135-19da688d2a1e","Type":"ContainerStarted","Data":"099aa6447c7eb2f7abee9517700a8ad369c692c48bd9eb28ee3bb2d679b3a02e"} Nov 22 09:00:21 crc kubenswrapper[4789]: I1122 09:00:21.023805 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29396700-5m7wf" podStartSLOduration=21.023772861 podStartE2EDuration="21.023772861s" podCreationTimestamp="2025-11-22 09:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 09:00:21.016722801 +0000 UTC m=+3835.251123084" watchObservedRunningTime="2025-11-22 09:00:21.023772861 +0000 UTC m=+3835.258173134" Nov 22 09:00:22 crc kubenswrapper[4789]: I1122 09:00:22.965893 4789 scope.go:117] "RemoveContainer" containerID="9a037c1a2417030bb4790bc38e61b3464365daf9c9e9963ba880fe8c29455d3b" Nov 22 09:00:22 crc kubenswrapper[4789]: E1122 09:00:22.966646 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:00:23 crc kubenswrapper[4789]: I1122 09:00:23.019442 4789 generic.go:334] "Generic (PLEG): container finished" podID="b530e47e-9170-41a3-9135-19da688d2a1e" containerID="099aa6447c7eb2f7abee9517700a8ad369c692c48bd9eb28ee3bb2d679b3a02e" exitCode=0 Nov 22 09:00:23 crc kubenswrapper[4789]: I1122 09:00:23.019501 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29396700-5m7wf" event={"ID":"b530e47e-9170-41a3-9135-19da688d2a1e","Type":"ContainerDied","Data":"099aa6447c7eb2f7abee9517700a8ad369c692c48bd9eb28ee3bb2d679b3a02e"} Nov 22 09:00:24 crc kubenswrapper[4789]: I1122 09:00:24.462511 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396700-5m7wf" Nov 22 09:00:24 crc kubenswrapper[4789]: I1122 09:00:24.636165 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dkdwm\" (UniqueName: \"kubernetes.io/projected/b530e47e-9170-41a3-9135-19da688d2a1e-kube-api-access-dkdwm\") pod \"b530e47e-9170-41a3-9135-19da688d2a1e\" (UID: \"b530e47e-9170-41a3-9135-19da688d2a1e\") " Nov 22 09:00:24 crc kubenswrapper[4789]: I1122 09:00:24.636404 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b530e47e-9170-41a3-9135-19da688d2a1e-config-volume\") pod \"b530e47e-9170-41a3-9135-19da688d2a1e\" (UID: \"b530e47e-9170-41a3-9135-19da688d2a1e\") " Nov 22 09:00:24 crc kubenswrapper[4789]: I1122 09:00:24.636515 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b530e47e-9170-41a3-9135-19da688d2a1e-secret-volume\") pod \"b530e47e-9170-41a3-9135-19da688d2a1e\" (UID: \"b530e47e-9170-41a3-9135-19da688d2a1e\") " Nov 22 09:00:24 crc kubenswrapper[4789]: I1122 09:00:24.638214 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b530e47e-9170-41a3-9135-19da688d2a1e-config-volume" (OuterVolumeSpecName: "config-volume") pod "b530e47e-9170-41a3-9135-19da688d2a1e" (UID: "b530e47e-9170-41a3-9135-19da688d2a1e"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 09:00:24 crc kubenswrapper[4789]: I1122 09:00:24.644242 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b530e47e-9170-41a3-9135-19da688d2a1e-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "b530e47e-9170-41a3-9135-19da688d2a1e" (UID: "b530e47e-9170-41a3-9135-19da688d2a1e"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 09:00:24 crc kubenswrapper[4789]: I1122 09:00:24.644564 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b530e47e-9170-41a3-9135-19da688d2a1e-kube-api-access-dkdwm" (OuterVolumeSpecName: "kube-api-access-dkdwm") pod "b530e47e-9170-41a3-9135-19da688d2a1e" (UID: "b530e47e-9170-41a3-9135-19da688d2a1e"). InnerVolumeSpecName "kube-api-access-dkdwm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 09:00:24 crc kubenswrapper[4789]: I1122 09:00:24.738888 4789 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b530e47e-9170-41a3-9135-19da688d2a1e-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 22 09:00:24 crc kubenswrapper[4789]: I1122 09:00:24.738922 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dkdwm\" (UniqueName: \"kubernetes.io/projected/b530e47e-9170-41a3-9135-19da688d2a1e-kube-api-access-dkdwm\") on node \"crc\" DevicePath \"\"" Nov 22 09:00:24 crc kubenswrapper[4789]: I1122 09:00:24.738931 4789 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b530e47e-9170-41a3-9135-19da688d2a1e-config-volume\") on node \"crc\" DevicePath \"\"" Nov 22 09:00:25 crc kubenswrapper[4789]: I1122 09:00:25.038398 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29396700-5m7wf" event={"ID":"b530e47e-9170-41a3-9135-19da688d2a1e","Type":"ContainerDied","Data":"bc2436ee7e4faae5e90eac855b6206337d133e5fc40856a48e332257b74330d9"} Nov 22 09:00:25 crc kubenswrapper[4789]: I1122 09:00:25.038727 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bc2436ee7e4faae5e90eac855b6206337d133e5fc40856a48e332257b74330d9" Nov 22 09:00:25 crc kubenswrapper[4789]: I1122 09:00:25.038472 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396700-5m7wf" Nov 22 09:00:25 crc kubenswrapper[4789]: I1122 09:00:25.102191 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396655-pm4q9"] Nov 22 09:00:25 crc kubenswrapper[4789]: I1122 09:00:25.110562 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396655-pm4q9"] Nov 22 09:00:25 crc kubenswrapper[4789]: I1122 09:00:25.975487 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2a71b7c2-042d-4b11-8b37-867c4dddddf2" path="/var/lib/kubelet/pods/2a71b7c2-042d-4b11-8b37-867c4dddddf2/volumes" Nov 22 09:00:28 crc kubenswrapper[4789]: I1122 09:00:28.063199 4789 scope.go:117] "RemoveContainer" containerID="f79b8f35c30c6903423dfa326fc73e1e524773a5acfa2c6d1ab55bb96502e49b" Nov 22 09:00:28 crc kubenswrapper[4789]: I1122 09:00:28.098991 4789 scope.go:117] "RemoveContainer" containerID="18527ce10ae16f3927ffa1f263ae903cc74829ed9695fa52e8ea021020d91dbc" Nov 22 09:00:28 crc kubenswrapper[4789]: I1122 09:00:28.125971 4789 scope.go:117] "RemoveContainer" containerID="254a79e2a772d7a8f8af426f899fa26e8d86e738a8471b562508261c46095767" Nov 22 09:00:31 crc kubenswrapper[4789]: I1122 09:00:31.968975 4789 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 22 09:00:36 crc kubenswrapper[4789]: I1122 09:00:36.965432 4789 scope.go:117] "RemoveContainer" containerID="9a037c1a2417030bb4790bc38e61b3464365daf9c9e9963ba880fe8c29455d3b" Nov 22 09:00:36 crc kubenswrapper[4789]: E1122 09:00:36.966304 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:00:43 crc kubenswrapper[4789]: I1122 09:00:43.319068 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Nov 22 09:00:45 crc kubenswrapper[4789]: I1122 09:00:45.220288 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest-s00-full" event={"ID":"18bbf966-a509-4faa-a8de-28045e5c01b4","Type":"ContainerStarted","Data":"629153cbbe0e8ed9142a669788309296f3b1bcc479a907dbaa59d0624a764760"} Nov 22 09:00:45 crc kubenswrapper[4789]: I1122 09:00:45.243001 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest-s00-full" podStartSLOduration=4.241854513 podStartE2EDuration="1m9.242981005s" podCreationTimestamp="2025-11-22 08:59:36 +0000 UTC" firstStartedPulling="2025-11-22 08:59:38.314784031 +0000 UTC m=+3792.549184304" lastFinishedPulling="2025-11-22 09:00:43.315910523 +0000 UTC m=+3857.550310796" observedRunningTime="2025-11-22 09:00:45.238229927 +0000 UTC m=+3859.472630220" watchObservedRunningTime="2025-11-22 09:00:45.242981005 +0000 UTC m=+3859.477381278" Nov 22 09:00:51 crc kubenswrapper[4789]: I1122 09:00:51.965291 4789 scope.go:117] "RemoveContainer" containerID="9a037c1a2417030bb4790bc38e61b3464365daf9c9e9963ba880fe8c29455d3b" Nov 22 09:00:51 crc kubenswrapper[4789]: E1122 09:00:51.966216 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:01:00 crc kubenswrapper[4789]: I1122 09:01:00.147795 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29396701-s2qrc"] Nov 22 09:01:00 crc kubenswrapper[4789]: E1122 09:01:00.148902 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b530e47e-9170-41a3-9135-19da688d2a1e" containerName="collect-profiles" Nov 22 09:01:00 crc kubenswrapper[4789]: I1122 09:01:00.148922 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="b530e47e-9170-41a3-9135-19da688d2a1e" containerName="collect-profiles" Nov 22 09:01:00 crc kubenswrapper[4789]: I1122 09:01:00.149154 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="b530e47e-9170-41a3-9135-19da688d2a1e" containerName="collect-profiles" Nov 22 09:01:00 crc kubenswrapper[4789]: I1122 09:01:00.149961 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29396701-s2qrc" Nov 22 09:01:00 crc kubenswrapper[4789]: I1122 09:01:00.156972 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29396701-s2qrc"] Nov 22 09:01:00 crc kubenswrapper[4789]: I1122 09:01:00.263208 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/716edb71-e43b-4217-86a7-adab440be7f5-config-data\") pod \"keystone-cron-29396701-s2qrc\" (UID: \"716edb71-e43b-4217-86a7-adab440be7f5\") " pod="openstack/keystone-cron-29396701-s2qrc" Nov 22 09:01:00 crc kubenswrapper[4789]: I1122 09:01:00.263263 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/716edb71-e43b-4217-86a7-adab440be7f5-fernet-keys\") pod \"keystone-cron-29396701-s2qrc\" (UID: \"716edb71-e43b-4217-86a7-adab440be7f5\") " pod="openstack/keystone-cron-29396701-s2qrc" Nov 22 09:01:00 crc kubenswrapper[4789]: I1122 09:01:00.263607 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m9j2z\" (UniqueName: \"kubernetes.io/projected/716edb71-e43b-4217-86a7-adab440be7f5-kube-api-access-m9j2z\") pod \"keystone-cron-29396701-s2qrc\" (UID: \"716edb71-e43b-4217-86a7-adab440be7f5\") " pod="openstack/keystone-cron-29396701-s2qrc" Nov 22 09:01:00 crc kubenswrapper[4789]: I1122 09:01:00.263812 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/716edb71-e43b-4217-86a7-adab440be7f5-combined-ca-bundle\") pod \"keystone-cron-29396701-s2qrc\" (UID: \"716edb71-e43b-4217-86a7-adab440be7f5\") " pod="openstack/keystone-cron-29396701-s2qrc" Nov 22 09:01:00 crc kubenswrapper[4789]: I1122 09:01:00.365393 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m9j2z\" (UniqueName: \"kubernetes.io/projected/716edb71-e43b-4217-86a7-adab440be7f5-kube-api-access-m9j2z\") pod \"keystone-cron-29396701-s2qrc\" (UID: \"716edb71-e43b-4217-86a7-adab440be7f5\") " pod="openstack/keystone-cron-29396701-s2qrc" Nov 22 09:01:00 crc kubenswrapper[4789]: I1122 09:01:00.365468 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/716edb71-e43b-4217-86a7-adab440be7f5-combined-ca-bundle\") pod \"keystone-cron-29396701-s2qrc\" (UID: \"716edb71-e43b-4217-86a7-adab440be7f5\") " pod="openstack/keystone-cron-29396701-s2qrc" Nov 22 09:01:00 crc kubenswrapper[4789]: I1122 09:01:00.365611 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/716edb71-e43b-4217-86a7-adab440be7f5-config-data\") pod \"keystone-cron-29396701-s2qrc\" (UID: \"716edb71-e43b-4217-86a7-adab440be7f5\") " pod="openstack/keystone-cron-29396701-s2qrc" Nov 22 09:01:00 crc kubenswrapper[4789]: I1122 09:01:00.365639 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/716edb71-e43b-4217-86a7-adab440be7f5-fernet-keys\") pod \"keystone-cron-29396701-s2qrc\" (UID: \"716edb71-e43b-4217-86a7-adab440be7f5\") " pod="openstack/keystone-cron-29396701-s2qrc" Nov 22 09:01:00 crc kubenswrapper[4789]: I1122 09:01:00.371880 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/716edb71-e43b-4217-86a7-adab440be7f5-fernet-keys\") pod \"keystone-cron-29396701-s2qrc\" (UID: \"716edb71-e43b-4217-86a7-adab440be7f5\") " pod="openstack/keystone-cron-29396701-s2qrc" Nov 22 09:01:00 crc kubenswrapper[4789]: I1122 09:01:00.372296 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/716edb71-e43b-4217-86a7-adab440be7f5-combined-ca-bundle\") pod \"keystone-cron-29396701-s2qrc\" (UID: \"716edb71-e43b-4217-86a7-adab440be7f5\") " pod="openstack/keystone-cron-29396701-s2qrc" Nov 22 09:01:00 crc kubenswrapper[4789]: I1122 09:01:00.372614 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/716edb71-e43b-4217-86a7-adab440be7f5-config-data\") pod \"keystone-cron-29396701-s2qrc\" (UID: \"716edb71-e43b-4217-86a7-adab440be7f5\") " pod="openstack/keystone-cron-29396701-s2qrc" Nov 22 09:01:00 crc kubenswrapper[4789]: I1122 09:01:00.397953 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m9j2z\" (UniqueName: \"kubernetes.io/projected/716edb71-e43b-4217-86a7-adab440be7f5-kube-api-access-m9j2z\") pod \"keystone-cron-29396701-s2qrc\" (UID: \"716edb71-e43b-4217-86a7-adab440be7f5\") " pod="openstack/keystone-cron-29396701-s2qrc" Nov 22 09:01:00 crc kubenswrapper[4789]: I1122 09:01:00.469737 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29396701-s2qrc" Nov 22 09:01:00 crc kubenswrapper[4789]: I1122 09:01:00.936741 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29396701-s2qrc"] Nov 22 09:01:01 crc kubenswrapper[4789]: I1122 09:01:01.361540 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29396701-s2qrc" event={"ID":"716edb71-e43b-4217-86a7-adab440be7f5","Type":"ContainerStarted","Data":"e31d07be00299e4e7a3475f9de6a3d2e0f7020319e35f3a239f5eab47d558f15"} Nov 22 09:01:01 crc kubenswrapper[4789]: I1122 09:01:01.361589 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29396701-s2qrc" event={"ID":"716edb71-e43b-4217-86a7-adab440be7f5","Type":"ContainerStarted","Data":"9f90be2e54d47b5dcc26fc7c2f20689ccbee74de21950e8acf74c493f76522a9"} Nov 22 09:01:01 crc kubenswrapper[4789]: I1122 09:01:01.386455 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29396701-s2qrc" podStartSLOduration=1.386431189 podStartE2EDuration="1.386431189s" podCreationTimestamp="2025-11-22 09:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 09:01:01.379072452 +0000 UTC m=+3875.613472735" watchObservedRunningTime="2025-11-22 09:01:01.386431189 +0000 UTC m=+3875.620831462" Nov 22 09:01:03 crc kubenswrapper[4789]: I1122 09:01:03.965820 4789 scope.go:117] "RemoveContainer" containerID="9a037c1a2417030bb4790bc38e61b3464365daf9c9e9963ba880fe8c29455d3b" Nov 22 09:01:03 crc kubenswrapper[4789]: E1122 09:01:03.966937 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:01:04 crc kubenswrapper[4789]: I1122 09:01:04.388926 4789 generic.go:334] "Generic (PLEG): container finished" podID="716edb71-e43b-4217-86a7-adab440be7f5" containerID="e31d07be00299e4e7a3475f9de6a3d2e0f7020319e35f3a239f5eab47d558f15" exitCode=0 Nov 22 09:01:04 crc kubenswrapper[4789]: I1122 09:01:04.389089 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29396701-s2qrc" event={"ID":"716edb71-e43b-4217-86a7-adab440be7f5","Type":"ContainerDied","Data":"e31d07be00299e4e7a3475f9de6a3d2e0f7020319e35f3a239f5eab47d558f15"} Nov 22 09:01:05 crc kubenswrapper[4789]: I1122 09:01:05.733456 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29396701-s2qrc" Nov 22 09:01:05 crc kubenswrapper[4789]: I1122 09:01:05.769889 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/716edb71-e43b-4217-86a7-adab440be7f5-combined-ca-bundle\") pod \"716edb71-e43b-4217-86a7-adab440be7f5\" (UID: \"716edb71-e43b-4217-86a7-adab440be7f5\") " Nov 22 09:01:05 crc kubenswrapper[4789]: I1122 09:01:05.770243 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m9j2z\" (UniqueName: \"kubernetes.io/projected/716edb71-e43b-4217-86a7-adab440be7f5-kube-api-access-m9j2z\") pod \"716edb71-e43b-4217-86a7-adab440be7f5\" (UID: \"716edb71-e43b-4217-86a7-adab440be7f5\") " Nov 22 09:01:05 crc kubenswrapper[4789]: I1122 09:01:05.770345 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/716edb71-e43b-4217-86a7-adab440be7f5-fernet-keys\") pod \"716edb71-e43b-4217-86a7-adab440be7f5\" (UID: \"716edb71-e43b-4217-86a7-adab440be7f5\") " Nov 22 09:01:05 crc kubenswrapper[4789]: I1122 09:01:05.770390 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/716edb71-e43b-4217-86a7-adab440be7f5-config-data\") pod \"716edb71-e43b-4217-86a7-adab440be7f5\" (UID: \"716edb71-e43b-4217-86a7-adab440be7f5\") " Nov 22 09:01:05 crc kubenswrapper[4789]: I1122 09:01:05.776812 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/716edb71-e43b-4217-86a7-adab440be7f5-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "716edb71-e43b-4217-86a7-adab440be7f5" (UID: "716edb71-e43b-4217-86a7-adab440be7f5"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 09:01:05 crc kubenswrapper[4789]: I1122 09:01:05.779176 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/716edb71-e43b-4217-86a7-adab440be7f5-kube-api-access-m9j2z" (OuterVolumeSpecName: "kube-api-access-m9j2z") pod "716edb71-e43b-4217-86a7-adab440be7f5" (UID: "716edb71-e43b-4217-86a7-adab440be7f5"). InnerVolumeSpecName "kube-api-access-m9j2z". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 09:01:05 crc kubenswrapper[4789]: I1122 09:01:05.807524 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/716edb71-e43b-4217-86a7-adab440be7f5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "716edb71-e43b-4217-86a7-adab440be7f5" (UID: "716edb71-e43b-4217-86a7-adab440be7f5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 09:01:05 crc kubenswrapper[4789]: I1122 09:01:05.830740 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/716edb71-e43b-4217-86a7-adab440be7f5-config-data" (OuterVolumeSpecName: "config-data") pod "716edb71-e43b-4217-86a7-adab440be7f5" (UID: "716edb71-e43b-4217-86a7-adab440be7f5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 09:01:05 crc kubenswrapper[4789]: I1122 09:01:05.873487 4789 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/716edb71-e43b-4217-86a7-adab440be7f5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 09:01:05 crc kubenswrapper[4789]: I1122 09:01:05.873521 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m9j2z\" (UniqueName: \"kubernetes.io/projected/716edb71-e43b-4217-86a7-adab440be7f5-kube-api-access-m9j2z\") on node \"crc\" DevicePath \"\"" Nov 22 09:01:05 crc kubenswrapper[4789]: I1122 09:01:05.873533 4789 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/716edb71-e43b-4217-86a7-adab440be7f5-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 22 09:01:05 crc kubenswrapper[4789]: I1122 09:01:05.873543 4789 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/716edb71-e43b-4217-86a7-adab440be7f5-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 09:01:06 crc kubenswrapper[4789]: I1122 09:01:06.416802 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29396701-s2qrc" event={"ID":"716edb71-e43b-4217-86a7-adab440be7f5","Type":"ContainerDied","Data":"9f90be2e54d47b5dcc26fc7c2f20689ccbee74de21950e8acf74c493f76522a9"} Nov 22 09:01:06 crc kubenswrapper[4789]: I1122 09:01:06.417110 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9f90be2e54d47b5dcc26fc7c2f20689ccbee74de21950e8acf74c493f76522a9" Nov 22 09:01:06 crc kubenswrapper[4789]: I1122 09:01:06.416867 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29396701-s2qrc" Nov 22 09:01:18 crc kubenswrapper[4789]: I1122 09:01:18.965349 4789 scope.go:117] "RemoveContainer" containerID="9a037c1a2417030bb4790bc38e61b3464365daf9c9e9963ba880fe8c29455d3b" Nov 22 09:01:18 crc kubenswrapper[4789]: E1122 09:01:18.966212 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:01:29 crc kubenswrapper[4789]: I1122 09:01:29.965725 4789 scope.go:117] "RemoveContainer" containerID="9a037c1a2417030bb4790bc38e61b3464365daf9c9e9963ba880fe8c29455d3b" Nov 22 09:01:29 crc kubenswrapper[4789]: E1122 09:01:29.966554 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:01:40 crc kubenswrapper[4789]: I1122 09:01:40.965082 4789 scope.go:117] "RemoveContainer" containerID="9a037c1a2417030bb4790bc38e61b3464365daf9c9e9963ba880fe8c29455d3b" Nov 22 09:01:40 crc kubenswrapper[4789]: E1122 09:01:40.965899 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:01:55 crc kubenswrapper[4789]: I1122 09:01:55.964898 4789 scope.go:117] "RemoveContainer" containerID="9a037c1a2417030bb4790bc38e61b3464365daf9c9e9963ba880fe8c29455d3b" Nov 22 09:01:55 crc kubenswrapper[4789]: E1122 09:01:55.965694 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:02:09 crc kubenswrapper[4789]: I1122 09:02:09.965627 4789 scope.go:117] "RemoveContainer" containerID="9a037c1a2417030bb4790bc38e61b3464365daf9c9e9963ba880fe8c29455d3b" Nov 22 09:02:10 crc kubenswrapper[4789]: I1122 09:02:10.988278 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" event={"ID":"58aa8071-7ecc-4692-8789-2db04ec70647","Type":"ContainerStarted","Data":"d6a9b65a2a78ffb9dd4c2e3e8f411e6945bd81cadebfc9cb238b42ee17ccc01c"} Nov 22 09:02:41 crc kubenswrapper[4789]: I1122 09:02:41.314798 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-62lrk"] Nov 22 09:02:41 crc kubenswrapper[4789]: E1122 09:02:41.316010 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="716edb71-e43b-4217-86a7-adab440be7f5" containerName="keystone-cron" Nov 22 09:02:41 crc kubenswrapper[4789]: I1122 09:02:41.316032 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="716edb71-e43b-4217-86a7-adab440be7f5" containerName="keystone-cron" Nov 22 09:02:41 crc kubenswrapper[4789]: I1122 09:02:41.316328 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="716edb71-e43b-4217-86a7-adab440be7f5" containerName="keystone-cron" Nov 22 09:02:41 crc kubenswrapper[4789]: I1122 09:02:41.320922 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-62lrk" Nov 22 09:02:41 crc kubenswrapper[4789]: I1122 09:02:41.333049 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-62lrk"] Nov 22 09:02:41 crc kubenswrapper[4789]: I1122 09:02:41.346834 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c5d26b89-9c68-4778-a34f-9d405c789e94-catalog-content\") pod \"redhat-marketplace-62lrk\" (UID: \"c5d26b89-9c68-4778-a34f-9d405c789e94\") " pod="openshift-marketplace/redhat-marketplace-62lrk" Nov 22 09:02:41 crc kubenswrapper[4789]: I1122 09:02:41.346939 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c5d26b89-9c68-4778-a34f-9d405c789e94-utilities\") pod \"redhat-marketplace-62lrk\" (UID: \"c5d26b89-9c68-4778-a34f-9d405c789e94\") " pod="openshift-marketplace/redhat-marketplace-62lrk" Nov 22 09:02:41 crc kubenswrapper[4789]: I1122 09:02:41.347160 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sg9ll\" (UniqueName: \"kubernetes.io/projected/c5d26b89-9c68-4778-a34f-9d405c789e94-kube-api-access-sg9ll\") pod \"redhat-marketplace-62lrk\" (UID: \"c5d26b89-9c68-4778-a34f-9d405c789e94\") " pod="openshift-marketplace/redhat-marketplace-62lrk" Nov 22 09:02:41 crc kubenswrapper[4789]: I1122 09:02:41.449162 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sg9ll\" (UniqueName: \"kubernetes.io/projected/c5d26b89-9c68-4778-a34f-9d405c789e94-kube-api-access-sg9ll\") pod \"redhat-marketplace-62lrk\" (UID: \"c5d26b89-9c68-4778-a34f-9d405c789e94\") " pod="openshift-marketplace/redhat-marketplace-62lrk" Nov 22 09:02:41 crc kubenswrapper[4789]: I1122 09:02:41.449310 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c5d26b89-9c68-4778-a34f-9d405c789e94-catalog-content\") pod \"redhat-marketplace-62lrk\" (UID: \"c5d26b89-9c68-4778-a34f-9d405c789e94\") " pod="openshift-marketplace/redhat-marketplace-62lrk" Nov 22 09:02:41 crc kubenswrapper[4789]: I1122 09:02:41.449342 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c5d26b89-9c68-4778-a34f-9d405c789e94-utilities\") pod \"redhat-marketplace-62lrk\" (UID: \"c5d26b89-9c68-4778-a34f-9d405c789e94\") " pod="openshift-marketplace/redhat-marketplace-62lrk" Nov 22 09:02:41 crc kubenswrapper[4789]: I1122 09:02:41.449875 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c5d26b89-9c68-4778-a34f-9d405c789e94-utilities\") pod \"redhat-marketplace-62lrk\" (UID: \"c5d26b89-9c68-4778-a34f-9d405c789e94\") " pod="openshift-marketplace/redhat-marketplace-62lrk" Nov 22 09:02:41 crc kubenswrapper[4789]: I1122 09:02:41.450106 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c5d26b89-9c68-4778-a34f-9d405c789e94-catalog-content\") pod \"redhat-marketplace-62lrk\" (UID: \"c5d26b89-9c68-4778-a34f-9d405c789e94\") " pod="openshift-marketplace/redhat-marketplace-62lrk" Nov 22 09:02:41 crc kubenswrapper[4789]: I1122 09:02:41.470416 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sg9ll\" (UniqueName: \"kubernetes.io/projected/c5d26b89-9c68-4778-a34f-9d405c789e94-kube-api-access-sg9ll\") pod \"redhat-marketplace-62lrk\" (UID: \"c5d26b89-9c68-4778-a34f-9d405c789e94\") " pod="openshift-marketplace/redhat-marketplace-62lrk" Nov 22 09:02:41 crc kubenswrapper[4789]: I1122 09:02:41.652331 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-62lrk" Nov 22 09:02:42 crc kubenswrapper[4789]: I1122 09:02:42.133194 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-62lrk"] Nov 22 09:02:42 crc kubenswrapper[4789]: I1122 09:02:42.247217 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-62lrk" event={"ID":"c5d26b89-9c68-4778-a34f-9d405c789e94","Type":"ContainerStarted","Data":"156a948fde02633036cdf50164f0a06f5c836ff2b3f6f64a7921d244ee2e3b24"} Nov 22 09:02:43 crc kubenswrapper[4789]: I1122 09:02:43.258295 4789 generic.go:334] "Generic (PLEG): container finished" podID="c5d26b89-9c68-4778-a34f-9d405c789e94" containerID="a361b5b0bc8986e7fa7600cc314909feda98acbc763cf8a49ef0fc4617045389" exitCode=0 Nov 22 09:02:43 crc kubenswrapper[4789]: I1122 09:02:43.258343 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-62lrk" event={"ID":"c5d26b89-9c68-4778-a34f-9d405c789e94","Type":"ContainerDied","Data":"a361b5b0bc8986e7fa7600cc314909feda98acbc763cf8a49ef0fc4617045389"} Nov 22 09:02:45 crc kubenswrapper[4789]: I1122 09:02:45.284660 4789 generic.go:334] "Generic (PLEG): container finished" podID="c5d26b89-9c68-4778-a34f-9d405c789e94" containerID="6bdeeaa680cd4a5caf82635cc28c4a70848bcd63ff1a60315af4ab0f4fdb0a1b" exitCode=0 Nov 22 09:02:45 crc kubenswrapper[4789]: I1122 09:02:45.285303 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-62lrk" event={"ID":"c5d26b89-9c68-4778-a34f-9d405c789e94","Type":"ContainerDied","Data":"6bdeeaa680cd4a5caf82635cc28c4a70848bcd63ff1a60315af4ab0f4fdb0a1b"} Nov 22 09:02:46 crc kubenswrapper[4789]: I1122 09:02:46.298583 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-62lrk" event={"ID":"c5d26b89-9c68-4778-a34f-9d405c789e94","Type":"ContainerStarted","Data":"f1466c7d8406b6727db0bcff4f187dd50d2533090f637b59aa045354eb91773c"} Nov 22 09:02:46 crc kubenswrapper[4789]: I1122 09:02:46.327348 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-62lrk" podStartSLOduration=2.9089355169999997 podStartE2EDuration="5.32732195s" podCreationTimestamp="2025-11-22 09:02:41 +0000 UTC" firstStartedPulling="2025-11-22 09:02:43.261515716 +0000 UTC m=+3977.495915989" lastFinishedPulling="2025-11-22 09:02:45.679902149 +0000 UTC m=+3979.914302422" observedRunningTime="2025-11-22 09:02:46.316537346 +0000 UTC m=+3980.550937639" watchObservedRunningTime="2025-11-22 09:02:46.32732195 +0000 UTC m=+3980.561722233" Nov 22 09:02:51 crc kubenswrapper[4789]: I1122 09:02:51.653380 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-62lrk" Nov 22 09:02:51 crc kubenswrapper[4789]: I1122 09:02:51.653840 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-62lrk" Nov 22 09:02:51 crc kubenswrapper[4789]: I1122 09:02:51.709185 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-62lrk" Nov 22 09:02:52 crc kubenswrapper[4789]: I1122 09:02:52.391983 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-62lrk" Nov 22 09:02:52 crc kubenswrapper[4789]: I1122 09:02:52.434981 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-62lrk"] Nov 22 09:02:54 crc kubenswrapper[4789]: I1122 09:02:54.363743 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-62lrk" podUID="c5d26b89-9c68-4778-a34f-9d405c789e94" containerName="registry-server" containerID="cri-o://f1466c7d8406b6727db0bcff4f187dd50d2533090f637b59aa045354eb91773c" gracePeriod=2 Nov 22 09:02:54 crc kubenswrapper[4789]: I1122 09:02:54.963954 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-62lrk" Nov 22 09:02:55 crc kubenswrapper[4789]: I1122 09:02:55.027232 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sg9ll\" (UniqueName: \"kubernetes.io/projected/c5d26b89-9c68-4778-a34f-9d405c789e94-kube-api-access-sg9ll\") pod \"c5d26b89-9c68-4778-a34f-9d405c789e94\" (UID: \"c5d26b89-9c68-4778-a34f-9d405c789e94\") " Nov 22 09:02:55 crc kubenswrapper[4789]: I1122 09:02:55.028014 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c5d26b89-9c68-4778-a34f-9d405c789e94-catalog-content\") pod \"c5d26b89-9c68-4778-a34f-9d405c789e94\" (UID: \"c5d26b89-9c68-4778-a34f-9d405c789e94\") " Nov 22 09:02:55 crc kubenswrapper[4789]: I1122 09:02:55.028093 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c5d26b89-9c68-4778-a34f-9d405c789e94-utilities\") pod \"c5d26b89-9c68-4778-a34f-9d405c789e94\" (UID: \"c5d26b89-9c68-4778-a34f-9d405c789e94\") " Nov 22 09:02:55 crc kubenswrapper[4789]: I1122 09:02:55.029201 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c5d26b89-9c68-4778-a34f-9d405c789e94-utilities" (OuterVolumeSpecName: "utilities") pod "c5d26b89-9c68-4778-a34f-9d405c789e94" (UID: "c5d26b89-9c68-4778-a34f-9d405c789e94"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 09:02:55 crc kubenswrapper[4789]: I1122 09:02:55.031058 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c5d26b89-9c68-4778-a34f-9d405c789e94-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 09:02:55 crc kubenswrapper[4789]: I1122 09:02:55.035615 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c5d26b89-9c68-4778-a34f-9d405c789e94-kube-api-access-sg9ll" (OuterVolumeSpecName: "kube-api-access-sg9ll") pod "c5d26b89-9c68-4778-a34f-9d405c789e94" (UID: "c5d26b89-9c68-4778-a34f-9d405c789e94"). InnerVolumeSpecName "kube-api-access-sg9ll". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 09:02:55 crc kubenswrapper[4789]: I1122 09:02:55.047158 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c5d26b89-9c68-4778-a34f-9d405c789e94-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c5d26b89-9c68-4778-a34f-9d405c789e94" (UID: "c5d26b89-9c68-4778-a34f-9d405c789e94"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 09:02:55 crc kubenswrapper[4789]: I1122 09:02:55.133797 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sg9ll\" (UniqueName: \"kubernetes.io/projected/c5d26b89-9c68-4778-a34f-9d405c789e94-kube-api-access-sg9ll\") on node \"crc\" DevicePath \"\"" Nov 22 09:02:55 crc kubenswrapper[4789]: I1122 09:02:55.133838 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c5d26b89-9c68-4778-a34f-9d405c789e94-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 09:02:55 crc kubenswrapper[4789]: I1122 09:02:55.379316 4789 generic.go:334] "Generic (PLEG): container finished" podID="c5d26b89-9c68-4778-a34f-9d405c789e94" containerID="f1466c7d8406b6727db0bcff4f187dd50d2533090f637b59aa045354eb91773c" exitCode=0 Nov 22 09:02:55 crc kubenswrapper[4789]: I1122 09:02:55.379385 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-62lrk" event={"ID":"c5d26b89-9c68-4778-a34f-9d405c789e94","Type":"ContainerDied","Data":"f1466c7d8406b6727db0bcff4f187dd50d2533090f637b59aa045354eb91773c"} Nov 22 09:02:55 crc kubenswrapper[4789]: I1122 09:02:55.379417 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-62lrk" Nov 22 09:02:55 crc kubenswrapper[4789]: I1122 09:02:55.379448 4789 scope.go:117] "RemoveContainer" containerID="f1466c7d8406b6727db0bcff4f187dd50d2533090f637b59aa045354eb91773c" Nov 22 09:02:55 crc kubenswrapper[4789]: I1122 09:02:55.379430 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-62lrk" event={"ID":"c5d26b89-9c68-4778-a34f-9d405c789e94","Type":"ContainerDied","Data":"156a948fde02633036cdf50164f0a06f5c836ff2b3f6f64a7921d244ee2e3b24"} Nov 22 09:02:55 crc kubenswrapper[4789]: I1122 09:02:55.420359 4789 scope.go:117] "RemoveContainer" containerID="6bdeeaa680cd4a5caf82635cc28c4a70848bcd63ff1a60315af4ab0f4fdb0a1b" Nov 22 09:02:55 crc kubenswrapper[4789]: I1122 09:02:55.443877 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-62lrk"] Nov 22 09:02:55 crc kubenswrapper[4789]: I1122 09:02:55.453624 4789 scope.go:117] "RemoveContainer" containerID="a361b5b0bc8986e7fa7600cc314909feda98acbc763cf8a49ef0fc4617045389" Nov 22 09:02:55 crc kubenswrapper[4789]: I1122 09:02:55.460529 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-62lrk"] Nov 22 09:02:55 crc kubenswrapper[4789]: I1122 09:02:55.509085 4789 scope.go:117] "RemoveContainer" containerID="f1466c7d8406b6727db0bcff4f187dd50d2533090f637b59aa045354eb91773c" Nov 22 09:02:55 crc kubenswrapper[4789]: E1122 09:02:55.509810 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f1466c7d8406b6727db0bcff4f187dd50d2533090f637b59aa045354eb91773c\": container with ID starting with f1466c7d8406b6727db0bcff4f187dd50d2533090f637b59aa045354eb91773c not found: ID does not exist" containerID="f1466c7d8406b6727db0bcff4f187dd50d2533090f637b59aa045354eb91773c" Nov 22 09:02:55 crc kubenswrapper[4789]: I1122 09:02:55.509941 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f1466c7d8406b6727db0bcff4f187dd50d2533090f637b59aa045354eb91773c"} err="failed to get container status \"f1466c7d8406b6727db0bcff4f187dd50d2533090f637b59aa045354eb91773c\": rpc error: code = NotFound desc = could not find container \"f1466c7d8406b6727db0bcff4f187dd50d2533090f637b59aa045354eb91773c\": container with ID starting with f1466c7d8406b6727db0bcff4f187dd50d2533090f637b59aa045354eb91773c not found: ID does not exist" Nov 22 09:02:55 crc kubenswrapper[4789]: I1122 09:02:55.510062 4789 scope.go:117] "RemoveContainer" containerID="6bdeeaa680cd4a5caf82635cc28c4a70848bcd63ff1a60315af4ab0f4fdb0a1b" Nov 22 09:02:55 crc kubenswrapper[4789]: E1122 09:02:55.510882 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6bdeeaa680cd4a5caf82635cc28c4a70848bcd63ff1a60315af4ab0f4fdb0a1b\": container with ID starting with 6bdeeaa680cd4a5caf82635cc28c4a70848bcd63ff1a60315af4ab0f4fdb0a1b not found: ID does not exist" containerID="6bdeeaa680cd4a5caf82635cc28c4a70848bcd63ff1a60315af4ab0f4fdb0a1b" Nov 22 09:02:55 crc kubenswrapper[4789]: I1122 09:02:55.510994 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6bdeeaa680cd4a5caf82635cc28c4a70848bcd63ff1a60315af4ab0f4fdb0a1b"} err="failed to get container status \"6bdeeaa680cd4a5caf82635cc28c4a70848bcd63ff1a60315af4ab0f4fdb0a1b\": rpc error: code = NotFound desc = could not find container \"6bdeeaa680cd4a5caf82635cc28c4a70848bcd63ff1a60315af4ab0f4fdb0a1b\": container with ID starting with 6bdeeaa680cd4a5caf82635cc28c4a70848bcd63ff1a60315af4ab0f4fdb0a1b not found: ID does not exist" Nov 22 09:02:55 crc kubenswrapper[4789]: I1122 09:02:55.511095 4789 scope.go:117] "RemoveContainer" containerID="a361b5b0bc8986e7fa7600cc314909feda98acbc763cf8a49ef0fc4617045389" Nov 22 09:02:55 crc kubenswrapper[4789]: E1122 09:02:55.511833 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a361b5b0bc8986e7fa7600cc314909feda98acbc763cf8a49ef0fc4617045389\": container with ID starting with a361b5b0bc8986e7fa7600cc314909feda98acbc763cf8a49ef0fc4617045389 not found: ID does not exist" containerID="a361b5b0bc8986e7fa7600cc314909feda98acbc763cf8a49ef0fc4617045389" Nov 22 09:02:55 crc kubenswrapper[4789]: I1122 09:02:55.511913 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a361b5b0bc8986e7fa7600cc314909feda98acbc763cf8a49ef0fc4617045389"} err="failed to get container status \"a361b5b0bc8986e7fa7600cc314909feda98acbc763cf8a49ef0fc4617045389\": rpc error: code = NotFound desc = could not find container \"a361b5b0bc8986e7fa7600cc314909feda98acbc763cf8a49ef0fc4617045389\": container with ID starting with a361b5b0bc8986e7fa7600cc314909feda98acbc763cf8a49ef0fc4617045389 not found: ID does not exist" Nov 22 09:02:55 crc kubenswrapper[4789]: I1122 09:02:55.980068 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c5d26b89-9c68-4778-a34f-9d405c789e94" path="/var/lib/kubelet/pods/c5d26b89-9c68-4778-a34f-9d405c789e94/volumes" Nov 22 09:03:03 crc kubenswrapper[4789]: I1122 09:03:03.429828 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-txxxv"] Nov 22 09:03:03 crc kubenswrapper[4789]: E1122 09:03:03.430701 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5d26b89-9c68-4778-a34f-9d405c789e94" containerName="extract-content" Nov 22 09:03:03 crc kubenswrapper[4789]: I1122 09:03:03.430713 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5d26b89-9c68-4778-a34f-9d405c789e94" containerName="extract-content" Nov 22 09:03:03 crc kubenswrapper[4789]: E1122 09:03:03.430731 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5d26b89-9c68-4778-a34f-9d405c789e94" containerName="extract-utilities" Nov 22 09:03:03 crc kubenswrapper[4789]: I1122 09:03:03.430737 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5d26b89-9c68-4778-a34f-9d405c789e94" containerName="extract-utilities" Nov 22 09:03:03 crc kubenswrapper[4789]: E1122 09:03:03.430773 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5d26b89-9c68-4778-a34f-9d405c789e94" containerName="registry-server" Nov 22 09:03:03 crc kubenswrapper[4789]: I1122 09:03:03.430784 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5d26b89-9c68-4778-a34f-9d405c789e94" containerName="registry-server" Nov 22 09:03:03 crc kubenswrapper[4789]: I1122 09:03:03.431019 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5d26b89-9c68-4778-a34f-9d405c789e94" containerName="registry-server" Nov 22 09:03:03 crc kubenswrapper[4789]: I1122 09:03:03.432536 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-txxxv" Nov 22 09:03:03 crc kubenswrapper[4789]: I1122 09:03:03.445457 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-txxxv"] Nov 22 09:03:03 crc kubenswrapper[4789]: I1122 09:03:03.526193 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94f2a362-a63e-4935-b310-7924567d558d-catalog-content\") pod \"community-operators-txxxv\" (UID: \"94f2a362-a63e-4935-b310-7924567d558d\") " pod="openshift-marketplace/community-operators-txxxv" Nov 22 09:03:03 crc kubenswrapper[4789]: I1122 09:03:03.526280 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94f2a362-a63e-4935-b310-7924567d558d-utilities\") pod \"community-operators-txxxv\" (UID: \"94f2a362-a63e-4935-b310-7924567d558d\") " pod="openshift-marketplace/community-operators-txxxv" Nov 22 09:03:03 crc kubenswrapper[4789]: I1122 09:03:03.526308 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vn5xs\" (UniqueName: \"kubernetes.io/projected/94f2a362-a63e-4935-b310-7924567d558d-kube-api-access-vn5xs\") pod \"community-operators-txxxv\" (UID: \"94f2a362-a63e-4935-b310-7924567d558d\") " pod="openshift-marketplace/community-operators-txxxv" Nov 22 09:03:03 crc kubenswrapper[4789]: I1122 09:03:03.628426 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94f2a362-a63e-4935-b310-7924567d558d-catalog-content\") pod \"community-operators-txxxv\" (UID: \"94f2a362-a63e-4935-b310-7924567d558d\") " pod="openshift-marketplace/community-operators-txxxv" Nov 22 09:03:03 crc kubenswrapper[4789]: I1122 09:03:03.628507 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94f2a362-a63e-4935-b310-7924567d558d-utilities\") pod \"community-operators-txxxv\" (UID: \"94f2a362-a63e-4935-b310-7924567d558d\") " pod="openshift-marketplace/community-operators-txxxv" Nov 22 09:03:03 crc kubenswrapper[4789]: I1122 09:03:03.628534 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vn5xs\" (UniqueName: \"kubernetes.io/projected/94f2a362-a63e-4935-b310-7924567d558d-kube-api-access-vn5xs\") pod \"community-operators-txxxv\" (UID: \"94f2a362-a63e-4935-b310-7924567d558d\") " pod="openshift-marketplace/community-operators-txxxv" Nov 22 09:03:03 crc kubenswrapper[4789]: I1122 09:03:03.628961 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94f2a362-a63e-4935-b310-7924567d558d-utilities\") pod \"community-operators-txxxv\" (UID: \"94f2a362-a63e-4935-b310-7924567d558d\") " pod="openshift-marketplace/community-operators-txxxv" Nov 22 09:03:03 crc kubenswrapper[4789]: I1122 09:03:03.629021 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94f2a362-a63e-4935-b310-7924567d558d-catalog-content\") pod \"community-operators-txxxv\" (UID: \"94f2a362-a63e-4935-b310-7924567d558d\") " pod="openshift-marketplace/community-operators-txxxv" Nov 22 09:03:03 crc kubenswrapper[4789]: I1122 09:03:03.648362 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vn5xs\" (UniqueName: \"kubernetes.io/projected/94f2a362-a63e-4935-b310-7924567d558d-kube-api-access-vn5xs\") pod \"community-operators-txxxv\" (UID: \"94f2a362-a63e-4935-b310-7924567d558d\") " pod="openshift-marketplace/community-operators-txxxv" Nov 22 09:03:03 crc kubenswrapper[4789]: I1122 09:03:03.758653 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-txxxv" Nov 22 09:03:04 crc kubenswrapper[4789]: I1122 09:03:04.236888 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-txxxv"] Nov 22 09:03:04 crc kubenswrapper[4789]: I1122 09:03:04.469610 4789 generic.go:334] "Generic (PLEG): container finished" podID="94f2a362-a63e-4935-b310-7924567d558d" containerID="6eb00cee1c00cf9a8a7e48859973b2757608cbd17c8315ab3a20c45501df31d8" exitCode=0 Nov 22 09:03:04 crc kubenswrapper[4789]: I1122 09:03:04.469648 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-txxxv" event={"ID":"94f2a362-a63e-4935-b310-7924567d558d","Type":"ContainerDied","Data":"6eb00cee1c00cf9a8a7e48859973b2757608cbd17c8315ab3a20c45501df31d8"} Nov 22 09:03:04 crc kubenswrapper[4789]: I1122 09:03:04.469674 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-txxxv" event={"ID":"94f2a362-a63e-4935-b310-7924567d558d","Type":"ContainerStarted","Data":"a64c400e12e2914d886f1f5e5518d244ee4fa5eb05d8dcadf058da88967eacec"} Nov 22 09:03:05 crc kubenswrapper[4789]: I1122 09:03:05.480178 4789 generic.go:334] "Generic (PLEG): container finished" podID="94f2a362-a63e-4935-b310-7924567d558d" containerID="d3d7cac49a0a1b232b67d58cde158015472d107ef6f5396e755ef3152110e171" exitCode=0 Nov 22 09:03:05 crc kubenswrapper[4789]: I1122 09:03:05.480258 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-txxxv" event={"ID":"94f2a362-a63e-4935-b310-7924567d558d","Type":"ContainerDied","Data":"d3d7cac49a0a1b232b67d58cde158015472d107ef6f5396e755ef3152110e171"} Nov 22 09:03:06 crc kubenswrapper[4789]: I1122 09:03:06.491579 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-txxxv" event={"ID":"94f2a362-a63e-4935-b310-7924567d558d","Type":"ContainerStarted","Data":"6d079abd1f040e8c82d0eabffa096fd21d8b68a105188a44e7c943b2958924eb"} Nov 22 09:03:13 crc kubenswrapper[4789]: I1122 09:03:13.759274 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-txxxv" Nov 22 09:03:13 crc kubenswrapper[4789]: I1122 09:03:13.759912 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-txxxv" Nov 22 09:03:13 crc kubenswrapper[4789]: I1122 09:03:13.807643 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-txxxv" Nov 22 09:03:13 crc kubenswrapper[4789]: I1122 09:03:13.827216 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-txxxv" podStartSLOduration=9.385156494 podStartE2EDuration="10.827195532s" podCreationTimestamp="2025-11-22 09:03:03 +0000 UTC" firstStartedPulling="2025-11-22 09:03:04.471909579 +0000 UTC m=+3998.706309852" lastFinishedPulling="2025-11-22 09:03:05.913948607 +0000 UTC m=+4000.148348890" observedRunningTime="2025-11-22 09:03:06.510211605 +0000 UTC m=+4000.744611878" watchObservedRunningTime="2025-11-22 09:03:13.827195532 +0000 UTC m=+4008.061595805" Nov 22 09:03:14 crc kubenswrapper[4789]: I1122 09:03:14.608371 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-txxxv" Nov 22 09:03:14 crc kubenswrapper[4789]: I1122 09:03:14.662218 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-txxxv"] Nov 22 09:03:16 crc kubenswrapper[4789]: I1122 09:03:16.574501 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-txxxv" podUID="94f2a362-a63e-4935-b310-7924567d558d" containerName="registry-server" containerID="cri-o://6d079abd1f040e8c82d0eabffa096fd21d8b68a105188a44e7c943b2958924eb" gracePeriod=2 Nov 22 09:03:17 crc kubenswrapper[4789]: I1122 09:03:17.587299 4789 generic.go:334] "Generic (PLEG): container finished" podID="94f2a362-a63e-4935-b310-7924567d558d" containerID="6d079abd1f040e8c82d0eabffa096fd21d8b68a105188a44e7c943b2958924eb" exitCode=0 Nov 22 09:03:17 crc kubenswrapper[4789]: I1122 09:03:17.587383 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-txxxv" event={"ID":"94f2a362-a63e-4935-b310-7924567d558d","Type":"ContainerDied","Data":"6d079abd1f040e8c82d0eabffa096fd21d8b68a105188a44e7c943b2958924eb"} Nov 22 09:03:18 crc kubenswrapper[4789]: I1122 09:03:18.321889 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-txxxv" Nov 22 09:03:18 crc kubenswrapper[4789]: I1122 09:03:18.506912 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vn5xs\" (UniqueName: \"kubernetes.io/projected/94f2a362-a63e-4935-b310-7924567d558d-kube-api-access-vn5xs\") pod \"94f2a362-a63e-4935-b310-7924567d558d\" (UID: \"94f2a362-a63e-4935-b310-7924567d558d\") " Nov 22 09:03:18 crc kubenswrapper[4789]: I1122 09:03:18.507065 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94f2a362-a63e-4935-b310-7924567d558d-utilities\") pod \"94f2a362-a63e-4935-b310-7924567d558d\" (UID: \"94f2a362-a63e-4935-b310-7924567d558d\") " Nov 22 09:03:18 crc kubenswrapper[4789]: I1122 09:03:18.507198 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94f2a362-a63e-4935-b310-7924567d558d-catalog-content\") pod \"94f2a362-a63e-4935-b310-7924567d558d\" (UID: \"94f2a362-a63e-4935-b310-7924567d558d\") " Nov 22 09:03:18 crc kubenswrapper[4789]: I1122 09:03:18.508181 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/94f2a362-a63e-4935-b310-7924567d558d-utilities" (OuterVolumeSpecName: "utilities") pod "94f2a362-a63e-4935-b310-7924567d558d" (UID: "94f2a362-a63e-4935-b310-7924567d558d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 09:03:18 crc kubenswrapper[4789]: I1122 09:03:18.513091 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94f2a362-a63e-4935-b310-7924567d558d-kube-api-access-vn5xs" (OuterVolumeSpecName: "kube-api-access-vn5xs") pod "94f2a362-a63e-4935-b310-7924567d558d" (UID: "94f2a362-a63e-4935-b310-7924567d558d"). InnerVolumeSpecName "kube-api-access-vn5xs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 09:03:18 crc kubenswrapper[4789]: I1122 09:03:18.563739 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/94f2a362-a63e-4935-b310-7924567d558d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "94f2a362-a63e-4935-b310-7924567d558d" (UID: "94f2a362-a63e-4935-b310-7924567d558d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 09:03:18 crc kubenswrapper[4789]: I1122 09:03:18.598980 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-txxxv" event={"ID":"94f2a362-a63e-4935-b310-7924567d558d","Type":"ContainerDied","Data":"a64c400e12e2914d886f1f5e5518d244ee4fa5eb05d8dcadf058da88967eacec"} Nov 22 09:03:18 crc kubenswrapper[4789]: I1122 09:03:18.599046 4789 scope.go:117] "RemoveContainer" containerID="6d079abd1f040e8c82d0eabffa096fd21d8b68a105188a44e7c943b2958924eb" Nov 22 09:03:18 crc kubenswrapper[4789]: I1122 09:03:18.599056 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-txxxv" Nov 22 09:03:18 crc kubenswrapper[4789]: I1122 09:03:18.610254 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vn5xs\" (UniqueName: \"kubernetes.io/projected/94f2a362-a63e-4935-b310-7924567d558d-kube-api-access-vn5xs\") on node \"crc\" DevicePath \"\"" Nov 22 09:03:18 crc kubenswrapper[4789]: I1122 09:03:18.610314 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94f2a362-a63e-4935-b310-7924567d558d-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 09:03:18 crc kubenswrapper[4789]: I1122 09:03:18.610330 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94f2a362-a63e-4935-b310-7924567d558d-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 09:03:18 crc kubenswrapper[4789]: I1122 09:03:18.626191 4789 scope.go:117] "RemoveContainer" containerID="d3d7cac49a0a1b232b67d58cde158015472d107ef6f5396e755ef3152110e171" Nov 22 09:03:18 crc kubenswrapper[4789]: I1122 09:03:18.644183 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-txxxv"] Nov 22 09:03:18 crc kubenswrapper[4789]: I1122 09:03:18.654728 4789 scope.go:117] "RemoveContainer" containerID="6eb00cee1c00cf9a8a7e48859973b2757608cbd17c8315ab3a20c45501df31d8" Nov 22 09:03:18 crc kubenswrapper[4789]: I1122 09:03:18.656611 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-txxxv"] Nov 22 09:03:19 crc kubenswrapper[4789]: I1122 09:03:19.988841 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="94f2a362-a63e-4935-b310-7924567d558d" path="/var/lib/kubelet/pods/94f2a362-a63e-4935-b310-7924567d558d/volumes" Nov 22 09:03:22 crc kubenswrapper[4789]: I1122 09:03:22.044091 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-db-create-2rtqh"] Nov 22 09:03:22 crc kubenswrapper[4789]: I1122 09:03:22.056411 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-b552-account-create-js8sh"] Nov 22 09:03:22 crc kubenswrapper[4789]: I1122 09:03:22.067078 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-db-create-2rtqh"] Nov 22 09:03:22 crc kubenswrapper[4789]: I1122 09:03:22.078034 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-b552-account-create-js8sh"] Nov 22 09:03:23 crc kubenswrapper[4789]: I1122 09:03:23.978106 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5aa41617-1c85-4ccf-b21c-cc5bd6da4902" path="/var/lib/kubelet/pods/5aa41617-1c85-4ccf-b21c-cc5bd6da4902/volumes" Nov 22 09:03:23 crc kubenswrapper[4789]: I1122 09:03:23.979416 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fc0bbde6-8d1e-4bd6-ab42-2c0549fa8792" path="/var/lib/kubelet/pods/fc0bbde6-8d1e-4bd6-ab42-2c0549fa8792/volumes" Nov 22 09:03:28 crc kubenswrapper[4789]: I1122 09:03:28.358023 4789 scope.go:117] "RemoveContainer" containerID="51c3198d42a0b4bc7b03f39d5f91f5e164345794c02ae80a3f8806a0b1a4ef07" Nov 22 09:03:28 crc kubenswrapper[4789]: I1122 09:03:28.382976 4789 scope.go:117] "RemoveContainer" containerID="0d9ed7230b97e612c06435edfe9acc03c8828ac2700f3e2d6269238323be0f0f" Nov 22 09:03:54 crc kubenswrapper[4789]: I1122 09:03:54.052428 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-db-sync-jbx7v"] Nov 22 09:03:54 crc kubenswrapper[4789]: I1122 09:03:54.060733 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-db-sync-jbx7v"] Nov 22 09:03:55 crc kubenswrapper[4789]: I1122 09:03:55.985900 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ff92fe7-35b8-43a0-9d62-cc825720ae76" path="/var/lib/kubelet/pods/6ff92fe7-35b8-43a0-9d62-cc825720ae76/volumes" Nov 22 09:04:28 crc kubenswrapper[4789]: I1122 09:04:28.516025 4789 scope.go:117] "RemoveContainer" containerID="b979cfbd66f08683cb938e63ddf2a4a2cc18432260fbd2a605f434b33ba54a0d" Nov 22 09:04:35 crc kubenswrapper[4789]: I1122 09:04:35.372429 4789 patch_prober.go:28] interesting pod/machine-config-daemon-pxvrm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 09:04:35 crc kubenswrapper[4789]: I1122 09:04:35.373043 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 09:04:52 crc kubenswrapper[4789]: I1122 09:04:52.651349 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-kxg26"] Nov 22 09:04:52 crc kubenswrapper[4789]: E1122 09:04:52.653473 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94f2a362-a63e-4935-b310-7924567d558d" containerName="registry-server" Nov 22 09:04:52 crc kubenswrapper[4789]: I1122 09:04:52.653549 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="94f2a362-a63e-4935-b310-7924567d558d" containerName="registry-server" Nov 22 09:04:52 crc kubenswrapper[4789]: E1122 09:04:52.653637 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94f2a362-a63e-4935-b310-7924567d558d" containerName="extract-utilities" Nov 22 09:04:52 crc kubenswrapper[4789]: I1122 09:04:52.653782 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="94f2a362-a63e-4935-b310-7924567d558d" containerName="extract-utilities" Nov 22 09:04:52 crc kubenswrapper[4789]: E1122 09:04:52.653872 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94f2a362-a63e-4935-b310-7924567d558d" containerName="extract-content" Nov 22 09:04:52 crc kubenswrapper[4789]: I1122 09:04:52.653971 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="94f2a362-a63e-4935-b310-7924567d558d" containerName="extract-content" Nov 22 09:04:52 crc kubenswrapper[4789]: I1122 09:04:52.654225 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="94f2a362-a63e-4935-b310-7924567d558d" containerName="registry-server" Nov 22 09:04:52 crc kubenswrapper[4789]: I1122 09:04:52.655733 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kxg26" Nov 22 09:04:52 crc kubenswrapper[4789]: I1122 09:04:52.662558 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kxg26"] Nov 22 09:04:52 crc kubenswrapper[4789]: I1122 09:04:52.764484 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sdtvk\" (UniqueName: \"kubernetes.io/projected/7bbac2a0-87be-45f4-89c9-bff766aca960-kube-api-access-sdtvk\") pod \"redhat-operators-kxg26\" (UID: \"7bbac2a0-87be-45f4-89c9-bff766aca960\") " pod="openshift-marketplace/redhat-operators-kxg26" Nov 22 09:04:52 crc kubenswrapper[4789]: I1122 09:04:52.764925 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7bbac2a0-87be-45f4-89c9-bff766aca960-catalog-content\") pod \"redhat-operators-kxg26\" (UID: \"7bbac2a0-87be-45f4-89c9-bff766aca960\") " pod="openshift-marketplace/redhat-operators-kxg26" Nov 22 09:04:52 crc kubenswrapper[4789]: I1122 09:04:52.764984 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7bbac2a0-87be-45f4-89c9-bff766aca960-utilities\") pod \"redhat-operators-kxg26\" (UID: \"7bbac2a0-87be-45f4-89c9-bff766aca960\") " pod="openshift-marketplace/redhat-operators-kxg26" Nov 22 09:04:52 crc kubenswrapper[4789]: I1122 09:04:52.867000 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sdtvk\" (UniqueName: \"kubernetes.io/projected/7bbac2a0-87be-45f4-89c9-bff766aca960-kube-api-access-sdtvk\") pod \"redhat-operators-kxg26\" (UID: \"7bbac2a0-87be-45f4-89c9-bff766aca960\") " pod="openshift-marketplace/redhat-operators-kxg26" Nov 22 09:04:52 crc kubenswrapper[4789]: I1122 09:04:52.867064 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7bbac2a0-87be-45f4-89c9-bff766aca960-catalog-content\") pod \"redhat-operators-kxg26\" (UID: \"7bbac2a0-87be-45f4-89c9-bff766aca960\") " pod="openshift-marketplace/redhat-operators-kxg26" Nov 22 09:04:52 crc kubenswrapper[4789]: I1122 09:04:52.867115 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7bbac2a0-87be-45f4-89c9-bff766aca960-utilities\") pod \"redhat-operators-kxg26\" (UID: \"7bbac2a0-87be-45f4-89c9-bff766aca960\") " pod="openshift-marketplace/redhat-operators-kxg26" Nov 22 09:04:52 crc kubenswrapper[4789]: I1122 09:04:52.867894 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7bbac2a0-87be-45f4-89c9-bff766aca960-utilities\") pod \"redhat-operators-kxg26\" (UID: \"7bbac2a0-87be-45f4-89c9-bff766aca960\") " pod="openshift-marketplace/redhat-operators-kxg26" Nov 22 09:04:52 crc kubenswrapper[4789]: I1122 09:04:52.868468 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7bbac2a0-87be-45f4-89c9-bff766aca960-catalog-content\") pod \"redhat-operators-kxg26\" (UID: \"7bbac2a0-87be-45f4-89c9-bff766aca960\") " pod="openshift-marketplace/redhat-operators-kxg26" Nov 22 09:04:52 crc kubenswrapper[4789]: I1122 09:04:52.890199 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sdtvk\" (UniqueName: \"kubernetes.io/projected/7bbac2a0-87be-45f4-89c9-bff766aca960-kube-api-access-sdtvk\") pod \"redhat-operators-kxg26\" (UID: \"7bbac2a0-87be-45f4-89c9-bff766aca960\") " pod="openshift-marketplace/redhat-operators-kxg26" Nov 22 09:04:52 crc kubenswrapper[4789]: I1122 09:04:52.983363 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kxg26" Nov 22 09:04:53 crc kubenswrapper[4789]: I1122 09:04:53.479585 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kxg26"] Nov 22 09:04:54 crc kubenswrapper[4789]: I1122 09:04:54.450880 4789 generic.go:334] "Generic (PLEG): container finished" podID="7bbac2a0-87be-45f4-89c9-bff766aca960" containerID="2a6d1c990a85b53ef326deb0030066f483827b5064dc926e1c6862e441eff3a2" exitCode=0 Nov 22 09:04:54 crc kubenswrapper[4789]: I1122 09:04:54.450957 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kxg26" event={"ID":"7bbac2a0-87be-45f4-89c9-bff766aca960","Type":"ContainerDied","Data":"2a6d1c990a85b53ef326deb0030066f483827b5064dc926e1c6862e441eff3a2"} Nov 22 09:04:54 crc kubenswrapper[4789]: I1122 09:04:54.451609 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kxg26" event={"ID":"7bbac2a0-87be-45f4-89c9-bff766aca960","Type":"ContainerStarted","Data":"5e1e9214aaf5b84abc37c3279cd5ed142f11a44d43634d14e7b0082068743454"} Nov 22 09:04:56 crc kubenswrapper[4789]: I1122 09:04:56.471481 4789 generic.go:334] "Generic (PLEG): container finished" podID="7bbac2a0-87be-45f4-89c9-bff766aca960" containerID="e99fe546b127a4bb5fb72488338944bc381d62f45dcfb1471bb541f06d16ad19" exitCode=0 Nov 22 09:04:56 crc kubenswrapper[4789]: I1122 09:04:56.471567 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kxg26" event={"ID":"7bbac2a0-87be-45f4-89c9-bff766aca960","Type":"ContainerDied","Data":"e99fe546b127a4bb5fb72488338944bc381d62f45dcfb1471bb541f06d16ad19"} Nov 22 09:04:58 crc kubenswrapper[4789]: I1122 09:04:58.492689 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kxg26" event={"ID":"7bbac2a0-87be-45f4-89c9-bff766aca960","Type":"ContainerStarted","Data":"fb5e1798a800d675c1e2e65eda1e0008713fb814050f2f31c90ef43600f18841"} Nov 22 09:04:58 crc kubenswrapper[4789]: I1122 09:04:58.517557 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-kxg26" podStartSLOduration=3.963068793 podStartE2EDuration="6.517538667s" podCreationTimestamp="2025-11-22 09:04:52 +0000 UTC" firstStartedPulling="2025-11-22 09:04:54.45354191 +0000 UTC m=+4108.687942183" lastFinishedPulling="2025-11-22 09:04:57.008011784 +0000 UTC m=+4111.242412057" observedRunningTime="2025-11-22 09:04:58.508905019 +0000 UTC m=+4112.743305292" watchObservedRunningTime="2025-11-22 09:04:58.517538667 +0000 UTC m=+4112.751938940" Nov 22 09:05:02 crc kubenswrapper[4789]: I1122 09:05:02.983706 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-kxg26" Nov 22 09:05:02 crc kubenswrapper[4789]: I1122 09:05:02.984390 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-kxg26" Nov 22 09:05:04 crc kubenswrapper[4789]: I1122 09:05:04.040868 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-kxg26" podUID="7bbac2a0-87be-45f4-89c9-bff766aca960" containerName="registry-server" probeResult="failure" output=< Nov 22 09:05:04 crc kubenswrapper[4789]: timeout: failed to connect service ":50051" within 1s Nov 22 09:05:04 crc kubenswrapper[4789]: > Nov 22 09:05:05 crc kubenswrapper[4789]: I1122 09:05:05.372103 4789 patch_prober.go:28] interesting pod/machine-config-daemon-pxvrm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 09:05:05 crc kubenswrapper[4789]: I1122 09:05:05.372453 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 09:05:13 crc kubenswrapper[4789]: I1122 09:05:13.042885 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-kxg26" Nov 22 09:05:13 crc kubenswrapper[4789]: I1122 09:05:13.093367 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-kxg26" Nov 22 09:05:13 crc kubenswrapper[4789]: I1122 09:05:13.284139 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kxg26"] Nov 22 09:05:14 crc kubenswrapper[4789]: I1122 09:05:14.635129 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-kxg26" podUID="7bbac2a0-87be-45f4-89c9-bff766aca960" containerName="registry-server" containerID="cri-o://fb5e1798a800d675c1e2e65eda1e0008713fb814050f2f31c90ef43600f18841" gracePeriod=2 Nov 22 09:05:15 crc kubenswrapper[4789]: I1122 09:05:15.644836 4789 generic.go:334] "Generic (PLEG): container finished" podID="7bbac2a0-87be-45f4-89c9-bff766aca960" containerID="fb5e1798a800d675c1e2e65eda1e0008713fb814050f2f31c90ef43600f18841" exitCode=0 Nov 22 09:05:15 crc kubenswrapper[4789]: I1122 09:05:15.645040 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kxg26" event={"ID":"7bbac2a0-87be-45f4-89c9-bff766aca960","Type":"ContainerDied","Data":"fb5e1798a800d675c1e2e65eda1e0008713fb814050f2f31c90ef43600f18841"} Nov 22 09:05:16 crc kubenswrapper[4789]: I1122 09:05:16.194831 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kxg26" Nov 22 09:05:16 crc kubenswrapper[4789]: I1122 09:05:16.233659 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7bbac2a0-87be-45f4-89c9-bff766aca960-catalog-content\") pod \"7bbac2a0-87be-45f4-89c9-bff766aca960\" (UID: \"7bbac2a0-87be-45f4-89c9-bff766aca960\") " Nov 22 09:05:16 crc kubenswrapper[4789]: I1122 09:05:16.233771 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7bbac2a0-87be-45f4-89c9-bff766aca960-utilities\") pod \"7bbac2a0-87be-45f4-89c9-bff766aca960\" (UID: \"7bbac2a0-87be-45f4-89c9-bff766aca960\") " Nov 22 09:05:16 crc kubenswrapper[4789]: I1122 09:05:16.233872 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sdtvk\" (UniqueName: \"kubernetes.io/projected/7bbac2a0-87be-45f4-89c9-bff766aca960-kube-api-access-sdtvk\") pod \"7bbac2a0-87be-45f4-89c9-bff766aca960\" (UID: \"7bbac2a0-87be-45f4-89c9-bff766aca960\") " Nov 22 09:05:16 crc kubenswrapper[4789]: I1122 09:05:16.234595 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7bbac2a0-87be-45f4-89c9-bff766aca960-utilities" (OuterVolumeSpecName: "utilities") pod "7bbac2a0-87be-45f4-89c9-bff766aca960" (UID: "7bbac2a0-87be-45f4-89c9-bff766aca960"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 09:05:16 crc kubenswrapper[4789]: I1122 09:05:16.240423 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bbac2a0-87be-45f4-89c9-bff766aca960-kube-api-access-sdtvk" (OuterVolumeSpecName: "kube-api-access-sdtvk") pod "7bbac2a0-87be-45f4-89c9-bff766aca960" (UID: "7bbac2a0-87be-45f4-89c9-bff766aca960"). InnerVolumeSpecName "kube-api-access-sdtvk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 09:05:16 crc kubenswrapper[4789]: I1122 09:05:16.325425 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7bbac2a0-87be-45f4-89c9-bff766aca960-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7bbac2a0-87be-45f4-89c9-bff766aca960" (UID: "7bbac2a0-87be-45f4-89c9-bff766aca960"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 09:05:16 crc kubenswrapper[4789]: I1122 09:05:16.336010 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sdtvk\" (UniqueName: \"kubernetes.io/projected/7bbac2a0-87be-45f4-89c9-bff766aca960-kube-api-access-sdtvk\") on node \"crc\" DevicePath \"\"" Nov 22 09:05:16 crc kubenswrapper[4789]: I1122 09:05:16.336046 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7bbac2a0-87be-45f4-89c9-bff766aca960-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 09:05:16 crc kubenswrapper[4789]: I1122 09:05:16.336077 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7bbac2a0-87be-45f4-89c9-bff766aca960-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 09:05:16 crc kubenswrapper[4789]: I1122 09:05:16.658131 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kxg26" event={"ID":"7bbac2a0-87be-45f4-89c9-bff766aca960","Type":"ContainerDied","Data":"5e1e9214aaf5b84abc37c3279cd5ed142f11a44d43634d14e7b0082068743454"} Nov 22 09:05:16 crc kubenswrapper[4789]: I1122 09:05:16.658178 4789 scope.go:117] "RemoveContainer" containerID="fb5e1798a800d675c1e2e65eda1e0008713fb814050f2f31c90ef43600f18841" Nov 22 09:05:16 crc kubenswrapper[4789]: I1122 09:05:16.658196 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kxg26" Nov 22 09:05:16 crc kubenswrapper[4789]: I1122 09:05:16.682348 4789 scope.go:117] "RemoveContainer" containerID="e99fe546b127a4bb5fb72488338944bc381d62f45dcfb1471bb541f06d16ad19" Nov 22 09:05:16 crc kubenswrapper[4789]: I1122 09:05:16.694465 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kxg26"] Nov 22 09:05:16 crc kubenswrapper[4789]: I1122 09:05:16.704301 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-kxg26"] Nov 22 09:05:16 crc kubenswrapper[4789]: I1122 09:05:16.724916 4789 scope.go:117] "RemoveContainer" containerID="2a6d1c990a85b53ef326deb0030066f483827b5064dc926e1c6862e441eff3a2" Nov 22 09:05:17 crc kubenswrapper[4789]: I1122 09:05:17.974842 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bbac2a0-87be-45f4-89c9-bff766aca960" path="/var/lib/kubelet/pods/7bbac2a0-87be-45f4-89c9-bff766aca960/volumes" Nov 22 09:05:35 crc kubenswrapper[4789]: I1122 09:05:35.371713 4789 patch_prober.go:28] interesting pod/machine-config-daemon-pxvrm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 09:05:35 crc kubenswrapper[4789]: I1122 09:05:35.372415 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 09:05:35 crc kubenswrapper[4789]: I1122 09:05:35.372473 4789 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" Nov 22 09:05:35 crc kubenswrapper[4789]: I1122 09:05:35.373512 4789 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d6a9b65a2a78ffb9dd4c2e3e8f411e6945bd81cadebfc9cb238b42ee17ccc01c"} pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 22 09:05:35 crc kubenswrapper[4789]: I1122 09:05:35.373585 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" containerID="cri-o://d6a9b65a2a78ffb9dd4c2e3e8f411e6945bd81cadebfc9cb238b42ee17ccc01c" gracePeriod=600 Nov 22 09:05:35 crc kubenswrapper[4789]: I1122 09:05:35.824028 4789 generic.go:334] "Generic (PLEG): container finished" podID="58aa8071-7ecc-4692-8789-2db04ec70647" containerID="d6a9b65a2a78ffb9dd4c2e3e8f411e6945bd81cadebfc9cb238b42ee17ccc01c" exitCode=0 Nov 22 09:05:35 crc kubenswrapper[4789]: I1122 09:05:35.824120 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" event={"ID":"58aa8071-7ecc-4692-8789-2db04ec70647","Type":"ContainerDied","Data":"d6a9b65a2a78ffb9dd4c2e3e8f411e6945bd81cadebfc9cb238b42ee17ccc01c"} Nov 22 09:05:35 crc kubenswrapper[4789]: I1122 09:05:35.824337 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" event={"ID":"58aa8071-7ecc-4692-8789-2db04ec70647","Type":"ContainerStarted","Data":"e9395b7f22fc562bccbab647e93ed06aa21f27b91666ef88922354e2f66bcf54"} Nov 22 09:05:35 crc kubenswrapper[4789]: I1122 09:05:35.824360 4789 scope.go:117] "RemoveContainer" containerID="9a037c1a2417030bb4790bc38e61b3464365daf9c9e9963ba880fe8c29455d3b" Nov 22 09:07:10 crc kubenswrapper[4789]: I1122 09:07:10.527932 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-xghwc"] Nov 22 09:07:10 crc kubenswrapper[4789]: E1122 09:07:10.528964 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7bbac2a0-87be-45f4-89c9-bff766aca960" containerName="extract-utilities" Nov 22 09:07:10 crc kubenswrapper[4789]: I1122 09:07:10.528984 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="7bbac2a0-87be-45f4-89c9-bff766aca960" containerName="extract-utilities" Nov 22 09:07:10 crc kubenswrapper[4789]: E1122 09:07:10.529009 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7bbac2a0-87be-45f4-89c9-bff766aca960" containerName="extract-content" Nov 22 09:07:10 crc kubenswrapper[4789]: I1122 09:07:10.529017 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="7bbac2a0-87be-45f4-89c9-bff766aca960" containerName="extract-content" Nov 22 09:07:10 crc kubenswrapper[4789]: E1122 09:07:10.529036 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7bbac2a0-87be-45f4-89c9-bff766aca960" containerName="registry-server" Nov 22 09:07:10 crc kubenswrapper[4789]: I1122 09:07:10.529044 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="7bbac2a0-87be-45f4-89c9-bff766aca960" containerName="registry-server" Nov 22 09:07:10 crc kubenswrapper[4789]: I1122 09:07:10.529301 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="7bbac2a0-87be-45f4-89c9-bff766aca960" containerName="registry-server" Nov 22 09:07:10 crc kubenswrapper[4789]: I1122 09:07:10.531013 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xghwc" Nov 22 09:07:10 crc kubenswrapper[4789]: I1122 09:07:10.553417 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xghwc"] Nov 22 09:07:10 crc kubenswrapper[4789]: I1122 09:07:10.582338 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fpz8z\" (UniqueName: \"kubernetes.io/projected/c7800900-e61a-4aa1-80f0-f68c945b3f66-kube-api-access-fpz8z\") pod \"certified-operators-xghwc\" (UID: \"c7800900-e61a-4aa1-80f0-f68c945b3f66\") " pod="openshift-marketplace/certified-operators-xghwc" Nov 22 09:07:10 crc kubenswrapper[4789]: I1122 09:07:10.582428 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c7800900-e61a-4aa1-80f0-f68c945b3f66-catalog-content\") pod \"certified-operators-xghwc\" (UID: \"c7800900-e61a-4aa1-80f0-f68c945b3f66\") " pod="openshift-marketplace/certified-operators-xghwc" Nov 22 09:07:10 crc kubenswrapper[4789]: I1122 09:07:10.582463 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c7800900-e61a-4aa1-80f0-f68c945b3f66-utilities\") pod \"certified-operators-xghwc\" (UID: \"c7800900-e61a-4aa1-80f0-f68c945b3f66\") " pod="openshift-marketplace/certified-operators-xghwc" Nov 22 09:07:10 crc kubenswrapper[4789]: I1122 09:07:10.684400 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fpz8z\" (UniqueName: \"kubernetes.io/projected/c7800900-e61a-4aa1-80f0-f68c945b3f66-kube-api-access-fpz8z\") pod \"certified-operators-xghwc\" (UID: \"c7800900-e61a-4aa1-80f0-f68c945b3f66\") " pod="openshift-marketplace/certified-operators-xghwc" Nov 22 09:07:10 crc kubenswrapper[4789]: I1122 09:07:10.684525 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c7800900-e61a-4aa1-80f0-f68c945b3f66-catalog-content\") pod \"certified-operators-xghwc\" (UID: \"c7800900-e61a-4aa1-80f0-f68c945b3f66\") " pod="openshift-marketplace/certified-operators-xghwc" Nov 22 09:07:10 crc kubenswrapper[4789]: I1122 09:07:10.684570 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c7800900-e61a-4aa1-80f0-f68c945b3f66-utilities\") pod \"certified-operators-xghwc\" (UID: \"c7800900-e61a-4aa1-80f0-f68c945b3f66\") " pod="openshift-marketplace/certified-operators-xghwc" Nov 22 09:07:10 crc kubenswrapper[4789]: I1122 09:07:10.685206 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c7800900-e61a-4aa1-80f0-f68c945b3f66-utilities\") pod \"certified-operators-xghwc\" (UID: \"c7800900-e61a-4aa1-80f0-f68c945b3f66\") " pod="openshift-marketplace/certified-operators-xghwc" Nov 22 09:07:10 crc kubenswrapper[4789]: I1122 09:07:10.685218 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c7800900-e61a-4aa1-80f0-f68c945b3f66-catalog-content\") pod \"certified-operators-xghwc\" (UID: \"c7800900-e61a-4aa1-80f0-f68c945b3f66\") " pod="openshift-marketplace/certified-operators-xghwc" Nov 22 09:07:10 crc kubenswrapper[4789]: I1122 09:07:10.725047 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fpz8z\" (UniqueName: \"kubernetes.io/projected/c7800900-e61a-4aa1-80f0-f68c945b3f66-kube-api-access-fpz8z\") pod \"certified-operators-xghwc\" (UID: \"c7800900-e61a-4aa1-80f0-f68c945b3f66\") " pod="openshift-marketplace/certified-operators-xghwc" Nov 22 09:07:10 crc kubenswrapper[4789]: I1122 09:07:10.853239 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xghwc" Nov 22 09:07:11 crc kubenswrapper[4789]: I1122 09:07:11.401932 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xghwc"] Nov 22 09:07:12 crc kubenswrapper[4789]: I1122 09:07:12.633629 4789 generic.go:334] "Generic (PLEG): container finished" podID="c7800900-e61a-4aa1-80f0-f68c945b3f66" containerID="ccf7157443856d8d5bf16e7f58734b5621d8d54ec62d5477ebd0245caf4ce24b" exitCode=0 Nov 22 09:07:12 crc kubenswrapper[4789]: I1122 09:07:12.633715 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xghwc" event={"ID":"c7800900-e61a-4aa1-80f0-f68c945b3f66","Type":"ContainerDied","Data":"ccf7157443856d8d5bf16e7f58734b5621d8d54ec62d5477ebd0245caf4ce24b"} Nov 22 09:07:12 crc kubenswrapper[4789]: I1122 09:07:12.634179 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xghwc" event={"ID":"c7800900-e61a-4aa1-80f0-f68c945b3f66","Type":"ContainerStarted","Data":"2bc7b26278d25fd06f23d7413913e37afa105296a2b017fcfe8887bdacbf7111"} Nov 22 09:07:12 crc kubenswrapper[4789]: I1122 09:07:12.636085 4789 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 22 09:07:14 crc kubenswrapper[4789]: I1122 09:07:14.653159 4789 generic.go:334] "Generic (PLEG): container finished" podID="c7800900-e61a-4aa1-80f0-f68c945b3f66" containerID="80498529b1a740ecfe7c6aed50ad3f10fb513987edeb9572e1732f26e8379d08" exitCode=0 Nov 22 09:07:14 crc kubenswrapper[4789]: I1122 09:07:14.653250 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xghwc" event={"ID":"c7800900-e61a-4aa1-80f0-f68c945b3f66","Type":"ContainerDied","Data":"80498529b1a740ecfe7c6aed50ad3f10fb513987edeb9572e1732f26e8379d08"} Nov 22 09:07:17 crc kubenswrapper[4789]: I1122 09:07:17.680518 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xghwc" event={"ID":"c7800900-e61a-4aa1-80f0-f68c945b3f66","Type":"ContainerStarted","Data":"656602eb58e49ff553e34fecf5f4632b6b7ad40bb866c3f02280c1214dd7b634"} Nov 22 09:07:17 crc kubenswrapper[4789]: I1122 09:07:17.706209 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-xghwc" podStartSLOduration=4.081858938 podStartE2EDuration="7.706190383s" podCreationTimestamp="2025-11-22 09:07:10 +0000 UTC" firstStartedPulling="2025-11-22 09:07:12.635483495 +0000 UTC m=+4246.869883768" lastFinishedPulling="2025-11-22 09:07:16.25981494 +0000 UTC m=+4250.494215213" observedRunningTime="2025-11-22 09:07:17.696150086 +0000 UTC m=+4251.930550359" watchObservedRunningTime="2025-11-22 09:07:17.706190383 +0000 UTC m=+4251.940590656" Nov 22 09:07:20 crc kubenswrapper[4789]: I1122 09:07:20.854215 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-xghwc" Nov 22 09:07:20 crc kubenswrapper[4789]: I1122 09:07:20.854854 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-xghwc" Nov 22 09:07:20 crc kubenswrapper[4789]: I1122 09:07:20.905625 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-xghwc" Nov 22 09:07:21 crc kubenswrapper[4789]: I1122 09:07:21.767676 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-xghwc" Nov 22 09:07:21 crc kubenswrapper[4789]: I1122 09:07:21.815121 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xghwc"] Nov 22 09:07:23 crc kubenswrapper[4789]: I1122 09:07:23.727296 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-xghwc" podUID="c7800900-e61a-4aa1-80f0-f68c945b3f66" containerName="registry-server" containerID="cri-o://656602eb58e49ff553e34fecf5f4632b6b7ad40bb866c3f02280c1214dd7b634" gracePeriod=2 Nov 22 09:07:24 crc kubenswrapper[4789]: I1122 09:07:24.382215 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xghwc" Nov 22 09:07:24 crc kubenswrapper[4789]: I1122 09:07:24.456442 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fpz8z\" (UniqueName: \"kubernetes.io/projected/c7800900-e61a-4aa1-80f0-f68c945b3f66-kube-api-access-fpz8z\") pod \"c7800900-e61a-4aa1-80f0-f68c945b3f66\" (UID: \"c7800900-e61a-4aa1-80f0-f68c945b3f66\") " Nov 22 09:07:24 crc kubenswrapper[4789]: I1122 09:07:24.456585 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c7800900-e61a-4aa1-80f0-f68c945b3f66-utilities\") pod \"c7800900-e61a-4aa1-80f0-f68c945b3f66\" (UID: \"c7800900-e61a-4aa1-80f0-f68c945b3f66\") " Nov 22 09:07:24 crc kubenswrapper[4789]: I1122 09:07:24.456645 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c7800900-e61a-4aa1-80f0-f68c945b3f66-catalog-content\") pod \"c7800900-e61a-4aa1-80f0-f68c945b3f66\" (UID: \"c7800900-e61a-4aa1-80f0-f68c945b3f66\") " Nov 22 09:07:24 crc kubenswrapper[4789]: I1122 09:07:24.457727 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c7800900-e61a-4aa1-80f0-f68c945b3f66-utilities" (OuterVolumeSpecName: "utilities") pod "c7800900-e61a-4aa1-80f0-f68c945b3f66" (UID: "c7800900-e61a-4aa1-80f0-f68c945b3f66"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 09:07:24 crc kubenswrapper[4789]: I1122 09:07:24.464227 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c7800900-e61a-4aa1-80f0-f68c945b3f66-kube-api-access-fpz8z" (OuterVolumeSpecName: "kube-api-access-fpz8z") pod "c7800900-e61a-4aa1-80f0-f68c945b3f66" (UID: "c7800900-e61a-4aa1-80f0-f68c945b3f66"). InnerVolumeSpecName "kube-api-access-fpz8z". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 09:07:24 crc kubenswrapper[4789]: I1122 09:07:24.503434 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c7800900-e61a-4aa1-80f0-f68c945b3f66-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c7800900-e61a-4aa1-80f0-f68c945b3f66" (UID: "c7800900-e61a-4aa1-80f0-f68c945b3f66"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 09:07:24 crc kubenswrapper[4789]: I1122 09:07:24.558882 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fpz8z\" (UniqueName: \"kubernetes.io/projected/c7800900-e61a-4aa1-80f0-f68c945b3f66-kube-api-access-fpz8z\") on node \"crc\" DevicePath \"\"" Nov 22 09:07:24 crc kubenswrapper[4789]: I1122 09:07:24.558913 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c7800900-e61a-4aa1-80f0-f68c945b3f66-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 09:07:24 crc kubenswrapper[4789]: I1122 09:07:24.558924 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c7800900-e61a-4aa1-80f0-f68c945b3f66-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 09:07:24 crc kubenswrapper[4789]: I1122 09:07:24.738371 4789 generic.go:334] "Generic (PLEG): container finished" podID="c7800900-e61a-4aa1-80f0-f68c945b3f66" containerID="656602eb58e49ff553e34fecf5f4632b6b7ad40bb866c3f02280c1214dd7b634" exitCode=0 Nov 22 09:07:24 crc kubenswrapper[4789]: I1122 09:07:24.738410 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xghwc" Nov 22 09:07:24 crc kubenswrapper[4789]: I1122 09:07:24.738426 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xghwc" event={"ID":"c7800900-e61a-4aa1-80f0-f68c945b3f66","Type":"ContainerDied","Data":"656602eb58e49ff553e34fecf5f4632b6b7ad40bb866c3f02280c1214dd7b634"} Nov 22 09:07:24 crc kubenswrapper[4789]: I1122 09:07:24.738798 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xghwc" event={"ID":"c7800900-e61a-4aa1-80f0-f68c945b3f66","Type":"ContainerDied","Data":"2bc7b26278d25fd06f23d7413913e37afa105296a2b017fcfe8887bdacbf7111"} Nov 22 09:07:24 crc kubenswrapper[4789]: I1122 09:07:24.738815 4789 scope.go:117] "RemoveContainer" containerID="656602eb58e49ff553e34fecf5f4632b6b7ad40bb866c3f02280c1214dd7b634" Nov 22 09:07:24 crc kubenswrapper[4789]: I1122 09:07:24.761025 4789 scope.go:117] "RemoveContainer" containerID="80498529b1a740ecfe7c6aed50ad3f10fb513987edeb9572e1732f26e8379d08" Nov 22 09:07:24 crc kubenswrapper[4789]: I1122 09:07:24.769078 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xghwc"] Nov 22 09:07:24 crc kubenswrapper[4789]: I1122 09:07:24.778487 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-xghwc"] Nov 22 09:07:25 crc kubenswrapper[4789]: I1122 09:07:25.010382 4789 scope.go:117] "RemoveContainer" containerID="ccf7157443856d8d5bf16e7f58734b5621d8d54ec62d5477ebd0245caf4ce24b" Nov 22 09:07:25 crc kubenswrapper[4789]: I1122 09:07:25.252715 4789 scope.go:117] "RemoveContainer" containerID="656602eb58e49ff553e34fecf5f4632b6b7ad40bb866c3f02280c1214dd7b634" Nov 22 09:07:25 crc kubenswrapper[4789]: E1122 09:07:25.253226 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"656602eb58e49ff553e34fecf5f4632b6b7ad40bb866c3f02280c1214dd7b634\": container with ID starting with 656602eb58e49ff553e34fecf5f4632b6b7ad40bb866c3f02280c1214dd7b634 not found: ID does not exist" containerID="656602eb58e49ff553e34fecf5f4632b6b7ad40bb866c3f02280c1214dd7b634" Nov 22 09:07:25 crc kubenswrapper[4789]: I1122 09:07:25.253264 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"656602eb58e49ff553e34fecf5f4632b6b7ad40bb866c3f02280c1214dd7b634"} err="failed to get container status \"656602eb58e49ff553e34fecf5f4632b6b7ad40bb866c3f02280c1214dd7b634\": rpc error: code = NotFound desc = could not find container \"656602eb58e49ff553e34fecf5f4632b6b7ad40bb866c3f02280c1214dd7b634\": container with ID starting with 656602eb58e49ff553e34fecf5f4632b6b7ad40bb866c3f02280c1214dd7b634 not found: ID does not exist" Nov 22 09:07:25 crc kubenswrapper[4789]: I1122 09:07:25.253286 4789 scope.go:117] "RemoveContainer" containerID="80498529b1a740ecfe7c6aed50ad3f10fb513987edeb9572e1732f26e8379d08" Nov 22 09:07:25 crc kubenswrapper[4789]: E1122 09:07:25.253626 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"80498529b1a740ecfe7c6aed50ad3f10fb513987edeb9572e1732f26e8379d08\": container with ID starting with 80498529b1a740ecfe7c6aed50ad3f10fb513987edeb9572e1732f26e8379d08 not found: ID does not exist" containerID="80498529b1a740ecfe7c6aed50ad3f10fb513987edeb9572e1732f26e8379d08" Nov 22 09:07:25 crc kubenswrapper[4789]: I1122 09:07:25.253645 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80498529b1a740ecfe7c6aed50ad3f10fb513987edeb9572e1732f26e8379d08"} err="failed to get container status \"80498529b1a740ecfe7c6aed50ad3f10fb513987edeb9572e1732f26e8379d08\": rpc error: code = NotFound desc = could not find container \"80498529b1a740ecfe7c6aed50ad3f10fb513987edeb9572e1732f26e8379d08\": container with ID starting with 80498529b1a740ecfe7c6aed50ad3f10fb513987edeb9572e1732f26e8379d08 not found: ID does not exist" Nov 22 09:07:25 crc kubenswrapper[4789]: I1122 09:07:25.253660 4789 scope.go:117] "RemoveContainer" containerID="ccf7157443856d8d5bf16e7f58734b5621d8d54ec62d5477ebd0245caf4ce24b" Nov 22 09:07:25 crc kubenswrapper[4789]: E1122 09:07:25.253983 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ccf7157443856d8d5bf16e7f58734b5621d8d54ec62d5477ebd0245caf4ce24b\": container with ID starting with ccf7157443856d8d5bf16e7f58734b5621d8d54ec62d5477ebd0245caf4ce24b not found: ID does not exist" containerID="ccf7157443856d8d5bf16e7f58734b5621d8d54ec62d5477ebd0245caf4ce24b" Nov 22 09:07:25 crc kubenswrapper[4789]: I1122 09:07:25.254014 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ccf7157443856d8d5bf16e7f58734b5621d8d54ec62d5477ebd0245caf4ce24b"} err="failed to get container status \"ccf7157443856d8d5bf16e7f58734b5621d8d54ec62d5477ebd0245caf4ce24b\": rpc error: code = NotFound desc = could not find container \"ccf7157443856d8d5bf16e7f58734b5621d8d54ec62d5477ebd0245caf4ce24b\": container with ID starting with ccf7157443856d8d5bf16e7f58734b5621d8d54ec62d5477ebd0245caf4ce24b not found: ID does not exist" Nov 22 09:07:25 crc kubenswrapper[4789]: I1122 09:07:25.997195 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c7800900-e61a-4aa1-80f0-f68c945b3f66" path="/var/lib/kubelet/pods/c7800900-e61a-4aa1-80f0-f68c945b3f66/volumes" Nov 22 09:07:35 crc kubenswrapper[4789]: I1122 09:07:35.371932 4789 patch_prober.go:28] interesting pod/machine-config-daemon-pxvrm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 09:07:35 crc kubenswrapper[4789]: I1122 09:07:35.372631 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 09:08:05 crc kubenswrapper[4789]: I1122 09:08:05.372277 4789 patch_prober.go:28] interesting pod/machine-config-daemon-pxvrm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 09:08:05 crc kubenswrapper[4789]: I1122 09:08:05.372875 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 09:08:35 crc kubenswrapper[4789]: I1122 09:08:35.372767 4789 patch_prober.go:28] interesting pod/machine-config-daemon-pxvrm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 09:08:35 crc kubenswrapper[4789]: I1122 09:08:35.373360 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 09:08:35 crc kubenswrapper[4789]: I1122 09:08:35.373418 4789 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" Nov 22 09:08:35 crc kubenswrapper[4789]: I1122 09:08:35.374287 4789 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e9395b7f22fc562bccbab647e93ed06aa21f27b91666ef88922354e2f66bcf54"} pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 22 09:08:35 crc kubenswrapper[4789]: I1122 09:08:35.374357 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" containerID="cri-o://e9395b7f22fc562bccbab647e93ed06aa21f27b91666ef88922354e2f66bcf54" gracePeriod=600 Nov 22 09:08:35 crc kubenswrapper[4789]: E1122 09:08:35.603717 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:08:36 crc kubenswrapper[4789]: I1122 09:08:36.432715 4789 generic.go:334] "Generic (PLEG): container finished" podID="58aa8071-7ecc-4692-8789-2db04ec70647" containerID="e9395b7f22fc562bccbab647e93ed06aa21f27b91666ef88922354e2f66bcf54" exitCode=0 Nov 22 09:08:36 crc kubenswrapper[4789]: I1122 09:08:36.432791 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" event={"ID":"58aa8071-7ecc-4692-8789-2db04ec70647","Type":"ContainerDied","Data":"e9395b7f22fc562bccbab647e93ed06aa21f27b91666ef88922354e2f66bcf54"} Nov 22 09:08:36 crc kubenswrapper[4789]: I1122 09:08:36.433146 4789 scope.go:117] "RemoveContainer" containerID="d6a9b65a2a78ffb9dd4c2e3e8f411e6945bd81cadebfc9cb238b42ee17ccc01c" Nov 22 09:08:36 crc kubenswrapper[4789]: I1122 09:08:36.433851 4789 scope.go:117] "RemoveContainer" containerID="e9395b7f22fc562bccbab647e93ed06aa21f27b91666ef88922354e2f66bcf54" Nov 22 09:08:36 crc kubenswrapper[4789]: E1122 09:08:36.434248 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:08:51 crc kubenswrapper[4789]: I1122 09:08:51.965732 4789 scope.go:117] "RemoveContainer" containerID="e9395b7f22fc562bccbab647e93ed06aa21f27b91666ef88922354e2f66bcf54" Nov 22 09:08:51 crc kubenswrapper[4789]: E1122 09:08:51.967621 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:09:06 crc kubenswrapper[4789]: I1122 09:09:06.966062 4789 scope.go:117] "RemoveContainer" containerID="e9395b7f22fc562bccbab647e93ed06aa21f27b91666ef88922354e2f66bcf54" Nov 22 09:09:06 crc kubenswrapper[4789]: E1122 09:09:06.966852 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:09:19 crc kubenswrapper[4789]: I1122 09:09:19.966381 4789 scope.go:117] "RemoveContainer" containerID="e9395b7f22fc562bccbab647e93ed06aa21f27b91666ef88922354e2f66bcf54" Nov 22 09:09:19 crc kubenswrapper[4789]: E1122 09:09:19.967568 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:09:31 crc kubenswrapper[4789]: I1122 09:09:31.965422 4789 scope.go:117] "RemoveContainer" containerID="e9395b7f22fc562bccbab647e93ed06aa21f27b91666ef88922354e2f66bcf54" Nov 22 09:09:31 crc kubenswrapper[4789]: E1122 09:09:31.966260 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:09:43 crc kubenswrapper[4789]: I1122 09:09:43.965937 4789 scope.go:117] "RemoveContainer" containerID="e9395b7f22fc562bccbab647e93ed06aa21f27b91666ef88922354e2f66bcf54" Nov 22 09:09:43 crc kubenswrapper[4789]: E1122 09:09:43.966761 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:09:57 crc kubenswrapper[4789]: I1122 09:09:57.972598 4789 scope.go:117] "RemoveContainer" containerID="e9395b7f22fc562bccbab647e93ed06aa21f27b91666ef88922354e2f66bcf54" Nov 22 09:09:57 crc kubenswrapper[4789]: E1122 09:09:57.973539 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:10:11 crc kubenswrapper[4789]: I1122 09:10:11.971688 4789 scope.go:117] "RemoveContainer" containerID="e9395b7f22fc562bccbab647e93ed06aa21f27b91666ef88922354e2f66bcf54" Nov 22 09:10:11 crc kubenswrapper[4789]: E1122 09:10:11.972491 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:10:22 crc kubenswrapper[4789]: I1122 09:10:22.965187 4789 scope.go:117] "RemoveContainer" containerID="e9395b7f22fc562bccbab647e93ed06aa21f27b91666ef88922354e2f66bcf54" Nov 22 09:10:22 crc kubenswrapper[4789]: E1122 09:10:22.965906 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:10:34 crc kubenswrapper[4789]: I1122 09:10:34.967859 4789 scope.go:117] "RemoveContainer" containerID="e9395b7f22fc562bccbab647e93ed06aa21f27b91666ef88922354e2f66bcf54" Nov 22 09:10:34 crc kubenswrapper[4789]: E1122 09:10:34.969149 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:10:48 crc kubenswrapper[4789]: I1122 09:10:48.965209 4789 scope.go:117] "RemoveContainer" containerID="e9395b7f22fc562bccbab647e93ed06aa21f27b91666ef88922354e2f66bcf54" Nov 22 09:10:48 crc kubenswrapper[4789]: E1122 09:10:48.966133 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:11:02 crc kubenswrapper[4789]: I1122 09:11:02.965575 4789 scope.go:117] "RemoveContainer" containerID="e9395b7f22fc562bccbab647e93ed06aa21f27b91666ef88922354e2f66bcf54" Nov 22 09:11:02 crc kubenswrapper[4789]: E1122 09:11:02.966327 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:11:17 crc kubenswrapper[4789]: I1122 09:11:17.972817 4789 scope.go:117] "RemoveContainer" containerID="e9395b7f22fc562bccbab647e93ed06aa21f27b91666ef88922354e2f66bcf54" Nov 22 09:11:17 crc kubenswrapper[4789]: E1122 09:11:17.973551 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:11:31 crc kubenswrapper[4789]: I1122 09:11:31.965172 4789 scope.go:117] "RemoveContainer" containerID="e9395b7f22fc562bccbab647e93ed06aa21f27b91666ef88922354e2f66bcf54" Nov 22 09:11:31 crc kubenswrapper[4789]: E1122 09:11:31.965967 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:11:44 crc kubenswrapper[4789]: I1122 09:11:44.965596 4789 scope.go:117] "RemoveContainer" containerID="e9395b7f22fc562bccbab647e93ed06aa21f27b91666ef88922354e2f66bcf54" Nov 22 09:11:44 crc kubenswrapper[4789]: E1122 09:11:44.966442 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:11:57 crc kubenswrapper[4789]: I1122 09:11:57.971885 4789 scope.go:117] "RemoveContainer" containerID="e9395b7f22fc562bccbab647e93ed06aa21f27b91666ef88922354e2f66bcf54" Nov 22 09:11:57 crc kubenswrapper[4789]: E1122 09:11:57.972870 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:12:11 crc kubenswrapper[4789]: I1122 09:12:11.966401 4789 scope.go:117] "RemoveContainer" containerID="e9395b7f22fc562bccbab647e93ed06aa21f27b91666ef88922354e2f66bcf54" Nov 22 09:12:11 crc kubenswrapper[4789]: E1122 09:12:11.968187 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:12:25 crc kubenswrapper[4789]: I1122 09:12:25.965099 4789 scope.go:117] "RemoveContainer" containerID="e9395b7f22fc562bccbab647e93ed06aa21f27b91666ef88922354e2f66bcf54" Nov 22 09:12:25 crc kubenswrapper[4789]: E1122 09:12:25.965819 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:12:37 crc kubenswrapper[4789]: I1122 09:12:37.974510 4789 scope.go:117] "RemoveContainer" containerID="e9395b7f22fc562bccbab647e93ed06aa21f27b91666ef88922354e2f66bcf54" Nov 22 09:12:37 crc kubenswrapper[4789]: E1122 09:12:37.975374 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:12:49 crc kubenswrapper[4789]: I1122 09:12:49.545391 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-xrv8z"] Nov 22 09:12:49 crc kubenswrapper[4789]: E1122 09:12:49.546650 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7800900-e61a-4aa1-80f0-f68c945b3f66" containerName="extract-content" Nov 22 09:12:49 crc kubenswrapper[4789]: I1122 09:12:49.546668 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7800900-e61a-4aa1-80f0-f68c945b3f66" containerName="extract-content" Nov 22 09:12:49 crc kubenswrapper[4789]: E1122 09:12:49.546685 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7800900-e61a-4aa1-80f0-f68c945b3f66" containerName="extract-utilities" Nov 22 09:12:49 crc kubenswrapper[4789]: I1122 09:12:49.546693 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7800900-e61a-4aa1-80f0-f68c945b3f66" containerName="extract-utilities" Nov 22 09:12:49 crc kubenswrapper[4789]: E1122 09:12:49.546707 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7800900-e61a-4aa1-80f0-f68c945b3f66" containerName="registry-server" Nov 22 09:12:49 crc kubenswrapper[4789]: I1122 09:12:49.546716 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7800900-e61a-4aa1-80f0-f68c945b3f66" containerName="registry-server" Nov 22 09:12:49 crc kubenswrapper[4789]: I1122 09:12:49.546979 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="c7800900-e61a-4aa1-80f0-f68c945b3f66" containerName="registry-server" Nov 22 09:12:49 crc kubenswrapper[4789]: I1122 09:12:49.548896 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xrv8z" Nov 22 09:12:49 crc kubenswrapper[4789]: I1122 09:12:49.577921 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xrv8z"] Nov 22 09:12:49 crc kubenswrapper[4789]: I1122 09:12:49.723079 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b36c461e-1e80-4f8c-8ec2-74abfb91a4f0-utilities\") pod \"redhat-marketplace-xrv8z\" (UID: \"b36c461e-1e80-4f8c-8ec2-74abfb91a4f0\") " pod="openshift-marketplace/redhat-marketplace-xrv8z" Nov 22 09:12:49 crc kubenswrapper[4789]: I1122 09:12:49.723485 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b36c461e-1e80-4f8c-8ec2-74abfb91a4f0-catalog-content\") pod \"redhat-marketplace-xrv8z\" (UID: \"b36c461e-1e80-4f8c-8ec2-74abfb91a4f0\") " pod="openshift-marketplace/redhat-marketplace-xrv8z" Nov 22 09:12:49 crc kubenswrapper[4789]: I1122 09:12:49.723534 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sr5g4\" (UniqueName: \"kubernetes.io/projected/b36c461e-1e80-4f8c-8ec2-74abfb91a4f0-kube-api-access-sr5g4\") pod \"redhat-marketplace-xrv8z\" (UID: \"b36c461e-1e80-4f8c-8ec2-74abfb91a4f0\") " pod="openshift-marketplace/redhat-marketplace-xrv8z" Nov 22 09:12:49 crc kubenswrapper[4789]: I1122 09:12:49.825557 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b36c461e-1e80-4f8c-8ec2-74abfb91a4f0-utilities\") pod \"redhat-marketplace-xrv8z\" (UID: \"b36c461e-1e80-4f8c-8ec2-74abfb91a4f0\") " pod="openshift-marketplace/redhat-marketplace-xrv8z" Nov 22 09:12:49 crc kubenswrapper[4789]: I1122 09:12:49.825637 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b36c461e-1e80-4f8c-8ec2-74abfb91a4f0-catalog-content\") pod \"redhat-marketplace-xrv8z\" (UID: \"b36c461e-1e80-4f8c-8ec2-74abfb91a4f0\") " pod="openshift-marketplace/redhat-marketplace-xrv8z" Nov 22 09:12:49 crc kubenswrapper[4789]: I1122 09:12:49.825675 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sr5g4\" (UniqueName: \"kubernetes.io/projected/b36c461e-1e80-4f8c-8ec2-74abfb91a4f0-kube-api-access-sr5g4\") pod \"redhat-marketplace-xrv8z\" (UID: \"b36c461e-1e80-4f8c-8ec2-74abfb91a4f0\") " pod="openshift-marketplace/redhat-marketplace-xrv8z" Nov 22 09:12:49 crc kubenswrapper[4789]: I1122 09:12:49.826067 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b36c461e-1e80-4f8c-8ec2-74abfb91a4f0-utilities\") pod \"redhat-marketplace-xrv8z\" (UID: \"b36c461e-1e80-4f8c-8ec2-74abfb91a4f0\") " pod="openshift-marketplace/redhat-marketplace-xrv8z" Nov 22 09:12:49 crc kubenswrapper[4789]: I1122 09:12:49.826110 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b36c461e-1e80-4f8c-8ec2-74abfb91a4f0-catalog-content\") pod \"redhat-marketplace-xrv8z\" (UID: \"b36c461e-1e80-4f8c-8ec2-74abfb91a4f0\") " pod="openshift-marketplace/redhat-marketplace-xrv8z" Nov 22 09:12:49 crc kubenswrapper[4789]: I1122 09:12:49.846931 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sr5g4\" (UniqueName: \"kubernetes.io/projected/b36c461e-1e80-4f8c-8ec2-74abfb91a4f0-kube-api-access-sr5g4\") pod \"redhat-marketplace-xrv8z\" (UID: \"b36c461e-1e80-4f8c-8ec2-74abfb91a4f0\") " pod="openshift-marketplace/redhat-marketplace-xrv8z" Nov 22 09:12:49 crc kubenswrapper[4789]: I1122 09:12:49.869740 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xrv8z" Nov 22 09:12:50 crc kubenswrapper[4789]: I1122 09:12:50.416025 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xrv8z"] Nov 22 09:12:50 crc kubenswrapper[4789]: I1122 09:12:50.591139 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xrv8z" event={"ID":"b36c461e-1e80-4f8c-8ec2-74abfb91a4f0","Type":"ContainerStarted","Data":"4118d7ecfdbeaa926677b020cebeed696a94a19b7db484273aeb8986d2da0db4"} Nov 22 09:12:50 crc kubenswrapper[4789]: I1122 09:12:50.965912 4789 scope.go:117] "RemoveContainer" containerID="e9395b7f22fc562bccbab647e93ed06aa21f27b91666ef88922354e2f66bcf54" Nov 22 09:12:50 crc kubenswrapper[4789]: E1122 09:12:50.966176 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:12:51 crc kubenswrapper[4789]: I1122 09:12:51.607175 4789 generic.go:334] "Generic (PLEG): container finished" podID="b36c461e-1e80-4f8c-8ec2-74abfb91a4f0" containerID="6310d94eab8d0f04fd4868283f8104f57138c7b38a1819ee800638c5825ae4f2" exitCode=0 Nov 22 09:12:51 crc kubenswrapper[4789]: I1122 09:12:51.607488 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xrv8z" event={"ID":"b36c461e-1e80-4f8c-8ec2-74abfb91a4f0","Type":"ContainerDied","Data":"6310d94eab8d0f04fd4868283f8104f57138c7b38a1819ee800638c5825ae4f2"} Nov 22 09:12:51 crc kubenswrapper[4789]: I1122 09:12:51.610309 4789 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 22 09:12:52 crc kubenswrapper[4789]: I1122 09:12:52.624265 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xrv8z" event={"ID":"b36c461e-1e80-4f8c-8ec2-74abfb91a4f0","Type":"ContainerStarted","Data":"6e4edf59ff68bb769e1e16428812845855300da350f17631296437911cee62e7"} Nov 22 09:12:53 crc kubenswrapper[4789]: I1122 09:12:53.637234 4789 generic.go:334] "Generic (PLEG): container finished" podID="b36c461e-1e80-4f8c-8ec2-74abfb91a4f0" containerID="6e4edf59ff68bb769e1e16428812845855300da350f17631296437911cee62e7" exitCode=0 Nov 22 09:12:53 crc kubenswrapper[4789]: I1122 09:12:53.637313 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xrv8z" event={"ID":"b36c461e-1e80-4f8c-8ec2-74abfb91a4f0","Type":"ContainerDied","Data":"6e4edf59ff68bb769e1e16428812845855300da350f17631296437911cee62e7"} Nov 22 09:12:54 crc kubenswrapper[4789]: I1122 09:12:54.650624 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xrv8z" event={"ID":"b36c461e-1e80-4f8c-8ec2-74abfb91a4f0","Type":"ContainerStarted","Data":"3b3f2eb4aa73d5ae399e38b45af896f083d276f5026692e5b0c159368e1e7d7a"} Nov 22 09:12:54 crc kubenswrapper[4789]: I1122 09:12:54.668279 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-xrv8z" podStartSLOduration=3.211102677 podStartE2EDuration="5.668257328s" podCreationTimestamp="2025-11-22 09:12:49 +0000 UTC" firstStartedPulling="2025-11-22 09:12:51.610007527 +0000 UTC m=+4585.844407800" lastFinishedPulling="2025-11-22 09:12:54.067162178 +0000 UTC m=+4588.301562451" observedRunningTime="2025-11-22 09:12:54.667089317 +0000 UTC m=+4588.901489610" watchObservedRunningTime="2025-11-22 09:12:54.668257328 +0000 UTC m=+4588.902657591" Nov 22 09:12:59 crc kubenswrapper[4789]: I1122 09:12:59.869961 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-xrv8z" Nov 22 09:12:59 crc kubenswrapper[4789]: I1122 09:12:59.870607 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-xrv8z" Nov 22 09:13:00 crc kubenswrapper[4789]: I1122 09:13:00.411766 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-xrv8z" Nov 22 09:13:00 crc kubenswrapper[4789]: I1122 09:13:00.747832 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-xrv8z" Nov 22 09:13:00 crc kubenswrapper[4789]: I1122 09:13:00.797653 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xrv8z"] Nov 22 09:13:02 crc kubenswrapper[4789]: I1122 09:13:02.714575 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-xrv8z" podUID="b36c461e-1e80-4f8c-8ec2-74abfb91a4f0" containerName="registry-server" containerID="cri-o://3b3f2eb4aa73d5ae399e38b45af896f083d276f5026692e5b0c159368e1e7d7a" gracePeriod=2 Nov 22 09:13:03 crc kubenswrapper[4789]: I1122 09:13:03.369386 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xrv8z" Nov 22 09:13:03 crc kubenswrapper[4789]: I1122 09:13:03.471637 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b36c461e-1e80-4f8c-8ec2-74abfb91a4f0-utilities\") pod \"b36c461e-1e80-4f8c-8ec2-74abfb91a4f0\" (UID: \"b36c461e-1e80-4f8c-8ec2-74abfb91a4f0\") " Nov 22 09:13:03 crc kubenswrapper[4789]: I1122 09:13:03.471723 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b36c461e-1e80-4f8c-8ec2-74abfb91a4f0-catalog-content\") pod \"b36c461e-1e80-4f8c-8ec2-74abfb91a4f0\" (UID: \"b36c461e-1e80-4f8c-8ec2-74abfb91a4f0\") " Nov 22 09:13:03 crc kubenswrapper[4789]: I1122 09:13:03.471829 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sr5g4\" (UniqueName: \"kubernetes.io/projected/b36c461e-1e80-4f8c-8ec2-74abfb91a4f0-kube-api-access-sr5g4\") pod \"b36c461e-1e80-4f8c-8ec2-74abfb91a4f0\" (UID: \"b36c461e-1e80-4f8c-8ec2-74abfb91a4f0\") " Nov 22 09:13:03 crc kubenswrapper[4789]: I1122 09:13:03.472826 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b36c461e-1e80-4f8c-8ec2-74abfb91a4f0-utilities" (OuterVolumeSpecName: "utilities") pod "b36c461e-1e80-4f8c-8ec2-74abfb91a4f0" (UID: "b36c461e-1e80-4f8c-8ec2-74abfb91a4f0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 09:13:03 crc kubenswrapper[4789]: I1122 09:13:03.480118 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b36c461e-1e80-4f8c-8ec2-74abfb91a4f0-kube-api-access-sr5g4" (OuterVolumeSpecName: "kube-api-access-sr5g4") pod "b36c461e-1e80-4f8c-8ec2-74abfb91a4f0" (UID: "b36c461e-1e80-4f8c-8ec2-74abfb91a4f0"). InnerVolumeSpecName "kube-api-access-sr5g4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 09:13:03 crc kubenswrapper[4789]: I1122 09:13:03.490548 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b36c461e-1e80-4f8c-8ec2-74abfb91a4f0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b36c461e-1e80-4f8c-8ec2-74abfb91a4f0" (UID: "b36c461e-1e80-4f8c-8ec2-74abfb91a4f0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 09:13:03 crc kubenswrapper[4789]: I1122 09:13:03.574557 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b36c461e-1e80-4f8c-8ec2-74abfb91a4f0-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 09:13:03 crc kubenswrapper[4789]: I1122 09:13:03.574601 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b36c461e-1e80-4f8c-8ec2-74abfb91a4f0-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 09:13:03 crc kubenswrapper[4789]: I1122 09:13:03.574617 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sr5g4\" (UniqueName: \"kubernetes.io/projected/b36c461e-1e80-4f8c-8ec2-74abfb91a4f0-kube-api-access-sr5g4\") on node \"crc\" DevicePath \"\"" Nov 22 09:13:03 crc kubenswrapper[4789]: I1122 09:13:03.726617 4789 generic.go:334] "Generic (PLEG): container finished" podID="b36c461e-1e80-4f8c-8ec2-74abfb91a4f0" containerID="3b3f2eb4aa73d5ae399e38b45af896f083d276f5026692e5b0c159368e1e7d7a" exitCode=0 Nov 22 09:13:03 crc kubenswrapper[4789]: I1122 09:13:03.726681 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xrv8z" Nov 22 09:13:03 crc kubenswrapper[4789]: I1122 09:13:03.726683 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xrv8z" event={"ID":"b36c461e-1e80-4f8c-8ec2-74abfb91a4f0","Type":"ContainerDied","Data":"3b3f2eb4aa73d5ae399e38b45af896f083d276f5026692e5b0c159368e1e7d7a"} Nov 22 09:13:03 crc kubenswrapper[4789]: I1122 09:13:03.726791 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xrv8z" event={"ID":"b36c461e-1e80-4f8c-8ec2-74abfb91a4f0","Type":"ContainerDied","Data":"4118d7ecfdbeaa926677b020cebeed696a94a19b7db484273aeb8986d2da0db4"} Nov 22 09:13:03 crc kubenswrapper[4789]: I1122 09:13:03.726825 4789 scope.go:117] "RemoveContainer" containerID="3b3f2eb4aa73d5ae399e38b45af896f083d276f5026692e5b0c159368e1e7d7a" Nov 22 09:13:03 crc kubenswrapper[4789]: I1122 09:13:03.754272 4789 scope.go:117] "RemoveContainer" containerID="6e4edf59ff68bb769e1e16428812845855300da350f17631296437911cee62e7" Nov 22 09:13:03 crc kubenswrapper[4789]: I1122 09:13:03.767912 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xrv8z"] Nov 22 09:13:03 crc kubenswrapper[4789]: I1122 09:13:03.780534 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-xrv8z"] Nov 22 09:13:03 crc kubenswrapper[4789]: I1122 09:13:03.785479 4789 scope.go:117] "RemoveContainer" containerID="6310d94eab8d0f04fd4868283f8104f57138c7b38a1819ee800638c5825ae4f2" Nov 22 09:13:03 crc kubenswrapper[4789]: I1122 09:13:03.830857 4789 scope.go:117] "RemoveContainer" containerID="3b3f2eb4aa73d5ae399e38b45af896f083d276f5026692e5b0c159368e1e7d7a" Nov 22 09:13:03 crc kubenswrapper[4789]: E1122 09:13:03.832152 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3b3f2eb4aa73d5ae399e38b45af896f083d276f5026692e5b0c159368e1e7d7a\": container with ID starting with 3b3f2eb4aa73d5ae399e38b45af896f083d276f5026692e5b0c159368e1e7d7a not found: ID does not exist" containerID="3b3f2eb4aa73d5ae399e38b45af896f083d276f5026692e5b0c159368e1e7d7a" Nov 22 09:13:03 crc kubenswrapper[4789]: I1122 09:13:03.832202 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b3f2eb4aa73d5ae399e38b45af896f083d276f5026692e5b0c159368e1e7d7a"} err="failed to get container status \"3b3f2eb4aa73d5ae399e38b45af896f083d276f5026692e5b0c159368e1e7d7a\": rpc error: code = NotFound desc = could not find container \"3b3f2eb4aa73d5ae399e38b45af896f083d276f5026692e5b0c159368e1e7d7a\": container with ID starting with 3b3f2eb4aa73d5ae399e38b45af896f083d276f5026692e5b0c159368e1e7d7a not found: ID does not exist" Nov 22 09:13:03 crc kubenswrapper[4789]: I1122 09:13:03.832228 4789 scope.go:117] "RemoveContainer" containerID="6e4edf59ff68bb769e1e16428812845855300da350f17631296437911cee62e7" Nov 22 09:13:03 crc kubenswrapper[4789]: E1122 09:13:03.832555 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6e4edf59ff68bb769e1e16428812845855300da350f17631296437911cee62e7\": container with ID starting with 6e4edf59ff68bb769e1e16428812845855300da350f17631296437911cee62e7 not found: ID does not exist" containerID="6e4edf59ff68bb769e1e16428812845855300da350f17631296437911cee62e7" Nov 22 09:13:03 crc kubenswrapper[4789]: I1122 09:13:03.832606 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e4edf59ff68bb769e1e16428812845855300da350f17631296437911cee62e7"} err="failed to get container status \"6e4edf59ff68bb769e1e16428812845855300da350f17631296437911cee62e7\": rpc error: code = NotFound desc = could not find container \"6e4edf59ff68bb769e1e16428812845855300da350f17631296437911cee62e7\": container with ID starting with 6e4edf59ff68bb769e1e16428812845855300da350f17631296437911cee62e7 not found: ID does not exist" Nov 22 09:13:03 crc kubenswrapper[4789]: I1122 09:13:03.832648 4789 scope.go:117] "RemoveContainer" containerID="6310d94eab8d0f04fd4868283f8104f57138c7b38a1819ee800638c5825ae4f2" Nov 22 09:13:03 crc kubenswrapper[4789]: E1122 09:13:03.833038 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6310d94eab8d0f04fd4868283f8104f57138c7b38a1819ee800638c5825ae4f2\": container with ID starting with 6310d94eab8d0f04fd4868283f8104f57138c7b38a1819ee800638c5825ae4f2 not found: ID does not exist" containerID="6310d94eab8d0f04fd4868283f8104f57138c7b38a1819ee800638c5825ae4f2" Nov 22 09:13:03 crc kubenswrapper[4789]: I1122 09:13:03.833072 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6310d94eab8d0f04fd4868283f8104f57138c7b38a1819ee800638c5825ae4f2"} err="failed to get container status \"6310d94eab8d0f04fd4868283f8104f57138c7b38a1819ee800638c5825ae4f2\": rpc error: code = NotFound desc = could not find container \"6310d94eab8d0f04fd4868283f8104f57138c7b38a1819ee800638c5825ae4f2\": container with ID starting with 6310d94eab8d0f04fd4868283f8104f57138c7b38a1819ee800638c5825ae4f2 not found: ID does not exist" Nov 22 09:13:03 crc kubenswrapper[4789]: I1122 09:13:03.979570 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b36c461e-1e80-4f8c-8ec2-74abfb91a4f0" path="/var/lib/kubelet/pods/b36c461e-1e80-4f8c-8ec2-74abfb91a4f0/volumes" Nov 22 09:13:05 crc kubenswrapper[4789]: I1122 09:13:05.965254 4789 scope.go:117] "RemoveContainer" containerID="e9395b7f22fc562bccbab647e93ed06aa21f27b91666ef88922354e2f66bcf54" Nov 22 09:13:05 crc kubenswrapper[4789]: E1122 09:13:05.965604 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:13:17 crc kubenswrapper[4789]: I1122 09:13:17.972766 4789 scope.go:117] "RemoveContainer" containerID="e9395b7f22fc562bccbab647e93ed06aa21f27b91666ef88922354e2f66bcf54" Nov 22 09:13:17 crc kubenswrapper[4789]: E1122 09:13:17.973567 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:13:29 crc kubenswrapper[4789]: I1122 09:13:29.965946 4789 scope.go:117] "RemoveContainer" containerID="e9395b7f22fc562bccbab647e93ed06aa21f27b91666ef88922354e2f66bcf54" Nov 22 09:13:29 crc kubenswrapper[4789]: E1122 09:13:29.967689 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:13:40 crc kubenswrapper[4789]: I1122 09:13:40.964813 4789 scope.go:117] "RemoveContainer" containerID="e9395b7f22fc562bccbab647e93ed06aa21f27b91666ef88922354e2f66bcf54" Nov 22 09:13:42 crc kubenswrapper[4789]: I1122 09:13:42.085907 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" event={"ID":"58aa8071-7ecc-4692-8789-2db04ec70647","Type":"ContainerStarted","Data":"ebe5a3882b79b00f381d091d16cb20463c15e848a3ab9e7c2941a9f043d441b1"} Nov 22 09:14:02 crc kubenswrapper[4789]: I1122 09:14:02.572465 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-kgwvx"] Nov 22 09:14:02 crc kubenswrapper[4789]: E1122 09:14:02.573495 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b36c461e-1e80-4f8c-8ec2-74abfb91a4f0" containerName="extract-utilities" Nov 22 09:14:02 crc kubenswrapper[4789]: I1122 09:14:02.573507 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="b36c461e-1e80-4f8c-8ec2-74abfb91a4f0" containerName="extract-utilities" Nov 22 09:14:02 crc kubenswrapper[4789]: E1122 09:14:02.573528 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b36c461e-1e80-4f8c-8ec2-74abfb91a4f0" containerName="extract-content" Nov 22 09:14:02 crc kubenswrapper[4789]: I1122 09:14:02.573534 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="b36c461e-1e80-4f8c-8ec2-74abfb91a4f0" containerName="extract-content" Nov 22 09:14:02 crc kubenswrapper[4789]: E1122 09:14:02.573561 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b36c461e-1e80-4f8c-8ec2-74abfb91a4f0" containerName="registry-server" Nov 22 09:14:02 crc kubenswrapper[4789]: I1122 09:14:02.573568 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="b36c461e-1e80-4f8c-8ec2-74abfb91a4f0" containerName="registry-server" Nov 22 09:14:02 crc kubenswrapper[4789]: I1122 09:14:02.573857 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="b36c461e-1e80-4f8c-8ec2-74abfb91a4f0" containerName="registry-server" Nov 22 09:14:02 crc kubenswrapper[4789]: I1122 09:14:02.575207 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kgwvx" Nov 22 09:14:02 crc kubenswrapper[4789]: I1122 09:14:02.610148 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kgwvx"] Nov 22 09:14:02 crc kubenswrapper[4789]: I1122 09:14:02.670325 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/79d3d7fe-cc71-49e3-ac0b-5cd4a53bd50e-catalog-content\") pod \"community-operators-kgwvx\" (UID: \"79d3d7fe-cc71-49e3-ac0b-5cd4a53bd50e\") " pod="openshift-marketplace/community-operators-kgwvx" Nov 22 09:14:02 crc kubenswrapper[4789]: I1122 09:14:02.670604 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-flzfc\" (UniqueName: \"kubernetes.io/projected/79d3d7fe-cc71-49e3-ac0b-5cd4a53bd50e-kube-api-access-flzfc\") pod \"community-operators-kgwvx\" (UID: \"79d3d7fe-cc71-49e3-ac0b-5cd4a53bd50e\") " pod="openshift-marketplace/community-operators-kgwvx" Nov 22 09:14:02 crc kubenswrapper[4789]: I1122 09:14:02.670820 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/79d3d7fe-cc71-49e3-ac0b-5cd4a53bd50e-utilities\") pod \"community-operators-kgwvx\" (UID: \"79d3d7fe-cc71-49e3-ac0b-5cd4a53bd50e\") " pod="openshift-marketplace/community-operators-kgwvx" Nov 22 09:14:02 crc kubenswrapper[4789]: I1122 09:14:02.772830 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/79d3d7fe-cc71-49e3-ac0b-5cd4a53bd50e-utilities\") pod \"community-operators-kgwvx\" (UID: \"79d3d7fe-cc71-49e3-ac0b-5cd4a53bd50e\") " pod="openshift-marketplace/community-operators-kgwvx" Nov 22 09:14:02 crc kubenswrapper[4789]: I1122 09:14:02.772933 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/79d3d7fe-cc71-49e3-ac0b-5cd4a53bd50e-catalog-content\") pod \"community-operators-kgwvx\" (UID: \"79d3d7fe-cc71-49e3-ac0b-5cd4a53bd50e\") " pod="openshift-marketplace/community-operators-kgwvx" Nov 22 09:14:02 crc kubenswrapper[4789]: I1122 09:14:02.772996 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-flzfc\" (UniqueName: \"kubernetes.io/projected/79d3d7fe-cc71-49e3-ac0b-5cd4a53bd50e-kube-api-access-flzfc\") pod \"community-operators-kgwvx\" (UID: \"79d3d7fe-cc71-49e3-ac0b-5cd4a53bd50e\") " pod="openshift-marketplace/community-operators-kgwvx" Nov 22 09:14:02 crc kubenswrapper[4789]: I1122 09:14:02.773816 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/79d3d7fe-cc71-49e3-ac0b-5cd4a53bd50e-utilities\") pod \"community-operators-kgwvx\" (UID: \"79d3d7fe-cc71-49e3-ac0b-5cd4a53bd50e\") " pod="openshift-marketplace/community-operators-kgwvx" Nov 22 09:14:02 crc kubenswrapper[4789]: I1122 09:14:02.773860 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/79d3d7fe-cc71-49e3-ac0b-5cd4a53bd50e-catalog-content\") pod \"community-operators-kgwvx\" (UID: \"79d3d7fe-cc71-49e3-ac0b-5cd4a53bd50e\") " pod="openshift-marketplace/community-operators-kgwvx" Nov 22 09:14:02 crc kubenswrapper[4789]: I1122 09:14:02.797309 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-flzfc\" (UniqueName: \"kubernetes.io/projected/79d3d7fe-cc71-49e3-ac0b-5cd4a53bd50e-kube-api-access-flzfc\") pod \"community-operators-kgwvx\" (UID: \"79d3d7fe-cc71-49e3-ac0b-5cd4a53bd50e\") " pod="openshift-marketplace/community-operators-kgwvx" Nov 22 09:14:02 crc kubenswrapper[4789]: I1122 09:14:02.896927 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kgwvx" Nov 22 09:14:03 crc kubenswrapper[4789]: I1122 09:14:03.668915 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kgwvx"] Nov 22 09:14:04 crc kubenswrapper[4789]: I1122 09:14:04.272142 4789 generic.go:334] "Generic (PLEG): container finished" podID="79d3d7fe-cc71-49e3-ac0b-5cd4a53bd50e" containerID="0109ee6fba0813922900ece8607b961bb7f91a935e0440a959972a58cc67854a" exitCode=0 Nov 22 09:14:04 crc kubenswrapper[4789]: I1122 09:14:04.272198 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kgwvx" event={"ID":"79d3d7fe-cc71-49e3-ac0b-5cd4a53bd50e","Type":"ContainerDied","Data":"0109ee6fba0813922900ece8607b961bb7f91a935e0440a959972a58cc67854a"} Nov 22 09:14:04 crc kubenswrapper[4789]: I1122 09:14:04.272624 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kgwvx" event={"ID":"79d3d7fe-cc71-49e3-ac0b-5cd4a53bd50e","Type":"ContainerStarted","Data":"2b01c0683166c91bb19618d7f9f0562b725204fcaf90815b26379e41db5df994"} Nov 22 09:14:06 crc kubenswrapper[4789]: I1122 09:14:06.294856 4789 generic.go:334] "Generic (PLEG): container finished" podID="79d3d7fe-cc71-49e3-ac0b-5cd4a53bd50e" containerID="f4083446f5ccdd2663c2db43b5ca71249adf9765730c7c1400fb2f01ba3bc3ce" exitCode=0 Nov 22 09:14:06 crc kubenswrapper[4789]: I1122 09:14:06.294891 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kgwvx" event={"ID":"79d3d7fe-cc71-49e3-ac0b-5cd4a53bd50e","Type":"ContainerDied","Data":"f4083446f5ccdd2663c2db43b5ca71249adf9765730c7c1400fb2f01ba3bc3ce"} Nov 22 09:14:08 crc kubenswrapper[4789]: I1122 09:14:08.315073 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kgwvx" event={"ID":"79d3d7fe-cc71-49e3-ac0b-5cd4a53bd50e","Type":"ContainerStarted","Data":"980ae25ecdee84be95f686c4fa0bbd279dd25857b3a23bf9f11e0a8e2a08377e"} Nov 22 09:14:08 crc kubenswrapper[4789]: I1122 09:14:08.339001 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-kgwvx" podStartSLOduration=3.531950874 podStartE2EDuration="6.338865298s" podCreationTimestamp="2025-11-22 09:14:02 +0000 UTC" firstStartedPulling="2025-11-22 09:14:04.283490505 +0000 UTC m=+4658.517890778" lastFinishedPulling="2025-11-22 09:14:07.090404919 +0000 UTC m=+4661.324805202" observedRunningTime="2025-11-22 09:14:08.331069591 +0000 UTC m=+4662.565469874" watchObservedRunningTime="2025-11-22 09:14:08.338865298 +0000 UTC m=+4662.573265571" Nov 22 09:14:12 crc kubenswrapper[4789]: I1122 09:14:12.897262 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-kgwvx" Nov 22 09:14:12 crc kubenswrapper[4789]: I1122 09:14:12.897817 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-kgwvx" Nov 22 09:14:12 crc kubenswrapper[4789]: I1122 09:14:12.945599 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-kgwvx" Nov 22 09:14:13 crc kubenswrapper[4789]: I1122 09:14:13.428975 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-kgwvx" Nov 22 09:14:13 crc kubenswrapper[4789]: I1122 09:14:13.494473 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-kgwvx"] Nov 22 09:14:15 crc kubenswrapper[4789]: I1122 09:14:15.368305 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-kgwvx" podUID="79d3d7fe-cc71-49e3-ac0b-5cd4a53bd50e" containerName="registry-server" containerID="cri-o://980ae25ecdee84be95f686c4fa0bbd279dd25857b3a23bf9f11e0a8e2a08377e" gracePeriod=2 Nov 22 09:14:16 crc kubenswrapper[4789]: I1122 09:14:16.069235 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kgwvx" Nov 22 09:14:16 crc kubenswrapper[4789]: I1122 09:14:16.153500 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/79d3d7fe-cc71-49e3-ac0b-5cd4a53bd50e-utilities\") pod \"79d3d7fe-cc71-49e3-ac0b-5cd4a53bd50e\" (UID: \"79d3d7fe-cc71-49e3-ac0b-5cd4a53bd50e\") " Nov 22 09:14:16 crc kubenswrapper[4789]: I1122 09:14:16.153641 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/79d3d7fe-cc71-49e3-ac0b-5cd4a53bd50e-catalog-content\") pod \"79d3d7fe-cc71-49e3-ac0b-5cd4a53bd50e\" (UID: \"79d3d7fe-cc71-49e3-ac0b-5cd4a53bd50e\") " Nov 22 09:14:16 crc kubenswrapper[4789]: I1122 09:14:16.153924 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-flzfc\" (UniqueName: \"kubernetes.io/projected/79d3d7fe-cc71-49e3-ac0b-5cd4a53bd50e-kube-api-access-flzfc\") pod \"79d3d7fe-cc71-49e3-ac0b-5cd4a53bd50e\" (UID: \"79d3d7fe-cc71-49e3-ac0b-5cd4a53bd50e\") " Nov 22 09:14:16 crc kubenswrapper[4789]: I1122 09:14:16.154703 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/79d3d7fe-cc71-49e3-ac0b-5cd4a53bd50e-utilities" (OuterVolumeSpecName: "utilities") pod "79d3d7fe-cc71-49e3-ac0b-5cd4a53bd50e" (UID: "79d3d7fe-cc71-49e3-ac0b-5cd4a53bd50e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 09:14:16 crc kubenswrapper[4789]: I1122 09:14:16.159598 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79d3d7fe-cc71-49e3-ac0b-5cd4a53bd50e-kube-api-access-flzfc" (OuterVolumeSpecName: "kube-api-access-flzfc") pod "79d3d7fe-cc71-49e3-ac0b-5cd4a53bd50e" (UID: "79d3d7fe-cc71-49e3-ac0b-5cd4a53bd50e"). InnerVolumeSpecName "kube-api-access-flzfc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 09:14:16 crc kubenswrapper[4789]: I1122 09:14:16.216185 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/79d3d7fe-cc71-49e3-ac0b-5cd4a53bd50e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "79d3d7fe-cc71-49e3-ac0b-5cd4a53bd50e" (UID: "79d3d7fe-cc71-49e3-ac0b-5cd4a53bd50e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 09:14:16 crc kubenswrapper[4789]: I1122 09:14:16.256090 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/79d3d7fe-cc71-49e3-ac0b-5cd4a53bd50e-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 09:14:16 crc kubenswrapper[4789]: I1122 09:14:16.256120 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-flzfc\" (UniqueName: \"kubernetes.io/projected/79d3d7fe-cc71-49e3-ac0b-5cd4a53bd50e-kube-api-access-flzfc\") on node \"crc\" DevicePath \"\"" Nov 22 09:14:16 crc kubenswrapper[4789]: I1122 09:14:16.256131 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/79d3d7fe-cc71-49e3-ac0b-5cd4a53bd50e-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 09:14:16 crc kubenswrapper[4789]: I1122 09:14:16.377772 4789 generic.go:334] "Generic (PLEG): container finished" podID="79d3d7fe-cc71-49e3-ac0b-5cd4a53bd50e" containerID="980ae25ecdee84be95f686c4fa0bbd279dd25857b3a23bf9f11e0a8e2a08377e" exitCode=0 Nov 22 09:14:16 crc kubenswrapper[4789]: I1122 09:14:16.377813 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kgwvx" event={"ID":"79d3d7fe-cc71-49e3-ac0b-5cd4a53bd50e","Type":"ContainerDied","Data":"980ae25ecdee84be95f686c4fa0bbd279dd25857b3a23bf9f11e0a8e2a08377e"} Nov 22 09:14:16 crc kubenswrapper[4789]: I1122 09:14:16.377838 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kgwvx" event={"ID":"79d3d7fe-cc71-49e3-ac0b-5cd4a53bd50e","Type":"ContainerDied","Data":"2b01c0683166c91bb19618d7f9f0562b725204fcaf90815b26379e41db5df994"} Nov 22 09:14:16 crc kubenswrapper[4789]: I1122 09:14:16.377859 4789 scope.go:117] "RemoveContainer" containerID="980ae25ecdee84be95f686c4fa0bbd279dd25857b3a23bf9f11e0a8e2a08377e" Nov 22 09:14:16 crc kubenswrapper[4789]: I1122 09:14:16.377982 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kgwvx" Nov 22 09:14:16 crc kubenswrapper[4789]: I1122 09:14:16.405174 4789 scope.go:117] "RemoveContainer" containerID="f4083446f5ccdd2663c2db43b5ca71249adf9765730c7c1400fb2f01ba3bc3ce" Nov 22 09:14:16 crc kubenswrapper[4789]: I1122 09:14:16.412622 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-kgwvx"] Nov 22 09:14:16 crc kubenswrapper[4789]: I1122 09:14:16.421381 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-kgwvx"] Nov 22 09:14:16 crc kubenswrapper[4789]: I1122 09:14:16.437813 4789 scope.go:117] "RemoveContainer" containerID="0109ee6fba0813922900ece8607b961bb7f91a935e0440a959972a58cc67854a" Nov 22 09:14:16 crc kubenswrapper[4789]: I1122 09:14:16.483648 4789 scope.go:117] "RemoveContainer" containerID="980ae25ecdee84be95f686c4fa0bbd279dd25857b3a23bf9f11e0a8e2a08377e" Nov 22 09:14:16 crc kubenswrapper[4789]: E1122 09:14:16.484275 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"980ae25ecdee84be95f686c4fa0bbd279dd25857b3a23bf9f11e0a8e2a08377e\": container with ID starting with 980ae25ecdee84be95f686c4fa0bbd279dd25857b3a23bf9f11e0a8e2a08377e not found: ID does not exist" containerID="980ae25ecdee84be95f686c4fa0bbd279dd25857b3a23bf9f11e0a8e2a08377e" Nov 22 09:14:16 crc kubenswrapper[4789]: I1122 09:14:16.484317 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"980ae25ecdee84be95f686c4fa0bbd279dd25857b3a23bf9f11e0a8e2a08377e"} err="failed to get container status \"980ae25ecdee84be95f686c4fa0bbd279dd25857b3a23bf9f11e0a8e2a08377e\": rpc error: code = NotFound desc = could not find container \"980ae25ecdee84be95f686c4fa0bbd279dd25857b3a23bf9f11e0a8e2a08377e\": container with ID starting with 980ae25ecdee84be95f686c4fa0bbd279dd25857b3a23bf9f11e0a8e2a08377e not found: ID does not exist" Nov 22 09:14:16 crc kubenswrapper[4789]: I1122 09:14:16.484347 4789 scope.go:117] "RemoveContainer" containerID="f4083446f5ccdd2663c2db43b5ca71249adf9765730c7c1400fb2f01ba3bc3ce" Nov 22 09:14:16 crc kubenswrapper[4789]: E1122 09:14:16.484886 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f4083446f5ccdd2663c2db43b5ca71249adf9765730c7c1400fb2f01ba3bc3ce\": container with ID starting with f4083446f5ccdd2663c2db43b5ca71249adf9765730c7c1400fb2f01ba3bc3ce not found: ID does not exist" containerID="f4083446f5ccdd2663c2db43b5ca71249adf9765730c7c1400fb2f01ba3bc3ce" Nov 22 09:14:16 crc kubenswrapper[4789]: I1122 09:14:16.484919 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f4083446f5ccdd2663c2db43b5ca71249adf9765730c7c1400fb2f01ba3bc3ce"} err="failed to get container status \"f4083446f5ccdd2663c2db43b5ca71249adf9765730c7c1400fb2f01ba3bc3ce\": rpc error: code = NotFound desc = could not find container \"f4083446f5ccdd2663c2db43b5ca71249adf9765730c7c1400fb2f01ba3bc3ce\": container with ID starting with f4083446f5ccdd2663c2db43b5ca71249adf9765730c7c1400fb2f01ba3bc3ce not found: ID does not exist" Nov 22 09:14:16 crc kubenswrapper[4789]: I1122 09:14:16.484941 4789 scope.go:117] "RemoveContainer" containerID="0109ee6fba0813922900ece8607b961bb7f91a935e0440a959972a58cc67854a" Nov 22 09:14:16 crc kubenswrapper[4789]: E1122 09:14:16.485235 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0109ee6fba0813922900ece8607b961bb7f91a935e0440a959972a58cc67854a\": container with ID starting with 0109ee6fba0813922900ece8607b961bb7f91a935e0440a959972a58cc67854a not found: ID does not exist" containerID="0109ee6fba0813922900ece8607b961bb7f91a935e0440a959972a58cc67854a" Nov 22 09:14:16 crc kubenswrapper[4789]: I1122 09:14:16.485255 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0109ee6fba0813922900ece8607b961bb7f91a935e0440a959972a58cc67854a"} err="failed to get container status \"0109ee6fba0813922900ece8607b961bb7f91a935e0440a959972a58cc67854a\": rpc error: code = NotFound desc = could not find container \"0109ee6fba0813922900ece8607b961bb7f91a935e0440a959972a58cc67854a\": container with ID starting with 0109ee6fba0813922900ece8607b961bb7f91a935e0440a959972a58cc67854a not found: ID does not exist" Nov 22 09:14:17 crc kubenswrapper[4789]: I1122 09:14:17.977485 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="79d3d7fe-cc71-49e3-ac0b-5cd4a53bd50e" path="/var/lib/kubelet/pods/79d3d7fe-cc71-49e3-ac0b-5cd4a53bd50e/volumes" Nov 22 09:15:00 crc kubenswrapper[4789]: I1122 09:15:00.144883 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396715-qcwt2"] Nov 22 09:15:00 crc kubenswrapper[4789]: E1122 09:15:00.146125 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79d3d7fe-cc71-49e3-ac0b-5cd4a53bd50e" containerName="extract-content" Nov 22 09:15:00 crc kubenswrapper[4789]: I1122 09:15:00.146144 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="79d3d7fe-cc71-49e3-ac0b-5cd4a53bd50e" containerName="extract-content" Nov 22 09:15:00 crc kubenswrapper[4789]: E1122 09:15:00.146165 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79d3d7fe-cc71-49e3-ac0b-5cd4a53bd50e" containerName="extract-utilities" Nov 22 09:15:00 crc kubenswrapper[4789]: I1122 09:15:00.146173 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="79d3d7fe-cc71-49e3-ac0b-5cd4a53bd50e" containerName="extract-utilities" Nov 22 09:15:00 crc kubenswrapper[4789]: E1122 09:15:00.146212 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79d3d7fe-cc71-49e3-ac0b-5cd4a53bd50e" containerName="registry-server" Nov 22 09:15:00 crc kubenswrapper[4789]: I1122 09:15:00.146221 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="79d3d7fe-cc71-49e3-ac0b-5cd4a53bd50e" containerName="registry-server" Nov 22 09:15:00 crc kubenswrapper[4789]: I1122 09:15:00.146440 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="79d3d7fe-cc71-49e3-ac0b-5cd4a53bd50e" containerName="registry-server" Nov 22 09:15:00 crc kubenswrapper[4789]: I1122 09:15:00.147244 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396715-qcwt2" Nov 22 09:15:00 crc kubenswrapper[4789]: I1122 09:15:00.149558 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 22 09:15:00 crc kubenswrapper[4789]: I1122 09:15:00.149929 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 22 09:15:00 crc kubenswrapper[4789]: I1122 09:15:00.159605 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396715-qcwt2"] Nov 22 09:15:00 crc kubenswrapper[4789]: I1122 09:15:00.208563 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/55bbc3c2-372c-4ce9-bbae-4b111aa57ca7-config-volume\") pod \"collect-profiles-29396715-qcwt2\" (UID: \"55bbc3c2-372c-4ce9-bbae-4b111aa57ca7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396715-qcwt2" Nov 22 09:15:00 crc kubenswrapper[4789]: I1122 09:15:00.208670 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/55bbc3c2-372c-4ce9-bbae-4b111aa57ca7-secret-volume\") pod \"collect-profiles-29396715-qcwt2\" (UID: \"55bbc3c2-372c-4ce9-bbae-4b111aa57ca7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396715-qcwt2" Nov 22 09:15:00 crc kubenswrapper[4789]: I1122 09:15:00.208780 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-825ph\" (UniqueName: \"kubernetes.io/projected/55bbc3c2-372c-4ce9-bbae-4b111aa57ca7-kube-api-access-825ph\") pod \"collect-profiles-29396715-qcwt2\" (UID: \"55bbc3c2-372c-4ce9-bbae-4b111aa57ca7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396715-qcwt2" Nov 22 09:15:00 crc kubenswrapper[4789]: I1122 09:15:00.311703 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-825ph\" (UniqueName: \"kubernetes.io/projected/55bbc3c2-372c-4ce9-bbae-4b111aa57ca7-kube-api-access-825ph\") pod \"collect-profiles-29396715-qcwt2\" (UID: \"55bbc3c2-372c-4ce9-bbae-4b111aa57ca7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396715-qcwt2" Nov 22 09:15:00 crc kubenswrapper[4789]: I1122 09:15:00.312000 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/55bbc3c2-372c-4ce9-bbae-4b111aa57ca7-config-volume\") pod \"collect-profiles-29396715-qcwt2\" (UID: \"55bbc3c2-372c-4ce9-bbae-4b111aa57ca7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396715-qcwt2" Nov 22 09:15:00 crc kubenswrapper[4789]: I1122 09:15:00.312075 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/55bbc3c2-372c-4ce9-bbae-4b111aa57ca7-secret-volume\") pod \"collect-profiles-29396715-qcwt2\" (UID: \"55bbc3c2-372c-4ce9-bbae-4b111aa57ca7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396715-qcwt2" Nov 22 09:15:00 crc kubenswrapper[4789]: I1122 09:15:00.313362 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/55bbc3c2-372c-4ce9-bbae-4b111aa57ca7-config-volume\") pod \"collect-profiles-29396715-qcwt2\" (UID: \"55bbc3c2-372c-4ce9-bbae-4b111aa57ca7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396715-qcwt2" Nov 22 09:15:00 crc kubenswrapper[4789]: I1122 09:15:00.318914 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/55bbc3c2-372c-4ce9-bbae-4b111aa57ca7-secret-volume\") pod \"collect-profiles-29396715-qcwt2\" (UID: \"55bbc3c2-372c-4ce9-bbae-4b111aa57ca7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396715-qcwt2" Nov 22 09:15:00 crc kubenswrapper[4789]: I1122 09:15:00.353197 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-825ph\" (UniqueName: \"kubernetes.io/projected/55bbc3c2-372c-4ce9-bbae-4b111aa57ca7-kube-api-access-825ph\") pod \"collect-profiles-29396715-qcwt2\" (UID: \"55bbc3c2-372c-4ce9-bbae-4b111aa57ca7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396715-qcwt2" Nov 22 09:15:00 crc kubenswrapper[4789]: I1122 09:15:00.473244 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396715-qcwt2" Nov 22 09:15:00 crc kubenswrapper[4789]: I1122 09:15:00.987604 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396715-qcwt2"] Nov 22 09:15:01 crc kubenswrapper[4789]: I1122 09:15:01.751864 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29396715-qcwt2" event={"ID":"55bbc3c2-372c-4ce9-bbae-4b111aa57ca7","Type":"ContainerStarted","Data":"abdc68875fb3d8f78fcd55346727187c234742fc0f7da6c9ef937ec8ca749d8d"} Nov 22 09:15:01 crc kubenswrapper[4789]: I1122 09:15:01.752319 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29396715-qcwt2" event={"ID":"55bbc3c2-372c-4ce9-bbae-4b111aa57ca7","Type":"ContainerStarted","Data":"3af97b5edfeac3309adb08053d413db33f4739c2616af29633bfc9ff6c2b7d61"} Nov 22 09:15:01 crc kubenswrapper[4789]: I1122 09:15:01.780161 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29396715-qcwt2" podStartSLOduration=1.780136248 podStartE2EDuration="1.780136248s" podCreationTimestamp="2025-11-22 09:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 09:15:01.767788589 +0000 UTC m=+4716.002188862" watchObservedRunningTime="2025-11-22 09:15:01.780136248 +0000 UTC m=+4716.014536521" Nov 22 09:15:02 crc kubenswrapper[4789]: I1122 09:15:02.762380 4789 generic.go:334] "Generic (PLEG): container finished" podID="55bbc3c2-372c-4ce9-bbae-4b111aa57ca7" containerID="abdc68875fb3d8f78fcd55346727187c234742fc0f7da6c9ef937ec8ca749d8d" exitCode=0 Nov 22 09:15:02 crc kubenswrapper[4789]: I1122 09:15:02.762424 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29396715-qcwt2" event={"ID":"55bbc3c2-372c-4ce9-bbae-4b111aa57ca7","Type":"ContainerDied","Data":"abdc68875fb3d8f78fcd55346727187c234742fc0f7da6c9ef937ec8ca749d8d"} Nov 22 09:15:04 crc kubenswrapper[4789]: I1122 09:15:04.311375 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396715-qcwt2" Nov 22 09:15:04 crc kubenswrapper[4789]: I1122 09:15:04.396861 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/55bbc3c2-372c-4ce9-bbae-4b111aa57ca7-secret-volume\") pod \"55bbc3c2-372c-4ce9-bbae-4b111aa57ca7\" (UID: \"55bbc3c2-372c-4ce9-bbae-4b111aa57ca7\") " Nov 22 09:15:04 crc kubenswrapper[4789]: I1122 09:15:04.397561 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/55bbc3c2-372c-4ce9-bbae-4b111aa57ca7-config-volume\") pod \"55bbc3c2-372c-4ce9-bbae-4b111aa57ca7\" (UID: \"55bbc3c2-372c-4ce9-bbae-4b111aa57ca7\") " Nov 22 09:15:04 crc kubenswrapper[4789]: I1122 09:15:04.398161 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-825ph\" (UniqueName: \"kubernetes.io/projected/55bbc3c2-372c-4ce9-bbae-4b111aa57ca7-kube-api-access-825ph\") pod \"55bbc3c2-372c-4ce9-bbae-4b111aa57ca7\" (UID: \"55bbc3c2-372c-4ce9-bbae-4b111aa57ca7\") " Nov 22 09:15:04 crc kubenswrapper[4789]: I1122 09:15:04.398490 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/55bbc3c2-372c-4ce9-bbae-4b111aa57ca7-config-volume" (OuterVolumeSpecName: "config-volume") pod "55bbc3c2-372c-4ce9-bbae-4b111aa57ca7" (UID: "55bbc3c2-372c-4ce9-bbae-4b111aa57ca7"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 09:15:04 crc kubenswrapper[4789]: I1122 09:15:04.399203 4789 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/55bbc3c2-372c-4ce9-bbae-4b111aa57ca7-config-volume\") on node \"crc\" DevicePath \"\"" Nov 22 09:15:04 crc kubenswrapper[4789]: I1122 09:15:04.409030 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55bbc3c2-372c-4ce9-bbae-4b111aa57ca7-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "55bbc3c2-372c-4ce9-bbae-4b111aa57ca7" (UID: "55bbc3c2-372c-4ce9-bbae-4b111aa57ca7"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 09:15:04 crc kubenswrapper[4789]: I1122 09:15:04.412462 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/55bbc3c2-372c-4ce9-bbae-4b111aa57ca7-kube-api-access-825ph" (OuterVolumeSpecName: "kube-api-access-825ph") pod "55bbc3c2-372c-4ce9-bbae-4b111aa57ca7" (UID: "55bbc3c2-372c-4ce9-bbae-4b111aa57ca7"). InnerVolumeSpecName "kube-api-access-825ph". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 09:15:04 crc kubenswrapper[4789]: I1122 09:15:04.503072 4789 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/55bbc3c2-372c-4ce9-bbae-4b111aa57ca7-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 22 09:15:04 crc kubenswrapper[4789]: I1122 09:15:04.503113 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-825ph\" (UniqueName: \"kubernetes.io/projected/55bbc3c2-372c-4ce9-bbae-4b111aa57ca7-kube-api-access-825ph\") on node \"crc\" DevicePath \"\"" Nov 22 09:15:04 crc kubenswrapper[4789]: I1122 09:15:04.783561 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29396715-qcwt2" event={"ID":"55bbc3c2-372c-4ce9-bbae-4b111aa57ca7","Type":"ContainerDied","Data":"3af97b5edfeac3309adb08053d413db33f4739c2616af29633bfc9ff6c2b7d61"} Nov 22 09:15:04 crc kubenswrapper[4789]: I1122 09:15:04.783611 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3af97b5edfeac3309adb08053d413db33f4739c2616af29633bfc9ff6c2b7d61" Nov 22 09:15:04 crc kubenswrapper[4789]: I1122 09:15:04.784150 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396715-qcwt2" Nov 22 09:15:04 crc kubenswrapper[4789]: I1122 09:15:04.847266 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396670-ltqvr"] Nov 22 09:15:04 crc kubenswrapper[4789]: I1122 09:15:04.855360 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396670-ltqvr"] Nov 22 09:15:05 crc kubenswrapper[4789]: I1122 09:15:05.977690 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="37c2f280-272f-45f0-993e-7d840ebf3b92" path="/var/lib/kubelet/pods/37c2f280-272f-45f0-993e-7d840ebf3b92/volumes" Nov 22 09:15:28 crc kubenswrapper[4789]: I1122 09:15:28.836163 4789 scope.go:117] "RemoveContainer" containerID="f9252464a45b6db0e11c046144f08da1b92ba249212199e6a571d3d7c1b15f13" Nov 22 09:15:43 crc kubenswrapper[4789]: I1122 09:15:43.269313 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-l8xgd"] Nov 22 09:15:43 crc kubenswrapper[4789]: E1122 09:15:43.270400 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55bbc3c2-372c-4ce9-bbae-4b111aa57ca7" containerName="collect-profiles" Nov 22 09:15:43 crc kubenswrapper[4789]: I1122 09:15:43.270421 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="55bbc3c2-372c-4ce9-bbae-4b111aa57ca7" containerName="collect-profiles" Nov 22 09:15:43 crc kubenswrapper[4789]: I1122 09:15:43.270653 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="55bbc3c2-372c-4ce9-bbae-4b111aa57ca7" containerName="collect-profiles" Nov 22 09:15:43 crc kubenswrapper[4789]: I1122 09:15:43.272024 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l8xgd" Nov 22 09:15:43 crc kubenswrapper[4789]: I1122 09:15:43.287870 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-l8xgd"] Nov 22 09:15:43 crc kubenswrapper[4789]: I1122 09:15:43.396046 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lm9rx\" (UniqueName: \"kubernetes.io/projected/6e49577b-5b45-455c-a135-395fa2dd2982-kube-api-access-lm9rx\") pod \"redhat-operators-l8xgd\" (UID: \"6e49577b-5b45-455c-a135-395fa2dd2982\") " pod="openshift-marketplace/redhat-operators-l8xgd" Nov 22 09:15:43 crc kubenswrapper[4789]: I1122 09:15:43.396583 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e49577b-5b45-455c-a135-395fa2dd2982-catalog-content\") pod \"redhat-operators-l8xgd\" (UID: \"6e49577b-5b45-455c-a135-395fa2dd2982\") " pod="openshift-marketplace/redhat-operators-l8xgd" Nov 22 09:15:43 crc kubenswrapper[4789]: I1122 09:15:43.396772 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e49577b-5b45-455c-a135-395fa2dd2982-utilities\") pod \"redhat-operators-l8xgd\" (UID: \"6e49577b-5b45-455c-a135-395fa2dd2982\") " pod="openshift-marketplace/redhat-operators-l8xgd" Nov 22 09:15:43 crc kubenswrapper[4789]: I1122 09:15:43.499023 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e49577b-5b45-455c-a135-395fa2dd2982-catalog-content\") pod \"redhat-operators-l8xgd\" (UID: \"6e49577b-5b45-455c-a135-395fa2dd2982\") " pod="openshift-marketplace/redhat-operators-l8xgd" Nov 22 09:15:43 crc kubenswrapper[4789]: I1122 09:15:43.499131 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e49577b-5b45-455c-a135-395fa2dd2982-utilities\") pod \"redhat-operators-l8xgd\" (UID: \"6e49577b-5b45-455c-a135-395fa2dd2982\") " pod="openshift-marketplace/redhat-operators-l8xgd" Nov 22 09:15:43 crc kubenswrapper[4789]: I1122 09:15:43.499226 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lm9rx\" (UniqueName: \"kubernetes.io/projected/6e49577b-5b45-455c-a135-395fa2dd2982-kube-api-access-lm9rx\") pod \"redhat-operators-l8xgd\" (UID: \"6e49577b-5b45-455c-a135-395fa2dd2982\") " pod="openshift-marketplace/redhat-operators-l8xgd" Nov 22 09:15:43 crc kubenswrapper[4789]: I1122 09:15:43.499471 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e49577b-5b45-455c-a135-395fa2dd2982-catalog-content\") pod \"redhat-operators-l8xgd\" (UID: \"6e49577b-5b45-455c-a135-395fa2dd2982\") " pod="openshift-marketplace/redhat-operators-l8xgd" Nov 22 09:15:43 crc kubenswrapper[4789]: I1122 09:15:43.499600 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e49577b-5b45-455c-a135-395fa2dd2982-utilities\") pod \"redhat-operators-l8xgd\" (UID: \"6e49577b-5b45-455c-a135-395fa2dd2982\") " pod="openshift-marketplace/redhat-operators-l8xgd" Nov 22 09:15:43 crc kubenswrapper[4789]: I1122 09:15:43.521126 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lm9rx\" (UniqueName: \"kubernetes.io/projected/6e49577b-5b45-455c-a135-395fa2dd2982-kube-api-access-lm9rx\") pod \"redhat-operators-l8xgd\" (UID: \"6e49577b-5b45-455c-a135-395fa2dd2982\") " pod="openshift-marketplace/redhat-operators-l8xgd" Nov 22 09:15:43 crc kubenswrapper[4789]: I1122 09:15:43.609125 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l8xgd" Nov 22 09:15:44 crc kubenswrapper[4789]: I1122 09:15:44.124816 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-l8xgd"] Nov 22 09:15:44 crc kubenswrapper[4789]: I1122 09:15:44.168386 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l8xgd" event={"ID":"6e49577b-5b45-455c-a135-395fa2dd2982","Type":"ContainerStarted","Data":"4903aac84e03fdcdba4598c3f6c82a929971705b6771b43e192ba2400a7947c4"} Nov 22 09:15:46 crc kubenswrapper[4789]: I1122 09:15:46.187515 4789 generic.go:334] "Generic (PLEG): container finished" podID="6e49577b-5b45-455c-a135-395fa2dd2982" containerID="e073791210d7347beb335e4ab171298ef6ecafad70d0bc17592c679359116b39" exitCode=0 Nov 22 09:15:46 crc kubenswrapper[4789]: I1122 09:15:46.187646 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l8xgd" event={"ID":"6e49577b-5b45-455c-a135-395fa2dd2982","Type":"ContainerDied","Data":"e073791210d7347beb335e4ab171298ef6ecafad70d0bc17592c679359116b39"} Nov 22 09:15:49 crc kubenswrapper[4789]: I1122 09:15:49.219987 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l8xgd" event={"ID":"6e49577b-5b45-455c-a135-395fa2dd2982","Type":"ContainerStarted","Data":"8ba36933cc0c1c94bad05fccdce9b7d606b5458ac79883e9b4c54e949a74ebb0"} Nov 22 09:15:50 crc kubenswrapper[4789]: I1122 09:15:50.230610 4789 generic.go:334] "Generic (PLEG): container finished" podID="6e49577b-5b45-455c-a135-395fa2dd2982" containerID="8ba36933cc0c1c94bad05fccdce9b7d606b5458ac79883e9b4c54e949a74ebb0" exitCode=0 Nov 22 09:15:50 crc kubenswrapper[4789]: I1122 09:15:50.230687 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l8xgd" event={"ID":"6e49577b-5b45-455c-a135-395fa2dd2982","Type":"ContainerDied","Data":"8ba36933cc0c1c94bad05fccdce9b7d606b5458ac79883e9b4c54e949a74ebb0"} Nov 22 09:15:57 crc kubenswrapper[4789]: I1122 09:15:57.309724 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l8xgd" event={"ID":"6e49577b-5b45-455c-a135-395fa2dd2982","Type":"ContainerStarted","Data":"9831bf52ba1ea2d010c0a8c04def331e9b9177aeb9a1aa99232d55f16041e4d0"} Nov 22 09:15:57 crc kubenswrapper[4789]: I1122 09:15:57.331931 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-l8xgd" podStartSLOduration=3.766794398 podStartE2EDuration="14.33190608s" podCreationTimestamp="2025-11-22 09:15:43 +0000 UTC" firstStartedPulling="2025-11-22 09:15:46.189413288 +0000 UTC m=+4760.423813561" lastFinishedPulling="2025-11-22 09:15:56.75452497 +0000 UTC m=+4770.988925243" observedRunningTime="2025-11-22 09:15:57.324410201 +0000 UTC m=+4771.558810494" watchObservedRunningTime="2025-11-22 09:15:57.33190608 +0000 UTC m=+4771.566306353" Nov 22 09:16:03 crc kubenswrapper[4789]: I1122 09:16:03.810148 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-l8xgd" Nov 22 09:16:03 crc kubenswrapper[4789]: I1122 09:16:03.811449 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-l8xgd" Nov 22 09:16:03 crc kubenswrapper[4789]: I1122 09:16:03.934718 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-l8xgd" Nov 22 09:16:04 crc kubenswrapper[4789]: I1122 09:16:04.418161 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-l8xgd" Nov 22 09:16:04 crc kubenswrapper[4789]: I1122 09:16:04.464399 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-l8xgd"] Nov 22 09:16:05 crc kubenswrapper[4789]: I1122 09:16:05.372145 4789 patch_prober.go:28] interesting pod/machine-config-daemon-pxvrm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 09:16:05 crc kubenswrapper[4789]: I1122 09:16:05.373445 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 09:16:06 crc kubenswrapper[4789]: I1122 09:16:06.381349 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-l8xgd" podUID="6e49577b-5b45-455c-a135-395fa2dd2982" containerName="registry-server" containerID="cri-o://9831bf52ba1ea2d010c0a8c04def331e9b9177aeb9a1aa99232d55f16041e4d0" gracePeriod=2 Nov 22 09:16:07 crc kubenswrapper[4789]: I1122 09:16:07.009423 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l8xgd" Nov 22 09:16:07 crc kubenswrapper[4789]: I1122 09:16:07.174649 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e49577b-5b45-455c-a135-395fa2dd2982-utilities\") pod \"6e49577b-5b45-455c-a135-395fa2dd2982\" (UID: \"6e49577b-5b45-455c-a135-395fa2dd2982\") " Nov 22 09:16:07 crc kubenswrapper[4789]: I1122 09:16:07.174790 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e49577b-5b45-455c-a135-395fa2dd2982-catalog-content\") pod \"6e49577b-5b45-455c-a135-395fa2dd2982\" (UID: \"6e49577b-5b45-455c-a135-395fa2dd2982\") " Nov 22 09:16:07 crc kubenswrapper[4789]: I1122 09:16:07.174899 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lm9rx\" (UniqueName: \"kubernetes.io/projected/6e49577b-5b45-455c-a135-395fa2dd2982-kube-api-access-lm9rx\") pod \"6e49577b-5b45-455c-a135-395fa2dd2982\" (UID: \"6e49577b-5b45-455c-a135-395fa2dd2982\") " Nov 22 09:16:07 crc kubenswrapper[4789]: I1122 09:16:07.175954 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6e49577b-5b45-455c-a135-395fa2dd2982-utilities" (OuterVolumeSpecName: "utilities") pod "6e49577b-5b45-455c-a135-395fa2dd2982" (UID: "6e49577b-5b45-455c-a135-395fa2dd2982"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 09:16:07 crc kubenswrapper[4789]: I1122 09:16:07.181112 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e49577b-5b45-455c-a135-395fa2dd2982-kube-api-access-lm9rx" (OuterVolumeSpecName: "kube-api-access-lm9rx") pod "6e49577b-5b45-455c-a135-395fa2dd2982" (UID: "6e49577b-5b45-455c-a135-395fa2dd2982"). InnerVolumeSpecName "kube-api-access-lm9rx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 09:16:07 crc kubenswrapper[4789]: I1122 09:16:07.276595 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lm9rx\" (UniqueName: \"kubernetes.io/projected/6e49577b-5b45-455c-a135-395fa2dd2982-kube-api-access-lm9rx\") on node \"crc\" DevicePath \"\"" Nov 22 09:16:07 crc kubenswrapper[4789]: I1122 09:16:07.276641 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e49577b-5b45-455c-a135-395fa2dd2982-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 09:16:07 crc kubenswrapper[4789]: I1122 09:16:07.280284 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6e49577b-5b45-455c-a135-395fa2dd2982-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6e49577b-5b45-455c-a135-395fa2dd2982" (UID: "6e49577b-5b45-455c-a135-395fa2dd2982"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 09:16:07 crc kubenswrapper[4789]: I1122 09:16:07.378803 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e49577b-5b45-455c-a135-395fa2dd2982-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 09:16:07 crc kubenswrapper[4789]: I1122 09:16:07.393502 4789 generic.go:334] "Generic (PLEG): container finished" podID="6e49577b-5b45-455c-a135-395fa2dd2982" containerID="9831bf52ba1ea2d010c0a8c04def331e9b9177aeb9a1aa99232d55f16041e4d0" exitCode=0 Nov 22 09:16:07 crc kubenswrapper[4789]: I1122 09:16:07.393548 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l8xgd" event={"ID":"6e49577b-5b45-455c-a135-395fa2dd2982","Type":"ContainerDied","Data":"9831bf52ba1ea2d010c0a8c04def331e9b9177aeb9a1aa99232d55f16041e4d0"} Nov 22 09:16:07 crc kubenswrapper[4789]: I1122 09:16:07.393562 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l8xgd" Nov 22 09:16:07 crc kubenswrapper[4789]: I1122 09:16:07.393578 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l8xgd" event={"ID":"6e49577b-5b45-455c-a135-395fa2dd2982","Type":"ContainerDied","Data":"4903aac84e03fdcdba4598c3f6c82a929971705b6771b43e192ba2400a7947c4"} Nov 22 09:16:07 crc kubenswrapper[4789]: I1122 09:16:07.393595 4789 scope.go:117] "RemoveContainer" containerID="9831bf52ba1ea2d010c0a8c04def331e9b9177aeb9a1aa99232d55f16041e4d0" Nov 22 09:16:07 crc kubenswrapper[4789]: I1122 09:16:07.425898 4789 scope.go:117] "RemoveContainer" containerID="8ba36933cc0c1c94bad05fccdce9b7d606b5458ac79883e9b4c54e949a74ebb0" Nov 22 09:16:07 crc kubenswrapper[4789]: I1122 09:16:07.430222 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-l8xgd"] Nov 22 09:16:07 crc kubenswrapper[4789]: I1122 09:16:07.440661 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-l8xgd"] Nov 22 09:16:07 crc kubenswrapper[4789]: I1122 09:16:07.452719 4789 scope.go:117] "RemoveContainer" containerID="e073791210d7347beb335e4ab171298ef6ecafad70d0bc17592c679359116b39" Nov 22 09:16:07 crc kubenswrapper[4789]: I1122 09:16:07.494380 4789 scope.go:117] "RemoveContainer" containerID="9831bf52ba1ea2d010c0a8c04def331e9b9177aeb9a1aa99232d55f16041e4d0" Nov 22 09:16:07 crc kubenswrapper[4789]: E1122 09:16:07.496333 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9831bf52ba1ea2d010c0a8c04def331e9b9177aeb9a1aa99232d55f16041e4d0\": container with ID starting with 9831bf52ba1ea2d010c0a8c04def331e9b9177aeb9a1aa99232d55f16041e4d0 not found: ID does not exist" containerID="9831bf52ba1ea2d010c0a8c04def331e9b9177aeb9a1aa99232d55f16041e4d0" Nov 22 09:16:07 crc kubenswrapper[4789]: I1122 09:16:07.496401 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9831bf52ba1ea2d010c0a8c04def331e9b9177aeb9a1aa99232d55f16041e4d0"} err="failed to get container status \"9831bf52ba1ea2d010c0a8c04def331e9b9177aeb9a1aa99232d55f16041e4d0\": rpc error: code = NotFound desc = could not find container \"9831bf52ba1ea2d010c0a8c04def331e9b9177aeb9a1aa99232d55f16041e4d0\": container with ID starting with 9831bf52ba1ea2d010c0a8c04def331e9b9177aeb9a1aa99232d55f16041e4d0 not found: ID does not exist" Nov 22 09:16:07 crc kubenswrapper[4789]: I1122 09:16:07.496435 4789 scope.go:117] "RemoveContainer" containerID="8ba36933cc0c1c94bad05fccdce9b7d606b5458ac79883e9b4c54e949a74ebb0" Nov 22 09:16:07 crc kubenswrapper[4789]: E1122 09:16:07.496788 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8ba36933cc0c1c94bad05fccdce9b7d606b5458ac79883e9b4c54e949a74ebb0\": container with ID starting with 8ba36933cc0c1c94bad05fccdce9b7d606b5458ac79883e9b4c54e949a74ebb0 not found: ID does not exist" containerID="8ba36933cc0c1c94bad05fccdce9b7d606b5458ac79883e9b4c54e949a74ebb0" Nov 22 09:16:07 crc kubenswrapper[4789]: I1122 09:16:07.496848 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8ba36933cc0c1c94bad05fccdce9b7d606b5458ac79883e9b4c54e949a74ebb0"} err="failed to get container status \"8ba36933cc0c1c94bad05fccdce9b7d606b5458ac79883e9b4c54e949a74ebb0\": rpc error: code = NotFound desc = could not find container \"8ba36933cc0c1c94bad05fccdce9b7d606b5458ac79883e9b4c54e949a74ebb0\": container with ID starting with 8ba36933cc0c1c94bad05fccdce9b7d606b5458ac79883e9b4c54e949a74ebb0 not found: ID does not exist" Nov 22 09:16:07 crc kubenswrapper[4789]: I1122 09:16:07.496870 4789 scope.go:117] "RemoveContainer" containerID="e073791210d7347beb335e4ab171298ef6ecafad70d0bc17592c679359116b39" Nov 22 09:16:07 crc kubenswrapper[4789]: E1122 09:16:07.498368 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e073791210d7347beb335e4ab171298ef6ecafad70d0bc17592c679359116b39\": container with ID starting with e073791210d7347beb335e4ab171298ef6ecafad70d0bc17592c679359116b39 not found: ID does not exist" containerID="e073791210d7347beb335e4ab171298ef6ecafad70d0bc17592c679359116b39" Nov 22 09:16:07 crc kubenswrapper[4789]: I1122 09:16:07.498403 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e073791210d7347beb335e4ab171298ef6ecafad70d0bc17592c679359116b39"} err="failed to get container status \"e073791210d7347beb335e4ab171298ef6ecafad70d0bc17592c679359116b39\": rpc error: code = NotFound desc = could not find container \"e073791210d7347beb335e4ab171298ef6ecafad70d0bc17592c679359116b39\": container with ID starting with e073791210d7347beb335e4ab171298ef6ecafad70d0bc17592c679359116b39 not found: ID does not exist" Nov 22 09:16:07 crc kubenswrapper[4789]: I1122 09:16:07.977605 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6e49577b-5b45-455c-a135-395fa2dd2982" path="/var/lib/kubelet/pods/6e49577b-5b45-455c-a135-395fa2dd2982/volumes" Nov 22 09:16:35 crc kubenswrapper[4789]: I1122 09:16:35.372080 4789 patch_prober.go:28] interesting pod/machine-config-daemon-pxvrm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 09:16:35 crc kubenswrapper[4789]: I1122 09:16:35.372684 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 09:17:05 crc kubenswrapper[4789]: I1122 09:17:05.372053 4789 patch_prober.go:28] interesting pod/machine-config-daemon-pxvrm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 09:17:05 crc kubenswrapper[4789]: I1122 09:17:05.372597 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 09:17:05 crc kubenswrapper[4789]: I1122 09:17:05.372643 4789 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" Nov 22 09:17:05 crc kubenswrapper[4789]: I1122 09:17:05.373377 4789 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ebe5a3882b79b00f381d091d16cb20463c15e848a3ab9e7c2941a9f043d441b1"} pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 22 09:17:05 crc kubenswrapper[4789]: I1122 09:17:05.373430 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" containerID="cri-o://ebe5a3882b79b00f381d091d16cb20463c15e848a3ab9e7c2941a9f043d441b1" gracePeriod=600 Nov 22 09:17:05 crc kubenswrapper[4789]: I1122 09:17:05.887200 4789 generic.go:334] "Generic (PLEG): container finished" podID="58aa8071-7ecc-4692-8789-2db04ec70647" containerID="ebe5a3882b79b00f381d091d16cb20463c15e848a3ab9e7c2941a9f043d441b1" exitCode=0 Nov 22 09:17:05 crc kubenswrapper[4789]: I1122 09:17:05.887343 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" event={"ID":"58aa8071-7ecc-4692-8789-2db04ec70647","Type":"ContainerDied","Data":"ebe5a3882b79b00f381d091d16cb20463c15e848a3ab9e7c2941a9f043d441b1"} Nov 22 09:17:05 crc kubenswrapper[4789]: I1122 09:17:05.887570 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" event={"ID":"58aa8071-7ecc-4692-8789-2db04ec70647","Type":"ContainerStarted","Data":"4240b6f705fa3ecd8a7faa7c66c58a040eaa92955c0d845a27bc191a2ab426e5"} Nov 22 09:17:05 crc kubenswrapper[4789]: I1122 09:17:05.887596 4789 scope.go:117] "RemoveContainer" containerID="e9395b7f22fc562bccbab647e93ed06aa21f27b91666ef88922354e2f66bcf54" Nov 22 09:17:43 crc kubenswrapper[4789]: I1122 09:17:43.531858 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-p7bj7"] Nov 22 09:17:43 crc kubenswrapper[4789]: E1122 09:17:43.532627 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e49577b-5b45-455c-a135-395fa2dd2982" containerName="extract-content" Nov 22 09:17:43 crc kubenswrapper[4789]: I1122 09:17:43.532639 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e49577b-5b45-455c-a135-395fa2dd2982" containerName="extract-content" Nov 22 09:17:43 crc kubenswrapper[4789]: E1122 09:17:43.532657 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e49577b-5b45-455c-a135-395fa2dd2982" containerName="registry-server" Nov 22 09:17:43 crc kubenswrapper[4789]: I1122 09:17:43.532663 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e49577b-5b45-455c-a135-395fa2dd2982" containerName="registry-server" Nov 22 09:17:43 crc kubenswrapper[4789]: E1122 09:17:43.532693 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e49577b-5b45-455c-a135-395fa2dd2982" containerName="extract-utilities" Nov 22 09:17:43 crc kubenswrapper[4789]: I1122 09:17:43.532700 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e49577b-5b45-455c-a135-395fa2dd2982" containerName="extract-utilities" Nov 22 09:17:43 crc kubenswrapper[4789]: I1122 09:17:43.532886 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e49577b-5b45-455c-a135-395fa2dd2982" containerName="registry-server" Nov 22 09:17:43 crc kubenswrapper[4789]: I1122 09:17:43.538081 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p7bj7" Nov 22 09:17:43 crc kubenswrapper[4789]: I1122 09:17:43.567076 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-p7bj7"] Nov 22 09:17:43 crc kubenswrapper[4789]: I1122 09:17:43.680998 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-md22s\" (UniqueName: \"kubernetes.io/projected/5b8408ac-6c74-415a-8ad7-6c75915440c2-kube-api-access-md22s\") pod \"certified-operators-p7bj7\" (UID: \"5b8408ac-6c74-415a-8ad7-6c75915440c2\") " pod="openshift-marketplace/certified-operators-p7bj7" Nov 22 09:17:43 crc kubenswrapper[4789]: I1122 09:17:43.681063 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b8408ac-6c74-415a-8ad7-6c75915440c2-utilities\") pod \"certified-operators-p7bj7\" (UID: \"5b8408ac-6c74-415a-8ad7-6c75915440c2\") " pod="openshift-marketplace/certified-operators-p7bj7" Nov 22 09:17:43 crc kubenswrapper[4789]: I1122 09:17:43.681138 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b8408ac-6c74-415a-8ad7-6c75915440c2-catalog-content\") pod \"certified-operators-p7bj7\" (UID: \"5b8408ac-6c74-415a-8ad7-6c75915440c2\") " pod="openshift-marketplace/certified-operators-p7bj7" Nov 22 09:17:43 crc kubenswrapper[4789]: I1122 09:17:43.782763 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-md22s\" (UniqueName: \"kubernetes.io/projected/5b8408ac-6c74-415a-8ad7-6c75915440c2-kube-api-access-md22s\") pod \"certified-operators-p7bj7\" (UID: \"5b8408ac-6c74-415a-8ad7-6c75915440c2\") " pod="openshift-marketplace/certified-operators-p7bj7" Nov 22 09:17:43 crc kubenswrapper[4789]: I1122 09:17:43.782815 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b8408ac-6c74-415a-8ad7-6c75915440c2-utilities\") pod \"certified-operators-p7bj7\" (UID: \"5b8408ac-6c74-415a-8ad7-6c75915440c2\") " pod="openshift-marketplace/certified-operators-p7bj7" Nov 22 09:17:43 crc kubenswrapper[4789]: I1122 09:17:43.782842 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b8408ac-6c74-415a-8ad7-6c75915440c2-catalog-content\") pod \"certified-operators-p7bj7\" (UID: \"5b8408ac-6c74-415a-8ad7-6c75915440c2\") " pod="openshift-marketplace/certified-operators-p7bj7" Nov 22 09:17:43 crc kubenswrapper[4789]: I1122 09:17:43.783247 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b8408ac-6c74-415a-8ad7-6c75915440c2-catalog-content\") pod \"certified-operators-p7bj7\" (UID: \"5b8408ac-6c74-415a-8ad7-6c75915440c2\") " pod="openshift-marketplace/certified-operators-p7bj7" Nov 22 09:17:43 crc kubenswrapper[4789]: I1122 09:17:43.783337 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b8408ac-6c74-415a-8ad7-6c75915440c2-utilities\") pod \"certified-operators-p7bj7\" (UID: \"5b8408ac-6c74-415a-8ad7-6c75915440c2\") " pod="openshift-marketplace/certified-operators-p7bj7" Nov 22 09:17:43 crc kubenswrapper[4789]: I1122 09:17:43.805156 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-md22s\" (UniqueName: \"kubernetes.io/projected/5b8408ac-6c74-415a-8ad7-6c75915440c2-kube-api-access-md22s\") pod \"certified-operators-p7bj7\" (UID: \"5b8408ac-6c74-415a-8ad7-6c75915440c2\") " pod="openshift-marketplace/certified-operators-p7bj7" Nov 22 09:17:43 crc kubenswrapper[4789]: I1122 09:17:43.872296 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p7bj7" Nov 22 09:17:44 crc kubenswrapper[4789]: I1122 09:17:44.375733 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-p7bj7"] Nov 22 09:17:45 crc kubenswrapper[4789]: I1122 09:17:45.218274 4789 generic.go:334] "Generic (PLEG): container finished" podID="5b8408ac-6c74-415a-8ad7-6c75915440c2" containerID="93e8c5d3b2eb5d240a807372e615866c383419c658957bb5fcefad00f30a9e6a" exitCode=0 Nov 22 09:17:45 crc kubenswrapper[4789]: I1122 09:17:45.218557 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p7bj7" event={"ID":"5b8408ac-6c74-415a-8ad7-6c75915440c2","Type":"ContainerDied","Data":"93e8c5d3b2eb5d240a807372e615866c383419c658957bb5fcefad00f30a9e6a"} Nov 22 09:17:45 crc kubenswrapper[4789]: I1122 09:17:45.218590 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p7bj7" event={"ID":"5b8408ac-6c74-415a-8ad7-6c75915440c2","Type":"ContainerStarted","Data":"fe1a35b6bcc1a53f3219f58de0a79804d0dbe8fe933c0d3eab19c91eeafca341"} Nov 22 09:17:46 crc kubenswrapper[4789]: I1122 09:17:46.231294 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p7bj7" event={"ID":"5b8408ac-6c74-415a-8ad7-6c75915440c2","Type":"ContainerStarted","Data":"0a038f1e3662a65f789fca57501866c94a778f9bb56b7a6367d3b0a0d0cec836"} Nov 22 09:17:48 crc kubenswrapper[4789]: I1122 09:17:48.251394 4789 generic.go:334] "Generic (PLEG): container finished" podID="5b8408ac-6c74-415a-8ad7-6c75915440c2" containerID="0a038f1e3662a65f789fca57501866c94a778f9bb56b7a6367d3b0a0d0cec836" exitCode=0 Nov 22 09:17:48 crc kubenswrapper[4789]: I1122 09:17:48.251845 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p7bj7" event={"ID":"5b8408ac-6c74-415a-8ad7-6c75915440c2","Type":"ContainerDied","Data":"0a038f1e3662a65f789fca57501866c94a778f9bb56b7a6367d3b0a0d0cec836"} Nov 22 09:17:49 crc kubenswrapper[4789]: I1122 09:17:49.264439 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p7bj7" event={"ID":"5b8408ac-6c74-415a-8ad7-6c75915440c2","Type":"ContainerStarted","Data":"9ea7ca14f5ef71bcf76d0cb2d54d58e240933bbfc116f5bc0375480d732519a2"} Nov 22 09:17:49 crc kubenswrapper[4789]: I1122 09:17:49.291382 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-p7bj7" podStartSLOduration=2.8329847409999998 podStartE2EDuration="6.291359753s" podCreationTimestamp="2025-11-22 09:17:43 +0000 UTC" firstStartedPulling="2025-11-22 09:17:45.220737635 +0000 UTC m=+4879.455137908" lastFinishedPulling="2025-11-22 09:17:48.679112647 +0000 UTC m=+4882.913512920" observedRunningTime="2025-11-22 09:17:49.282391095 +0000 UTC m=+4883.516791388" watchObservedRunningTime="2025-11-22 09:17:49.291359753 +0000 UTC m=+4883.525760026" Nov 22 09:17:53 crc kubenswrapper[4789]: I1122 09:17:53.873223 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-p7bj7" Nov 22 09:17:53 crc kubenswrapper[4789]: I1122 09:17:53.873817 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-p7bj7" Nov 22 09:17:53 crc kubenswrapper[4789]: I1122 09:17:53.924994 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-p7bj7" Nov 22 09:17:54 crc kubenswrapper[4789]: I1122 09:17:54.351607 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-p7bj7" Nov 22 09:17:54 crc kubenswrapper[4789]: I1122 09:17:54.405304 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-p7bj7"] Nov 22 09:17:56 crc kubenswrapper[4789]: I1122 09:17:56.322350 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-p7bj7" podUID="5b8408ac-6c74-415a-8ad7-6c75915440c2" containerName="registry-server" containerID="cri-o://9ea7ca14f5ef71bcf76d0cb2d54d58e240933bbfc116f5bc0375480d732519a2" gracePeriod=2 Nov 22 09:17:57 crc kubenswrapper[4789]: I1122 09:17:57.062028 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p7bj7" Nov 22 09:17:57 crc kubenswrapper[4789]: I1122 09:17:57.148947 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b8408ac-6c74-415a-8ad7-6c75915440c2-utilities\") pod \"5b8408ac-6c74-415a-8ad7-6c75915440c2\" (UID: \"5b8408ac-6c74-415a-8ad7-6c75915440c2\") " Nov 22 09:17:57 crc kubenswrapper[4789]: I1122 09:17:57.149094 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b8408ac-6c74-415a-8ad7-6c75915440c2-catalog-content\") pod \"5b8408ac-6c74-415a-8ad7-6c75915440c2\" (UID: \"5b8408ac-6c74-415a-8ad7-6c75915440c2\") " Nov 22 09:17:57 crc kubenswrapper[4789]: I1122 09:17:57.149337 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-md22s\" (UniqueName: \"kubernetes.io/projected/5b8408ac-6c74-415a-8ad7-6c75915440c2-kube-api-access-md22s\") pod \"5b8408ac-6c74-415a-8ad7-6c75915440c2\" (UID: \"5b8408ac-6c74-415a-8ad7-6c75915440c2\") " Nov 22 09:17:57 crc kubenswrapper[4789]: I1122 09:17:57.150179 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5b8408ac-6c74-415a-8ad7-6c75915440c2-utilities" (OuterVolumeSpecName: "utilities") pod "5b8408ac-6c74-415a-8ad7-6c75915440c2" (UID: "5b8408ac-6c74-415a-8ad7-6c75915440c2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 09:17:57 crc kubenswrapper[4789]: I1122 09:17:57.155309 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b8408ac-6c74-415a-8ad7-6c75915440c2-kube-api-access-md22s" (OuterVolumeSpecName: "kube-api-access-md22s") pod "5b8408ac-6c74-415a-8ad7-6c75915440c2" (UID: "5b8408ac-6c74-415a-8ad7-6c75915440c2"). InnerVolumeSpecName "kube-api-access-md22s". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 09:17:57 crc kubenswrapper[4789]: I1122 09:17:57.204382 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5b8408ac-6c74-415a-8ad7-6c75915440c2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5b8408ac-6c74-415a-8ad7-6c75915440c2" (UID: "5b8408ac-6c74-415a-8ad7-6c75915440c2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 09:17:57 crc kubenswrapper[4789]: I1122 09:17:57.252051 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-md22s\" (UniqueName: \"kubernetes.io/projected/5b8408ac-6c74-415a-8ad7-6c75915440c2-kube-api-access-md22s\") on node \"crc\" DevicePath \"\"" Nov 22 09:17:57 crc kubenswrapper[4789]: I1122 09:17:57.252095 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b8408ac-6c74-415a-8ad7-6c75915440c2-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 09:17:57 crc kubenswrapper[4789]: I1122 09:17:57.252114 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b8408ac-6c74-415a-8ad7-6c75915440c2-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 09:17:57 crc kubenswrapper[4789]: I1122 09:17:57.336608 4789 generic.go:334] "Generic (PLEG): container finished" podID="5b8408ac-6c74-415a-8ad7-6c75915440c2" containerID="9ea7ca14f5ef71bcf76d0cb2d54d58e240933bbfc116f5bc0375480d732519a2" exitCode=0 Nov 22 09:17:57 crc kubenswrapper[4789]: I1122 09:17:57.336675 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p7bj7" event={"ID":"5b8408ac-6c74-415a-8ad7-6c75915440c2","Type":"ContainerDied","Data":"9ea7ca14f5ef71bcf76d0cb2d54d58e240933bbfc116f5bc0375480d732519a2"} Nov 22 09:17:57 crc kubenswrapper[4789]: I1122 09:17:57.336716 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p7bj7" event={"ID":"5b8408ac-6c74-415a-8ad7-6c75915440c2","Type":"ContainerDied","Data":"fe1a35b6bcc1a53f3219f58de0a79804d0dbe8fe933c0d3eab19c91eeafca341"} Nov 22 09:17:57 crc kubenswrapper[4789]: I1122 09:17:57.336739 4789 scope.go:117] "RemoveContainer" containerID="9ea7ca14f5ef71bcf76d0cb2d54d58e240933bbfc116f5bc0375480d732519a2" Nov 22 09:17:57 crc kubenswrapper[4789]: I1122 09:17:57.336879 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p7bj7" Nov 22 09:17:57 crc kubenswrapper[4789]: I1122 09:17:57.383850 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-p7bj7"] Nov 22 09:17:57 crc kubenswrapper[4789]: I1122 09:17:57.388218 4789 scope.go:117] "RemoveContainer" containerID="0a038f1e3662a65f789fca57501866c94a778f9bb56b7a6367d3b0a0d0cec836" Nov 22 09:17:57 crc kubenswrapper[4789]: I1122 09:17:57.393592 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-p7bj7"] Nov 22 09:17:57 crc kubenswrapper[4789]: I1122 09:17:57.408930 4789 scope.go:117] "RemoveContainer" containerID="93e8c5d3b2eb5d240a807372e615866c383419c658957bb5fcefad00f30a9e6a" Nov 22 09:17:57 crc kubenswrapper[4789]: I1122 09:17:57.449878 4789 scope.go:117] "RemoveContainer" containerID="9ea7ca14f5ef71bcf76d0cb2d54d58e240933bbfc116f5bc0375480d732519a2" Nov 22 09:17:57 crc kubenswrapper[4789]: E1122 09:17:57.450412 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9ea7ca14f5ef71bcf76d0cb2d54d58e240933bbfc116f5bc0375480d732519a2\": container with ID starting with 9ea7ca14f5ef71bcf76d0cb2d54d58e240933bbfc116f5bc0375480d732519a2 not found: ID does not exist" containerID="9ea7ca14f5ef71bcf76d0cb2d54d58e240933bbfc116f5bc0375480d732519a2" Nov 22 09:17:57 crc kubenswrapper[4789]: I1122 09:17:57.450490 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ea7ca14f5ef71bcf76d0cb2d54d58e240933bbfc116f5bc0375480d732519a2"} err="failed to get container status \"9ea7ca14f5ef71bcf76d0cb2d54d58e240933bbfc116f5bc0375480d732519a2\": rpc error: code = NotFound desc = could not find container \"9ea7ca14f5ef71bcf76d0cb2d54d58e240933bbfc116f5bc0375480d732519a2\": container with ID starting with 9ea7ca14f5ef71bcf76d0cb2d54d58e240933bbfc116f5bc0375480d732519a2 not found: ID does not exist" Nov 22 09:17:57 crc kubenswrapper[4789]: I1122 09:17:57.450522 4789 scope.go:117] "RemoveContainer" containerID="0a038f1e3662a65f789fca57501866c94a778f9bb56b7a6367d3b0a0d0cec836" Nov 22 09:17:57 crc kubenswrapper[4789]: E1122 09:17:57.450955 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0a038f1e3662a65f789fca57501866c94a778f9bb56b7a6367d3b0a0d0cec836\": container with ID starting with 0a038f1e3662a65f789fca57501866c94a778f9bb56b7a6367d3b0a0d0cec836 not found: ID does not exist" containerID="0a038f1e3662a65f789fca57501866c94a778f9bb56b7a6367d3b0a0d0cec836" Nov 22 09:17:57 crc kubenswrapper[4789]: I1122 09:17:57.451111 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a038f1e3662a65f789fca57501866c94a778f9bb56b7a6367d3b0a0d0cec836"} err="failed to get container status \"0a038f1e3662a65f789fca57501866c94a778f9bb56b7a6367d3b0a0d0cec836\": rpc error: code = NotFound desc = could not find container \"0a038f1e3662a65f789fca57501866c94a778f9bb56b7a6367d3b0a0d0cec836\": container with ID starting with 0a038f1e3662a65f789fca57501866c94a778f9bb56b7a6367d3b0a0d0cec836 not found: ID does not exist" Nov 22 09:17:57 crc kubenswrapper[4789]: I1122 09:17:57.451139 4789 scope.go:117] "RemoveContainer" containerID="93e8c5d3b2eb5d240a807372e615866c383419c658957bb5fcefad00f30a9e6a" Nov 22 09:17:57 crc kubenswrapper[4789]: E1122 09:17:57.451429 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"93e8c5d3b2eb5d240a807372e615866c383419c658957bb5fcefad00f30a9e6a\": container with ID starting with 93e8c5d3b2eb5d240a807372e615866c383419c658957bb5fcefad00f30a9e6a not found: ID does not exist" containerID="93e8c5d3b2eb5d240a807372e615866c383419c658957bb5fcefad00f30a9e6a" Nov 22 09:17:57 crc kubenswrapper[4789]: I1122 09:17:57.451461 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93e8c5d3b2eb5d240a807372e615866c383419c658957bb5fcefad00f30a9e6a"} err="failed to get container status \"93e8c5d3b2eb5d240a807372e615866c383419c658957bb5fcefad00f30a9e6a\": rpc error: code = NotFound desc = could not find container \"93e8c5d3b2eb5d240a807372e615866c383419c658957bb5fcefad00f30a9e6a\": container with ID starting with 93e8c5d3b2eb5d240a807372e615866c383419c658957bb5fcefad00f30a9e6a not found: ID does not exist" Nov 22 09:17:57 crc kubenswrapper[4789]: I1122 09:17:57.976384 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b8408ac-6c74-415a-8ad7-6c75915440c2" path="/var/lib/kubelet/pods/5b8408ac-6c74-415a-8ad7-6c75915440c2/volumes" Nov 22 09:19:05 crc kubenswrapper[4789]: I1122 09:19:05.372704 4789 patch_prober.go:28] interesting pod/machine-config-daemon-pxvrm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 09:19:05 crc kubenswrapper[4789]: I1122 09:19:05.373274 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 09:19:35 crc kubenswrapper[4789]: I1122 09:19:35.372133 4789 patch_prober.go:28] interesting pod/machine-config-daemon-pxvrm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 09:19:35 crc kubenswrapper[4789]: I1122 09:19:35.372819 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 09:20:05 crc kubenswrapper[4789]: I1122 09:20:05.371649 4789 patch_prober.go:28] interesting pod/machine-config-daemon-pxvrm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 09:20:05 crc kubenswrapper[4789]: I1122 09:20:05.372104 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 09:20:05 crc kubenswrapper[4789]: I1122 09:20:05.372147 4789 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" Nov 22 09:20:05 crc kubenswrapper[4789]: I1122 09:20:05.372855 4789 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4240b6f705fa3ecd8a7faa7c66c58a040eaa92955c0d845a27bc191a2ab426e5"} pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 22 09:20:05 crc kubenswrapper[4789]: I1122 09:20:05.372908 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" containerID="cri-o://4240b6f705fa3ecd8a7faa7c66c58a040eaa92955c0d845a27bc191a2ab426e5" gracePeriod=600 Nov 22 09:20:05 crc kubenswrapper[4789]: E1122 09:20:05.510439 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:20:05 crc kubenswrapper[4789]: E1122 09:20:05.587140 4789 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod58aa8071_7ecc_4692_8789_2db04ec70647.slice/crio-4240b6f705fa3ecd8a7faa7c66c58a040eaa92955c0d845a27bc191a2ab426e5.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod58aa8071_7ecc_4692_8789_2db04ec70647.slice/crio-conmon-4240b6f705fa3ecd8a7faa7c66c58a040eaa92955c0d845a27bc191a2ab426e5.scope\": RecentStats: unable to find data in memory cache]" Nov 22 09:20:06 crc kubenswrapper[4789]: I1122 09:20:06.443138 4789 generic.go:334] "Generic (PLEG): container finished" podID="58aa8071-7ecc-4692-8789-2db04ec70647" containerID="4240b6f705fa3ecd8a7faa7c66c58a040eaa92955c0d845a27bc191a2ab426e5" exitCode=0 Nov 22 09:20:06 crc kubenswrapper[4789]: I1122 09:20:06.443190 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" event={"ID":"58aa8071-7ecc-4692-8789-2db04ec70647","Type":"ContainerDied","Data":"4240b6f705fa3ecd8a7faa7c66c58a040eaa92955c0d845a27bc191a2ab426e5"} Nov 22 09:20:06 crc kubenswrapper[4789]: I1122 09:20:06.443236 4789 scope.go:117] "RemoveContainer" containerID="ebe5a3882b79b00f381d091d16cb20463c15e848a3ab9e7c2941a9f043d441b1" Nov 22 09:20:06 crc kubenswrapper[4789]: I1122 09:20:06.443910 4789 scope.go:117] "RemoveContainer" containerID="4240b6f705fa3ecd8a7faa7c66c58a040eaa92955c0d845a27bc191a2ab426e5" Nov 22 09:20:06 crc kubenswrapper[4789]: E1122 09:20:06.444275 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:20:20 crc kubenswrapper[4789]: I1122 09:20:20.965485 4789 scope.go:117] "RemoveContainer" containerID="4240b6f705fa3ecd8a7faa7c66c58a040eaa92955c0d845a27bc191a2ab426e5" Nov 22 09:20:20 crc kubenswrapper[4789]: E1122 09:20:20.966345 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:20:34 crc kubenswrapper[4789]: I1122 09:20:34.965445 4789 scope.go:117] "RemoveContainer" containerID="4240b6f705fa3ecd8a7faa7c66c58a040eaa92955c0d845a27bc191a2ab426e5" Nov 22 09:20:34 crc kubenswrapper[4789]: E1122 09:20:34.966350 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:20:46 crc kubenswrapper[4789]: I1122 09:20:46.965056 4789 scope.go:117] "RemoveContainer" containerID="4240b6f705fa3ecd8a7faa7c66c58a040eaa92955c0d845a27bc191a2ab426e5" Nov 22 09:20:46 crc kubenswrapper[4789]: E1122 09:20:46.965834 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:21:00 crc kubenswrapper[4789]: I1122 09:21:00.964927 4789 scope.go:117] "RemoveContainer" containerID="4240b6f705fa3ecd8a7faa7c66c58a040eaa92955c0d845a27bc191a2ab426e5" Nov 22 09:21:00 crc kubenswrapper[4789]: E1122 09:21:00.965644 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:21:12 crc kubenswrapper[4789]: I1122 09:21:12.965512 4789 scope.go:117] "RemoveContainer" containerID="4240b6f705fa3ecd8a7faa7c66c58a040eaa92955c0d845a27bc191a2ab426e5" Nov 22 09:21:12 crc kubenswrapper[4789]: E1122 09:21:12.966462 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:21:26 crc kubenswrapper[4789]: I1122 09:21:26.965442 4789 scope.go:117] "RemoveContainer" containerID="4240b6f705fa3ecd8a7faa7c66c58a040eaa92955c0d845a27bc191a2ab426e5" Nov 22 09:21:26 crc kubenswrapper[4789]: E1122 09:21:26.966160 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:21:38 crc kubenswrapper[4789]: I1122 09:21:38.966473 4789 scope.go:117] "RemoveContainer" containerID="4240b6f705fa3ecd8a7faa7c66c58a040eaa92955c0d845a27bc191a2ab426e5" Nov 22 09:21:38 crc kubenswrapper[4789]: E1122 09:21:38.967223 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:21:52 crc kubenswrapper[4789]: I1122 09:21:52.965300 4789 scope.go:117] "RemoveContainer" containerID="4240b6f705fa3ecd8a7faa7c66c58a040eaa92955c0d845a27bc191a2ab426e5" Nov 22 09:21:52 crc kubenswrapper[4789]: E1122 09:21:52.966018 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:22:06 crc kubenswrapper[4789]: I1122 09:22:06.965746 4789 scope.go:117] "RemoveContainer" containerID="4240b6f705fa3ecd8a7faa7c66c58a040eaa92955c0d845a27bc191a2ab426e5" Nov 22 09:22:06 crc kubenswrapper[4789]: E1122 09:22:06.966711 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:22:19 crc kubenswrapper[4789]: I1122 09:22:19.965224 4789 scope.go:117] "RemoveContainer" containerID="4240b6f705fa3ecd8a7faa7c66c58a040eaa92955c0d845a27bc191a2ab426e5" Nov 22 09:22:19 crc kubenswrapper[4789]: E1122 09:22:19.966038 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:22:32 crc kubenswrapper[4789]: I1122 09:22:32.965354 4789 scope.go:117] "RemoveContainer" containerID="4240b6f705fa3ecd8a7faa7c66c58a040eaa92955c0d845a27bc191a2ab426e5" Nov 22 09:22:32 crc kubenswrapper[4789]: E1122 09:22:32.966112 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:22:44 crc kubenswrapper[4789]: I1122 09:22:44.965740 4789 scope.go:117] "RemoveContainer" containerID="4240b6f705fa3ecd8a7faa7c66c58a040eaa92955c0d845a27bc191a2ab426e5" Nov 22 09:22:44 crc kubenswrapper[4789]: E1122 09:22:44.966742 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:22:57 crc kubenswrapper[4789]: I1122 09:22:57.975716 4789 scope.go:117] "RemoveContainer" containerID="4240b6f705fa3ecd8a7faa7c66c58a040eaa92955c0d845a27bc191a2ab426e5" Nov 22 09:22:57 crc kubenswrapper[4789]: E1122 09:22:57.976398 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:23:12 crc kubenswrapper[4789]: I1122 09:23:12.964917 4789 scope.go:117] "RemoveContainer" containerID="4240b6f705fa3ecd8a7faa7c66c58a040eaa92955c0d845a27bc191a2ab426e5" Nov 22 09:23:12 crc kubenswrapper[4789]: E1122 09:23:12.965699 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:23:20 crc kubenswrapper[4789]: I1122 09:23:20.372053 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-sthhp"] Nov 22 09:23:20 crc kubenswrapper[4789]: E1122 09:23:20.374510 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b8408ac-6c74-415a-8ad7-6c75915440c2" containerName="registry-server" Nov 22 09:23:20 crc kubenswrapper[4789]: I1122 09:23:20.374634 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b8408ac-6c74-415a-8ad7-6c75915440c2" containerName="registry-server" Nov 22 09:23:20 crc kubenswrapper[4789]: E1122 09:23:20.374728 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b8408ac-6c74-415a-8ad7-6c75915440c2" containerName="extract-utilities" Nov 22 09:23:20 crc kubenswrapper[4789]: I1122 09:23:20.374848 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b8408ac-6c74-415a-8ad7-6c75915440c2" containerName="extract-utilities" Nov 22 09:23:20 crc kubenswrapper[4789]: E1122 09:23:20.374990 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b8408ac-6c74-415a-8ad7-6c75915440c2" containerName="extract-content" Nov 22 09:23:20 crc kubenswrapper[4789]: I1122 09:23:20.375077 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b8408ac-6c74-415a-8ad7-6c75915440c2" containerName="extract-content" Nov 22 09:23:20 crc kubenswrapper[4789]: I1122 09:23:20.375524 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b8408ac-6c74-415a-8ad7-6c75915440c2" containerName="registry-server" Nov 22 09:23:20 crc kubenswrapper[4789]: I1122 09:23:20.377258 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sthhp" Nov 22 09:23:20 crc kubenswrapper[4789]: I1122 09:23:20.385691 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-sthhp"] Nov 22 09:23:20 crc kubenswrapper[4789]: I1122 09:23:20.544554 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e1b2757-7f3d-4c68-93c5-f5d4daf34201-catalog-content\") pod \"redhat-marketplace-sthhp\" (UID: \"5e1b2757-7f3d-4c68-93c5-f5d4daf34201\") " pod="openshift-marketplace/redhat-marketplace-sthhp" Nov 22 09:23:20 crc kubenswrapper[4789]: I1122 09:23:20.544707 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e1b2757-7f3d-4c68-93c5-f5d4daf34201-utilities\") pod \"redhat-marketplace-sthhp\" (UID: \"5e1b2757-7f3d-4c68-93c5-f5d4daf34201\") " pod="openshift-marketplace/redhat-marketplace-sthhp" Nov 22 09:23:20 crc kubenswrapper[4789]: I1122 09:23:20.544734 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rs5vf\" (UniqueName: \"kubernetes.io/projected/5e1b2757-7f3d-4c68-93c5-f5d4daf34201-kube-api-access-rs5vf\") pod \"redhat-marketplace-sthhp\" (UID: \"5e1b2757-7f3d-4c68-93c5-f5d4daf34201\") " pod="openshift-marketplace/redhat-marketplace-sthhp" Nov 22 09:23:20 crc kubenswrapper[4789]: I1122 09:23:20.646116 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e1b2757-7f3d-4c68-93c5-f5d4daf34201-catalog-content\") pod \"redhat-marketplace-sthhp\" (UID: \"5e1b2757-7f3d-4c68-93c5-f5d4daf34201\") " pod="openshift-marketplace/redhat-marketplace-sthhp" Nov 22 09:23:20 crc kubenswrapper[4789]: I1122 09:23:20.646483 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rs5vf\" (UniqueName: \"kubernetes.io/projected/5e1b2757-7f3d-4c68-93c5-f5d4daf34201-kube-api-access-rs5vf\") pod \"redhat-marketplace-sthhp\" (UID: \"5e1b2757-7f3d-4c68-93c5-f5d4daf34201\") " pod="openshift-marketplace/redhat-marketplace-sthhp" Nov 22 09:23:20 crc kubenswrapper[4789]: I1122 09:23:20.646603 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e1b2757-7f3d-4c68-93c5-f5d4daf34201-utilities\") pod \"redhat-marketplace-sthhp\" (UID: \"5e1b2757-7f3d-4c68-93c5-f5d4daf34201\") " pod="openshift-marketplace/redhat-marketplace-sthhp" Nov 22 09:23:20 crc kubenswrapper[4789]: I1122 09:23:20.646744 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e1b2757-7f3d-4c68-93c5-f5d4daf34201-catalog-content\") pod \"redhat-marketplace-sthhp\" (UID: \"5e1b2757-7f3d-4c68-93c5-f5d4daf34201\") " pod="openshift-marketplace/redhat-marketplace-sthhp" Nov 22 09:23:20 crc kubenswrapper[4789]: I1122 09:23:20.647178 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e1b2757-7f3d-4c68-93c5-f5d4daf34201-utilities\") pod \"redhat-marketplace-sthhp\" (UID: \"5e1b2757-7f3d-4c68-93c5-f5d4daf34201\") " pod="openshift-marketplace/redhat-marketplace-sthhp" Nov 22 09:23:20 crc kubenswrapper[4789]: I1122 09:23:20.665396 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rs5vf\" (UniqueName: \"kubernetes.io/projected/5e1b2757-7f3d-4c68-93c5-f5d4daf34201-kube-api-access-rs5vf\") pod \"redhat-marketplace-sthhp\" (UID: \"5e1b2757-7f3d-4c68-93c5-f5d4daf34201\") " pod="openshift-marketplace/redhat-marketplace-sthhp" Nov 22 09:23:20 crc kubenswrapper[4789]: I1122 09:23:20.699266 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sthhp" Nov 22 09:23:21 crc kubenswrapper[4789]: I1122 09:23:21.143114 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-sthhp"] Nov 22 09:23:22 crc kubenswrapper[4789]: I1122 09:23:22.084258 4789 generic.go:334] "Generic (PLEG): container finished" podID="5e1b2757-7f3d-4c68-93c5-f5d4daf34201" containerID="e3455958a7427b2fafda0369092bbbcfd6294543c02d66b1e333be3bab0ec36f" exitCode=0 Nov 22 09:23:22 crc kubenswrapper[4789]: I1122 09:23:22.084313 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sthhp" event={"ID":"5e1b2757-7f3d-4c68-93c5-f5d4daf34201","Type":"ContainerDied","Data":"e3455958a7427b2fafda0369092bbbcfd6294543c02d66b1e333be3bab0ec36f"} Nov 22 09:23:22 crc kubenswrapper[4789]: I1122 09:23:22.084771 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sthhp" event={"ID":"5e1b2757-7f3d-4c68-93c5-f5d4daf34201","Type":"ContainerStarted","Data":"f55f4f8b29361a323dcecaced040ea93b52f6be93c78a33f0fed73739c8b67d0"} Nov 22 09:23:22 crc kubenswrapper[4789]: I1122 09:23:22.088243 4789 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 22 09:23:24 crc kubenswrapper[4789]: I1122 09:23:24.105915 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sthhp" event={"ID":"5e1b2757-7f3d-4c68-93c5-f5d4daf34201","Type":"ContainerStarted","Data":"f8c753b706c16e1719bef41d5b556c945d7cfc8cb28132deb0b1362fa0bf0745"} Nov 22 09:23:25 crc kubenswrapper[4789]: I1122 09:23:25.965396 4789 scope.go:117] "RemoveContainer" containerID="4240b6f705fa3ecd8a7faa7c66c58a040eaa92955c0d845a27bc191a2ab426e5" Nov 22 09:23:25 crc kubenswrapper[4789]: E1122 09:23:25.966155 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:23:26 crc kubenswrapper[4789]: I1122 09:23:26.125368 4789 generic.go:334] "Generic (PLEG): container finished" podID="5e1b2757-7f3d-4c68-93c5-f5d4daf34201" containerID="f8c753b706c16e1719bef41d5b556c945d7cfc8cb28132deb0b1362fa0bf0745" exitCode=0 Nov 22 09:23:26 crc kubenswrapper[4789]: I1122 09:23:26.125439 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sthhp" event={"ID":"5e1b2757-7f3d-4c68-93c5-f5d4daf34201","Type":"ContainerDied","Data":"f8c753b706c16e1719bef41d5b556c945d7cfc8cb28132deb0b1362fa0bf0745"} Nov 22 09:23:27 crc kubenswrapper[4789]: I1122 09:23:27.137391 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sthhp" event={"ID":"5e1b2757-7f3d-4c68-93c5-f5d4daf34201","Type":"ContainerStarted","Data":"98f45bef82901a22528aa1da5a57b09a37e377dc533da5c513ce9faac4b958e2"} Nov 22 09:23:27 crc kubenswrapper[4789]: I1122 09:23:27.162570 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-sthhp" podStartSLOduration=2.666079161 podStartE2EDuration="7.162541121s" podCreationTimestamp="2025-11-22 09:23:20 +0000 UTC" firstStartedPulling="2025-11-22 09:23:22.088037674 +0000 UTC m=+5216.322437947" lastFinishedPulling="2025-11-22 09:23:26.584499634 +0000 UTC m=+5220.818899907" observedRunningTime="2025-11-22 09:23:27.154776585 +0000 UTC m=+5221.389176868" watchObservedRunningTime="2025-11-22 09:23:27.162541121 +0000 UTC m=+5221.396941414" Nov 22 09:23:30 crc kubenswrapper[4789]: I1122 09:23:30.700215 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-sthhp" Nov 22 09:23:30 crc kubenswrapper[4789]: I1122 09:23:30.700660 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-sthhp" Nov 22 09:23:30 crc kubenswrapper[4789]: I1122 09:23:30.750016 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-sthhp" Nov 22 09:23:31 crc kubenswrapper[4789]: I1122 09:23:31.221842 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-sthhp" Nov 22 09:23:31 crc kubenswrapper[4789]: I1122 09:23:31.271984 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-sthhp"] Nov 22 09:23:33 crc kubenswrapper[4789]: I1122 09:23:33.190626 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-sthhp" podUID="5e1b2757-7f3d-4c68-93c5-f5d4daf34201" containerName="registry-server" containerID="cri-o://98f45bef82901a22528aa1da5a57b09a37e377dc533da5c513ce9faac4b958e2" gracePeriod=2 Nov 22 09:23:33 crc kubenswrapper[4789]: I1122 09:23:33.802612 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sthhp" Nov 22 09:23:33 crc kubenswrapper[4789]: I1122 09:23:33.901126 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e1b2757-7f3d-4c68-93c5-f5d4daf34201-utilities\") pod \"5e1b2757-7f3d-4c68-93c5-f5d4daf34201\" (UID: \"5e1b2757-7f3d-4c68-93c5-f5d4daf34201\") " Nov 22 09:23:33 crc kubenswrapper[4789]: I1122 09:23:33.901423 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e1b2757-7f3d-4c68-93c5-f5d4daf34201-catalog-content\") pod \"5e1b2757-7f3d-4c68-93c5-f5d4daf34201\" (UID: \"5e1b2757-7f3d-4c68-93c5-f5d4daf34201\") " Nov 22 09:23:33 crc kubenswrapper[4789]: I1122 09:23:33.901636 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rs5vf\" (UniqueName: \"kubernetes.io/projected/5e1b2757-7f3d-4c68-93c5-f5d4daf34201-kube-api-access-rs5vf\") pod \"5e1b2757-7f3d-4c68-93c5-f5d4daf34201\" (UID: \"5e1b2757-7f3d-4c68-93c5-f5d4daf34201\") " Nov 22 09:23:33 crc kubenswrapper[4789]: I1122 09:23:33.902945 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5e1b2757-7f3d-4c68-93c5-f5d4daf34201-utilities" (OuterVolumeSpecName: "utilities") pod "5e1b2757-7f3d-4c68-93c5-f5d4daf34201" (UID: "5e1b2757-7f3d-4c68-93c5-f5d4daf34201"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 09:23:33 crc kubenswrapper[4789]: I1122 09:23:33.907861 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e1b2757-7f3d-4c68-93c5-f5d4daf34201-kube-api-access-rs5vf" (OuterVolumeSpecName: "kube-api-access-rs5vf") pod "5e1b2757-7f3d-4c68-93c5-f5d4daf34201" (UID: "5e1b2757-7f3d-4c68-93c5-f5d4daf34201"). InnerVolumeSpecName "kube-api-access-rs5vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 09:23:33 crc kubenswrapper[4789]: I1122 09:23:33.922298 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5e1b2757-7f3d-4c68-93c5-f5d4daf34201-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5e1b2757-7f3d-4c68-93c5-f5d4daf34201" (UID: "5e1b2757-7f3d-4c68-93c5-f5d4daf34201"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 09:23:34 crc kubenswrapper[4789]: I1122 09:23:34.004604 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e1b2757-7f3d-4c68-93c5-f5d4daf34201-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 09:23:34 crc kubenswrapper[4789]: I1122 09:23:34.004656 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e1b2757-7f3d-4c68-93c5-f5d4daf34201-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 09:23:34 crc kubenswrapper[4789]: I1122 09:23:34.004673 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rs5vf\" (UniqueName: \"kubernetes.io/projected/5e1b2757-7f3d-4c68-93c5-f5d4daf34201-kube-api-access-rs5vf\") on node \"crc\" DevicePath \"\"" Nov 22 09:23:34 crc kubenswrapper[4789]: I1122 09:23:34.203497 4789 generic.go:334] "Generic (PLEG): container finished" podID="5e1b2757-7f3d-4c68-93c5-f5d4daf34201" containerID="98f45bef82901a22528aa1da5a57b09a37e377dc533da5c513ce9faac4b958e2" exitCode=0 Nov 22 09:23:34 crc kubenswrapper[4789]: I1122 09:23:34.203539 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sthhp" event={"ID":"5e1b2757-7f3d-4c68-93c5-f5d4daf34201","Type":"ContainerDied","Data":"98f45bef82901a22528aa1da5a57b09a37e377dc533da5c513ce9faac4b958e2"} Nov 22 09:23:34 crc kubenswrapper[4789]: I1122 09:23:34.203550 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sthhp" Nov 22 09:23:34 crc kubenswrapper[4789]: I1122 09:23:34.203583 4789 scope.go:117] "RemoveContainer" containerID="98f45bef82901a22528aa1da5a57b09a37e377dc533da5c513ce9faac4b958e2" Nov 22 09:23:34 crc kubenswrapper[4789]: I1122 09:23:34.203570 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sthhp" event={"ID":"5e1b2757-7f3d-4c68-93c5-f5d4daf34201","Type":"ContainerDied","Data":"f55f4f8b29361a323dcecaced040ea93b52f6be93c78a33f0fed73739c8b67d0"} Nov 22 09:23:34 crc kubenswrapper[4789]: I1122 09:23:34.227739 4789 scope.go:117] "RemoveContainer" containerID="f8c753b706c16e1719bef41d5b556c945d7cfc8cb28132deb0b1362fa0bf0745" Nov 22 09:23:34 crc kubenswrapper[4789]: I1122 09:23:34.232789 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-sthhp"] Nov 22 09:23:34 crc kubenswrapper[4789]: I1122 09:23:34.244229 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-sthhp"] Nov 22 09:23:34 crc kubenswrapper[4789]: I1122 09:23:34.710665 4789 scope.go:117] "RemoveContainer" containerID="e3455958a7427b2fafda0369092bbbcfd6294543c02d66b1e333be3bab0ec36f" Nov 22 09:23:34 crc kubenswrapper[4789]: I1122 09:23:34.787148 4789 scope.go:117] "RemoveContainer" containerID="98f45bef82901a22528aa1da5a57b09a37e377dc533da5c513ce9faac4b958e2" Nov 22 09:23:34 crc kubenswrapper[4789]: E1122 09:23:34.787656 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"98f45bef82901a22528aa1da5a57b09a37e377dc533da5c513ce9faac4b958e2\": container with ID starting with 98f45bef82901a22528aa1da5a57b09a37e377dc533da5c513ce9faac4b958e2 not found: ID does not exist" containerID="98f45bef82901a22528aa1da5a57b09a37e377dc533da5c513ce9faac4b958e2" Nov 22 09:23:34 crc kubenswrapper[4789]: I1122 09:23:34.787687 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"98f45bef82901a22528aa1da5a57b09a37e377dc533da5c513ce9faac4b958e2"} err="failed to get container status \"98f45bef82901a22528aa1da5a57b09a37e377dc533da5c513ce9faac4b958e2\": rpc error: code = NotFound desc = could not find container \"98f45bef82901a22528aa1da5a57b09a37e377dc533da5c513ce9faac4b958e2\": container with ID starting with 98f45bef82901a22528aa1da5a57b09a37e377dc533da5c513ce9faac4b958e2 not found: ID does not exist" Nov 22 09:23:34 crc kubenswrapper[4789]: I1122 09:23:34.787707 4789 scope.go:117] "RemoveContainer" containerID="f8c753b706c16e1719bef41d5b556c945d7cfc8cb28132deb0b1362fa0bf0745" Nov 22 09:23:34 crc kubenswrapper[4789]: E1122 09:23:34.788158 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f8c753b706c16e1719bef41d5b556c945d7cfc8cb28132deb0b1362fa0bf0745\": container with ID starting with f8c753b706c16e1719bef41d5b556c945d7cfc8cb28132deb0b1362fa0bf0745 not found: ID does not exist" containerID="f8c753b706c16e1719bef41d5b556c945d7cfc8cb28132deb0b1362fa0bf0745" Nov 22 09:23:34 crc kubenswrapper[4789]: I1122 09:23:34.788195 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f8c753b706c16e1719bef41d5b556c945d7cfc8cb28132deb0b1362fa0bf0745"} err="failed to get container status \"f8c753b706c16e1719bef41d5b556c945d7cfc8cb28132deb0b1362fa0bf0745\": rpc error: code = NotFound desc = could not find container \"f8c753b706c16e1719bef41d5b556c945d7cfc8cb28132deb0b1362fa0bf0745\": container with ID starting with f8c753b706c16e1719bef41d5b556c945d7cfc8cb28132deb0b1362fa0bf0745 not found: ID does not exist" Nov 22 09:23:34 crc kubenswrapper[4789]: I1122 09:23:34.788221 4789 scope.go:117] "RemoveContainer" containerID="e3455958a7427b2fafda0369092bbbcfd6294543c02d66b1e333be3bab0ec36f" Nov 22 09:23:34 crc kubenswrapper[4789]: E1122 09:23:34.788581 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e3455958a7427b2fafda0369092bbbcfd6294543c02d66b1e333be3bab0ec36f\": container with ID starting with e3455958a7427b2fafda0369092bbbcfd6294543c02d66b1e333be3bab0ec36f not found: ID does not exist" containerID="e3455958a7427b2fafda0369092bbbcfd6294543c02d66b1e333be3bab0ec36f" Nov 22 09:23:34 crc kubenswrapper[4789]: I1122 09:23:34.788627 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e3455958a7427b2fafda0369092bbbcfd6294543c02d66b1e333be3bab0ec36f"} err="failed to get container status \"e3455958a7427b2fafda0369092bbbcfd6294543c02d66b1e333be3bab0ec36f\": rpc error: code = NotFound desc = could not find container \"e3455958a7427b2fafda0369092bbbcfd6294543c02d66b1e333be3bab0ec36f\": container with ID starting with e3455958a7427b2fafda0369092bbbcfd6294543c02d66b1e333be3bab0ec36f not found: ID does not exist" Nov 22 09:23:35 crc kubenswrapper[4789]: I1122 09:23:35.981357 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5e1b2757-7f3d-4c68-93c5-f5d4daf34201" path="/var/lib/kubelet/pods/5e1b2757-7f3d-4c68-93c5-f5d4daf34201/volumes" Nov 22 09:23:40 crc kubenswrapper[4789]: I1122 09:23:40.965039 4789 scope.go:117] "RemoveContainer" containerID="4240b6f705fa3ecd8a7faa7c66c58a040eaa92955c0d845a27bc191a2ab426e5" Nov 22 09:23:40 crc kubenswrapper[4789]: E1122 09:23:40.965881 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:23:51 crc kubenswrapper[4789]: I1122 09:23:51.965210 4789 scope.go:117] "RemoveContainer" containerID="4240b6f705fa3ecd8a7faa7c66c58a040eaa92955c0d845a27bc191a2ab426e5" Nov 22 09:23:51 crc kubenswrapper[4789]: E1122 09:23:51.965962 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:24:03 crc kubenswrapper[4789]: I1122 09:24:03.966455 4789 scope.go:117] "RemoveContainer" containerID="4240b6f705fa3ecd8a7faa7c66c58a040eaa92955c0d845a27bc191a2ab426e5" Nov 22 09:24:03 crc kubenswrapper[4789]: E1122 09:24:03.967129 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:24:16 crc kubenswrapper[4789]: I1122 09:24:16.966175 4789 scope.go:117] "RemoveContainer" containerID="4240b6f705fa3ecd8a7faa7c66c58a040eaa92955c0d845a27bc191a2ab426e5" Nov 22 09:24:16 crc kubenswrapper[4789]: E1122 09:24:16.967246 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:24:31 crc kubenswrapper[4789]: I1122 09:24:31.965264 4789 scope.go:117] "RemoveContainer" containerID="4240b6f705fa3ecd8a7faa7c66c58a040eaa92955c0d845a27bc191a2ab426e5" Nov 22 09:24:31 crc kubenswrapper[4789]: E1122 09:24:31.966059 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:24:43 crc kubenswrapper[4789]: I1122 09:24:43.966073 4789 scope.go:117] "RemoveContainer" containerID="4240b6f705fa3ecd8a7faa7c66c58a040eaa92955c0d845a27bc191a2ab426e5" Nov 22 09:24:43 crc kubenswrapper[4789]: E1122 09:24:43.967057 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:24:47 crc kubenswrapper[4789]: I1122 09:24:47.648020 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-ntkqv"] Nov 22 09:24:47 crc kubenswrapper[4789]: E1122 09:24:47.649095 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e1b2757-7f3d-4c68-93c5-f5d4daf34201" containerName="extract-content" Nov 22 09:24:47 crc kubenswrapper[4789]: I1122 09:24:47.649113 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e1b2757-7f3d-4c68-93c5-f5d4daf34201" containerName="extract-content" Nov 22 09:24:47 crc kubenswrapper[4789]: E1122 09:24:47.649135 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e1b2757-7f3d-4c68-93c5-f5d4daf34201" containerName="registry-server" Nov 22 09:24:47 crc kubenswrapper[4789]: I1122 09:24:47.649144 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e1b2757-7f3d-4c68-93c5-f5d4daf34201" containerName="registry-server" Nov 22 09:24:47 crc kubenswrapper[4789]: E1122 09:24:47.649173 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e1b2757-7f3d-4c68-93c5-f5d4daf34201" containerName="extract-utilities" Nov 22 09:24:47 crc kubenswrapper[4789]: I1122 09:24:47.649184 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e1b2757-7f3d-4c68-93c5-f5d4daf34201" containerName="extract-utilities" Nov 22 09:24:47 crc kubenswrapper[4789]: I1122 09:24:47.649408 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e1b2757-7f3d-4c68-93c5-f5d4daf34201" containerName="registry-server" Nov 22 09:24:47 crc kubenswrapper[4789]: I1122 09:24:47.650979 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ntkqv" Nov 22 09:24:47 crc kubenswrapper[4789]: I1122 09:24:47.657846 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ntkqv"] Nov 22 09:24:47 crc kubenswrapper[4789]: I1122 09:24:47.802559 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9mxhg\" (UniqueName: \"kubernetes.io/projected/8e84b4c0-ffa0-4e8d-a84c-2e88d4e6ad2a-kube-api-access-9mxhg\") pod \"community-operators-ntkqv\" (UID: \"8e84b4c0-ffa0-4e8d-a84c-2e88d4e6ad2a\") " pod="openshift-marketplace/community-operators-ntkqv" Nov 22 09:24:47 crc kubenswrapper[4789]: I1122 09:24:47.802624 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e84b4c0-ffa0-4e8d-a84c-2e88d4e6ad2a-catalog-content\") pod \"community-operators-ntkqv\" (UID: \"8e84b4c0-ffa0-4e8d-a84c-2e88d4e6ad2a\") " pod="openshift-marketplace/community-operators-ntkqv" Nov 22 09:24:47 crc kubenswrapper[4789]: I1122 09:24:47.802677 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e84b4c0-ffa0-4e8d-a84c-2e88d4e6ad2a-utilities\") pod \"community-operators-ntkqv\" (UID: \"8e84b4c0-ffa0-4e8d-a84c-2e88d4e6ad2a\") " pod="openshift-marketplace/community-operators-ntkqv" Nov 22 09:24:47 crc kubenswrapper[4789]: I1122 09:24:47.904207 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9mxhg\" (UniqueName: \"kubernetes.io/projected/8e84b4c0-ffa0-4e8d-a84c-2e88d4e6ad2a-kube-api-access-9mxhg\") pod \"community-operators-ntkqv\" (UID: \"8e84b4c0-ffa0-4e8d-a84c-2e88d4e6ad2a\") " pod="openshift-marketplace/community-operators-ntkqv" Nov 22 09:24:47 crc kubenswrapper[4789]: I1122 09:24:47.904265 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e84b4c0-ffa0-4e8d-a84c-2e88d4e6ad2a-catalog-content\") pod \"community-operators-ntkqv\" (UID: \"8e84b4c0-ffa0-4e8d-a84c-2e88d4e6ad2a\") " pod="openshift-marketplace/community-operators-ntkqv" Nov 22 09:24:47 crc kubenswrapper[4789]: I1122 09:24:47.904305 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e84b4c0-ffa0-4e8d-a84c-2e88d4e6ad2a-utilities\") pod \"community-operators-ntkqv\" (UID: \"8e84b4c0-ffa0-4e8d-a84c-2e88d4e6ad2a\") " pod="openshift-marketplace/community-operators-ntkqv" Nov 22 09:24:47 crc kubenswrapper[4789]: I1122 09:24:47.904882 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e84b4c0-ffa0-4e8d-a84c-2e88d4e6ad2a-utilities\") pod \"community-operators-ntkqv\" (UID: \"8e84b4c0-ffa0-4e8d-a84c-2e88d4e6ad2a\") " pod="openshift-marketplace/community-operators-ntkqv" Nov 22 09:24:47 crc kubenswrapper[4789]: I1122 09:24:47.904953 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e84b4c0-ffa0-4e8d-a84c-2e88d4e6ad2a-catalog-content\") pod \"community-operators-ntkqv\" (UID: \"8e84b4c0-ffa0-4e8d-a84c-2e88d4e6ad2a\") " pod="openshift-marketplace/community-operators-ntkqv" Nov 22 09:24:47 crc kubenswrapper[4789]: I1122 09:24:47.923133 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9mxhg\" (UniqueName: \"kubernetes.io/projected/8e84b4c0-ffa0-4e8d-a84c-2e88d4e6ad2a-kube-api-access-9mxhg\") pod \"community-operators-ntkqv\" (UID: \"8e84b4c0-ffa0-4e8d-a84c-2e88d4e6ad2a\") " pod="openshift-marketplace/community-operators-ntkqv" Nov 22 09:24:47 crc kubenswrapper[4789]: I1122 09:24:47.990234 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ntkqv" Nov 22 09:24:48 crc kubenswrapper[4789]: I1122 09:24:48.526014 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ntkqv"] Nov 22 09:24:49 crc kubenswrapper[4789]: I1122 09:24:49.049698 4789 generic.go:334] "Generic (PLEG): container finished" podID="8e84b4c0-ffa0-4e8d-a84c-2e88d4e6ad2a" containerID="ca15459126ce9af55ffe8c48256293015f8062d66debee290125c0e008f7c75f" exitCode=0 Nov 22 09:24:49 crc kubenswrapper[4789]: I1122 09:24:49.049775 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ntkqv" event={"ID":"8e84b4c0-ffa0-4e8d-a84c-2e88d4e6ad2a","Type":"ContainerDied","Data":"ca15459126ce9af55ffe8c48256293015f8062d66debee290125c0e008f7c75f"} Nov 22 09:24:49 crc kubenswrapper[4789]: I1122 09:24:49.050123 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ntkqv" event={"ID":"8e84b4c0-ffa0-4e8d-a84c-2e88d4e6ad2a","Type":"ContainerStarted","Data":"99e35c3839eac92a5c3a92bef387df9cc0101016f15e4044bc0b4c99cae37744"} Nov 22 09:24:50 crc kubenswrapper[4789]: I1122 09:24:50.062046 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ntkqv" event={"ID":"8e84b4c0-ffa0-4e8d-a84c-2e88d4e6ad2a","Type":"ContainerStarted","Data":"48581cf44799e784e1934fcf287e204ce38059f9f7ba0d69ce1cc6f4fc70a41b"} Nov 22 09:24:52 crc kubenswrapper[4789]: I1122 09:24:52.080428 4789 generic.go:334] "Generic (PLEG): container finished" podID="8e84b4c0-ffa0-4e8d-a84c-2e88d4e6ad2a" containerID="48581cf44799e784e1934fcf287e204ce38059f9f7ba0d69ce1cc6f4fc70a41b" exitCode=0 Nov 22 09:24:52 crc kubenswrapper[4789]: I1122 09:24:52.080503 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ntkqv" event={"ID":"8e84b4c0-ffa0-4e8d-a84c-2e88d4e6ad2a","Type":"ContainerDied","Data":"48581cf44799e784e1934fcf287e204ce38059f9f7ba0d69ce1cc6f4fc70a41b"} Nov 22 09:24:53 crc kubenswrapper[4789]: I1122 09:24:53.092417 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ntkqv" event={"ID":"8e84b4c0-ffa0-4e8d-a84c-2e88d4e6ad2a","Type":"ContainerStarted","Data":"6cc20946df98a8fc8693d03fe5c415ca7042022a74949c8ded53894243ee497f"} Nov 22 09:24:55 crc kubenswrapper[4789]: I1122 09:24:55.966859 4789 scope.go:117] "RemoveContainer" containerID="4240b6f705fa3ecd8a7faa7c66c58a040eaa92955c0d845a27bc191a2ab426e5" Nov 22 09:24:55 crc kubenswrapper[4789]: E1122 09:24:55.967470 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:24:57 crc kubenswrapper[4789]: I1122 09:24:57.991701 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-ntkqv" Nov 22 09:24:57 crc kubenswrapper[4789]: I1122 09:24:57.992165 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-ntkqv" Nov 22 09:24:58 crc kubenswrapper[4789]: I1122 09:24:58.036880 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-ntkqv" Nov 22 09:24:58 crc kubenswrapper[4789]: I1122 09:24:58.057396 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-ntkqv" podStartSLOduration=7.616976486 podStartE2EDuration="11.057368828s" podCreationTimestamp="2025-11-22 09:24:47 +0000 UTC" firstStartedPulling="2025-11-22 09:24:49.051550616 +0000 UTC m=+5303.285950899" lastFinishedPulling="2025-11-22 09:24:52.491942958 +0000 UTC m=+5306.726343241" observedRunningTime="2025-11-22 09:24:53.114352932 +0000 UTC m=+5307.348753205" watchObservedRunningTime="2025-11-22 09:24:58.057368828 +0000 UTC m=+5312.291769101" Nov 22 09:24:58 crc kubenswrapper[4789]: I1122 09:24:58.171647 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-ntkqv" Nov 22 09:24:58 crc kubenswrapper[4789]: I1122 09:24:58.267428 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ntkqv"] Nov 22 09:25:00 crc kubenswrapper[4789]: I1122 09:25:00.155406 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-ntkqv" podUID="8e84b4c0-ffa0-4e8d-a84c-2e88d4e6ad2a" containerName="registry-server" containerID="cri-o://6cc20946df98a8fc8693d03fe5c415ca7042022a74949c8ded53894243ee497f" gracePeriod=2 Nov 22 09:25:01 crc kubenswrapper[4789]: I1122 09:25:01.213041 4789 generic.go:334] "Generic (PLEG): container finished" podID="8e84b4c0-ffa0-4e8d-a84c-2e88d4e6ad2a" containerID="6cc20946df98a8fc8693d03fe5c415ca7042022a74949c8ded53894243ee497f" exitCode=0 Nov 22 09:25:01 crc kubenswrapper[4789]: I1122 09:25:01.213411 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ntkqv" event={"ID":"8e84b4c0-ffa0-4e8d-a84c-2e88d4e6ad2a","Type":"ContainerDied","Data":"6cc20946df98a8fc8693d03fe5c415ca7042022a74949c8ded53894243ee497f"} Nov 22 09:25:01 crc kubenswrapper[4789]: I1122 09:25:01.530577 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ntkqv" Nov 22 09:25:01 crc kubenswrapper[4789]: I1122 09:25:01.688964 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e84b4c0-ffa0-4e8d-a84c-2e88d4e6ad2a-catalog-content\") pod \"8e84b4c0-ffa0-4e8d-a84c-2e88d4e6ad2a\" (UID: \"8e84b4c0-ffa0-4e8d-a84c-2e88d4e6ad2a\") " Nov 22 09:25:01 crc kubenswrapper[4789]: I1122 09:25:01.689469 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e84b4c0-ffa0-4e8d-a84c-2e88d4e6ad2a-utilities\") pod \"8e84b4c0-ffa0-4e8d-a84c-2e88d4e6ad2a\" (UID: \"8e84b4c0-ffa0-4e8d-a84c-2e88d4e6ad2a\") " Nov 22 09:25:01 crc kubenswrapper[4789]: I1122 09:25:01.689606 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9mxhg\" (UniqueName: \"kubernetes.io/projected/8e84b4c0-ffa0-4e8d-a84c-2e88d4e6ad2a-kube-api-access-9mxhg\") pod \"8e84b4c0-ffa0-4e8d-a84c-2e88d4e6ad2a\" (UID: \"8e84b4c0-ffa0-4e8d-a84c-2e88d4e6ad2a\") " Nov 22 09:25:01 crc kubenswrapper[4789]: I1122 09:25:01.690854 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8e84b4c0-ffa0-4e8d-a84c-2e88d4e6ad2a-utilities" (OuterVolumeSpecName: "utilities") pod "8e84b4c0-ffa0-4e8d-a84c-2e88d4e6ad2a" (UID: "8e84b4c0-ffa0-4e8d-a84c-2e88d4e6ad2a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 09:25:01 crc kubenswrapper[4789]: I1122 09:25:01.698180 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e84b4c0-ffa0-4e8d-a84c-2e88d4e6ad2a-kube-api-access-9mxhg" (OuterVolumeSpecName: "kube-api-access-9mxhg") pod "8e84b4c0-ffa0-4e8d-a84c-2e88d4e6ad2a" (UID: "8e84b4c0-ffa0-4e8d-a84c-2e88d4e6ad2a"). InnerVolumeSpecName "kube-api-access-9mxhg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 09:25:01 crc kubenswrapper[4789]: I1122 09:25:01.792296 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9mxhg\" (UniqueName: \"kubernetes.io/projected/8e84b4c0-ffa0-4e8d-a84c-2e88d4e6ad2a-kube-api-access-9mxhg\") on node \"crc\" DevicePath \"\"" Nov 22 09:25:01 crc kubenswrapper[4789]: I1122 09:25:01.792331 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e84b4c0-ffa0-4e8d-a84c-2e88d4e6ad2a-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 09:25:01 crc kubenswrapper[4789]: I1122 09:25:01.975554 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8e84b4c0-ffa0-4e8d-a84c-2e88d4e6ad2a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8e84b4c0-ffa0-4e8d-a84c-2e88d4e6ad2a" (UID: "8e84b4c0-ffa0-4e8d-a84c-2e88d4e6ad2a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 09:25:01 crc kubenswrapper[4789]: I1122 09:25:01.996030 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e84b4c0-ffa0-4e8d-a84c-2e88d4e6ad2a-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 09:25:02 crc kubenswrapper[4789]: I1122 09:25:02.223477 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ntkqv" event={"ID":"8e84b4c0-ffa0-4e8d-a84c-2e88d4e6ad2a","Type":"ContainerDied","Data":"99e35c3839eac92a5c3a92bef387df9cc0101016f15e4044bc0b4c99cae37744"} Nov 22 09:25:02 crc kubenswrapper[4789]: I1122 09:25:02.223545 4789 scope.go:117] "RemoveContainer" containerID="6cc20946df98a8fc8693d03fe5c415ca7042022a74949c8ded53894243ee497f" Nov 22 09:25:02 crc kubenswrapper[4789]: I1122 09:25:02.223542 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ntkqv" Nov 22 09:25:02 crc kubenswrapper[4789]: I1122 09:25:02.252127 4789 scope.go:117] "RemoveContainer" containerID="48581cf44799e784e1934fcf287e204ce38059f9f7ba0d69ce1cc6f4fc70a41b" Nov 22 09:25:02 crc kubenswrapper[4789]: I1122 09:25:02.259485 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ntkqv"] Nov 22 09:25:02 crc kubenswrapper[4789]: I1122 09:25:02.269421 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-ntkqv"] Nov 22 09:25:02 crc kubenswrapper[4789]: I1122 09:25:02.273060 4789 scope.go:117] "RemoveContainer" containerID="ca15459126ce9af55ffe8c48256293015f8062d66debee290125c0e008f7c75f" Nov 22 09:25:03 crc kubenswrapper[4789]: I1122 09:25:03.977316 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e84b4c0-ffa0-4e8d-a84c-2e88d4e6ad2a" path="/var/lib/kubelet/pods/8e84b4c0-ffa0-4e8d-a84c-2e88d4e6ad2a/volumes" Nov 22 09:25:09 crc kubenswrapper[4789]: I1122 09:25:09.965932 4789 scope.go:117] "RemoveContainer" containerID="4240b6f705fa3ecd8a7faa7c66c58a040eaa92955c0d845a27bc191a2ab426e5" Nov 22 09:25:10 crc kubenswrapper[4789]: I1122 09:25:10.306964 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" event={"ID":"58aa8071-7ecc-4692-8789-2db04ec70647","Type":"ContainerStarted","Data":"ed4b8b74d74a5feda58881cf2c15dfd3369874753b4acfeffa0f8739aacacab0"} Nov 22 09:27:35 crc kubenswrapper[4789]: I1122 09:27:35.372246 4789 patch_prober.go:28] interesting pod/machine-config-daemon-pxvrm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 09:27:35 crc kubenswrapper[4789]: I1122 09:27:35.373050 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 09:28:05 crc kubenswrapper[4789]: I1122 09:28:05.372803 4789 patch_prober.go:28] interesting pod/machine-config-daemon-pxvrm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 09:28:05 crc kubenswrapper[4789]: I1122 09:28:05.373248 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 09:28:35 crc kubenswrapper[4789]: I1122 09:28:35.372566 4789 patch_prober.go:28] interesting pod/machine-config-daemon-pxvrm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 09:28:35 crc kubenswrapper[4789]: I1122 09:28:35.373345 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 09:28:35 crc kubenswrapper[4789]: I1122 09:28:35.373421 4789 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" Nov 22 09:28:35 crc kubenswrapper[4789]: I1122 09:28:35.374519 4789 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ed4b8b74d74a5feda58881cf2c15dfd3369874753b4acfeffa0f8739aacacab0"} pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 22 09:28:35 crc kubenswrapper[4789]: I1122 09:28:35.374622 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" containerID="cri-o://ed4b8b74d74a5feda58881cf2c15dfd3369874753b4acfeffa0f8739aacacab0" gracePeriod=600 Nov 22 09:28:36 crc kubenswrapper[4789]: I1122 09:28:36.145509 4789 generic.go:334] "Generic (PLEG): container finished" podID="58aa8071-7ecc-4692-8789-2db04ec70647" containerID="ed4b8b74d74a5feda58881cf2c15dfd3369874753b4acfeffa0f8739aacacab0" exitCode=0 Nov 22 09:28:36 crc kubenswrapper[4789]: I1122 09:28:36.145590 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" event={"ID":"58aa8071-7ecc-4692-8789-2db04ec70647","Type":"ContainerDied","Data":"ed4b8b74d74a5feda58881cf2c15dfd3369874753b4acfeffa0f8739aacacab0"} Nov 22 09:28:36 crc kubenswrapper[4789]: I1122 09:28:36.146104 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" event={"ID":"58aa8071-7ecc-4692-8789-2db04ec70647","Type":"ContainerStarted","Data":"a2b76708b26cd8ba98b6d3c6d8e8d18687de1639e9e73083d0f8694b338937ca"} Nov 22 09:28:36 crc kubenswrapper[4789]: I1122 09:28:36.146125 4789 scope.go:117] "RemoveContainer" containerID="4240b6f705fa3ecd8a7faa7c66c58a040eaa92955c0d845a27bc191a2ab426e5" Nov 22 09:29:14 crc kubenswrapper[4789]: I1122 09:29:14.719613 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-n6xmf"] Nov 22 09:29:14 crc kubenswrapper[4789]: E1122 09:29:14.720637 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e84b4c0-ffa0-4e8d-a84c-2e88d4e6ad2a" containerName="registry-server" Nov 22 09:29:14 crc kubenswrapper[4789]: I1122 09:29:14.720652 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e84b4c0-ffa0-4e8d-a84c-2e88d4e6ad2a" containerName="registry-server" Nov 22 09:29:14 crc kubenswrapper[4789]: E1122 09:29:14.720702 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e84b4c0-ffa0-4e8d-a84c-2e88d4e6ad2a" containerName="extract-utilities" Nov 22 09:29:14 crc kubenswrapper[4789]: I1122 09:29:14.720711 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e84b4c0-ffa0-4e8d-a84c-2e88d4e6ad2a" containerName="extract-utilities" Nov 22 09:29:14 crc kubenswrapper[4789]: E1122 09:29:14.720729 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e84b4c0-ffa0-4e8d-a84c-2e88d4e6ad2a" containerName="extract-content" Nov 22 09:29:14 crc kubenswrapper[4789]: I1122 09:29:14.720742 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e84b4c0-ffa0-4e8d-a84c-2e88d4e6ad2a" containerName="extract-content" Nov 22 09:29:14 crc kubenswrapper[4789]: I1122 09:29:14.721084 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e84b4c0-ffa0-4e8d-a84c-2e88d4e6ad2a" containerName="registry-server" Nov 22 09:29:14 crc kubenswrapper[4789]: I1122 09:29:14.722662 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-n6xmf" Nov 22 09:29:14 crc kubenswrapper[4789]: I1122 09:29:14.733496 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-n6xmf"] Nov 22 09:29:14 crc kubenswrapper[4789]: I1122 09:29:14.869649 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80835696-3e50-4565-a2b2-2d207248b37b-catalog-content\") pod \"certified-operators-n6xmf\" (UID: \"80835696-3e50-4565-a2b2-2d207248b37b\") " pod="openshift-marketplace/certified-operators-n6xmf" Nov 22 09:29:14 crc kubenswrapper[4789]: I1122 09:29:14.869703 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80835696-3e50-4565-a2b2-2d207248b37b-utilities\") pod \"certified-operators-n6xmf\" (UID: \"80835696-3e50-4565-a2b2-2d207248b37b\") " pod="openshift-marketplace/certified-operators-n6xmf" Nov 22 09:29:14 crc kubenswrapper[4789]: I1122 09:29:14.869739 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jbnkh\" (UniqueName: \"kubernetes.io/projected/80835696-3e50-4565-a2b2-2d207248b37b-kube-api-access-jbnkh\") pod \"certified-operators-n6xmf\" (UID: \"80835696-3e50-4565-a2b2-2d207248b37b\") " pod="openshift-marketplace/certified-operators-n6xmf" Nov 22 09:29:14 crc kubenswrapper[4789]: I1122 09:29:14.972117 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80835696-3e50-4565-a2b2-2d207248b37b-catalog-content\") pod \"certified-operators-n6xmf\" (UID: \"80835696-3e50-4565-a2b2-2d207248b37b\") " pod="openshift-marketplace/certified-operators-n6xmf" Nov 22 09:29:14 crc kubenswrapper[4789]: I1122 09:29:14.972165 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80835696-3e50-4565-a2b2-2d207248b37b-utilities\") pod \"certified-operators-n6xmf\" (UID: \"80835696-3e50-4565-a2b2-2d207248b37b\") " pod="openshift-marketplace/certified-operators-n6xmf" Nov 22 09:29:14 crc kubenswrapper[4789]: I1122 09:29:14.972208 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jbnkh\" (UniqueName: \"kubernetes.io/projected/80835696-3e50-4565-a2b2-2d207248b37b-kube-api-access-jbnkh\") pod \"certified-operators-n6xmf\" (UID: \"80835696-3e50-4565-a2b2-2d207248b37b\") " pod="openshift-marketplace/certified-operators-n6xmf" Nov 22 09:29:14 crc kubenswrapper[4789]: I1122 09:29:14.972691 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80835696-3e50-4565-a2b2-2d207248b37b-catalog-content\") pod \"certified-operators-n6xmf\" (UID: \"80835696-3e50-4565-a2b2-2d207248b37b\") " pod="openshift-marketplace/certified-operators-n6xmf" Nov 22 09:29:14 crc kubenswrapper[4789]: I1122 09:29:14.972773 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80835696-3e50-4565-a2b2-2d207248b37b-utilities\") pod \"certified-operators-n6xmf\" (UID: \"80835696-3e50-4565-a2b2-2d207248b37b\") " pod="openshift-marketplace/certified-operators-n6xmf" Nov 22 09:29:14 crc kubenswrapper[4789]: I1122 09:29:14.996624 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jbnkh\" (UniqueName: \"kubernetes.io/projected/80835696-3e50-4565-a2b2-2d207248b37b-kube-api-access-jbnkh\") pod \"certified-operators-n6xmf\" (UID: \"80835696-3e50-4565-a2b2-2d207248b37b\") " pod="openshift-marketplace/certified-operators-n6xmf" Nov 22 09:29:15 crc kubenswrapper[4789]: I1122 09:29:15.040294 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-n6xmf" Nov 22 09:29:15 crc kubenswrapper[4789]: I1122 09:29:15.632374 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-n6xmf"] Nov 22 09:29:16 crc kubenswrapper[4789]: I1122 09:29:16.518764 4789 generic.go:334] "Generic (PLEG): container finished" podID="80835696-3e50-4565-a2b2-2d207248b37b" containerID="5c52cee4baf23ed6ead07f4cbd81488a8e1d0413a2e055399e2ea18ee612bc44" exitCode=0 Nov 22 09:29:16 crc kubenswrapper[4789]: I1122 09:29:16.519106 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n6xmf" event={"ID":"80835696-3e50-4565-a2b2-2d207248b37b","Type":"ContainerDied","Data":"5c52cee4baf23ed6ead07f4cbd81488a8e1d0413a2e055399e2ea18ee612bc44"} Nov 22 09:29:16 crc kubenswrapper[4789]: I1122 09:29:16.519132 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n6xmf" event={"ID":"80835696-3e50-4565-a2b2-2d207248b37b","Type":"ContainerStarted","Data":"392b9bbe080492370541abca1ffdbd92e7264a5c8a5e4e933a7dee516fdb720d"} Nov 22 09:29:16 crc kubenswrapper[4789]: I1122 09:29:16.524177 4789 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 22 09:29:17 crc kubenswrapper[4789]: I1122 09:29:17.110494 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-mlp25"] Nov 22 09:29:17 crc kubenswrapper[4789]: I1122 09:29:17.112357 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mlp25" Nov 22 09:29:17 crc kubenswrapper[4789]: I1122 09:29:17.123854 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mlp25"] Nov 22 09:29:17 crc kubenswrapper[4789]: I1122 09:29:17.213847 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ea14af8-8fc6-44cf-9131-243a30d4fb47-utilities\") pod \"redhat-operators-mlp25\" (UID: \"7ea14af8-8fc6-44cf-9131-243a30d4fb47\") " pod="openshift-marketplace/redhat-operators-mlp25" Nov 22 09:29:17 crc kubenswrapper[4789]: I1122 09:29:17.214026 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4xdw5\" (UniqueName: \"kubernetes.io/projected/7ea14af8-8fc6-44cf-9131-243a30d4fb47-kube-api-access-4xdw5\") pod \"redhat-operators-mlp25\" (UID: \"7ea14af8-8fc6-44cf-9131-243a30d4fb47\") " pod="openshift-marketplace/redhat-operators-mlp25" Nov 22 09:29:17 crc kubenswrapper[4789]: I1122 09:29:17.214099 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ea14af8-8fc6-44cf-9131-243a30d4fb47-catalog-content\") pod \"redhat-operators-mlp25\" (UID: \"7ea14af8-8fc6-44cf-9131-243a30d4fb47\") " pod="openshift-marketplace/redhat-operators-mlp25" Nov 22 09:29:17 crc kubenswrapper[4789]: I1122 09:29:17.315583 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4xdw5\" (UniqueName: \"kubernetes.io/projected/7ea14af8-8fc6-44cf-9131-243a30d4fb47-kube-api-access-4xdw5\") pod \"redhat-operators-mlp25\" (UID: \"7ea14af8-8fc6-44cf-9131-243a30d4fb47\") " pod="openshift-marketplace/redhat-operators-mlp25" Nov 22 09:29:17 crc kubenswrapper[4789]: I1122 09:29:17.316013 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ea14af8-8fc6-44cf-9131-243a30d4fb47-catalog-content\") pod \"redhat-operators-mlp25\" (UID: \"7ea14af8-8fc6-44cf-9131-243a30d4fb47\") " pod="openshift-marketplace/redhat-operators-mlp25" Nov 22 09:29:17 crc kubenswrapper[4789]: I1122 09:29:17.316090 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ea14af8-8fc6-44cf-9131-243a30d4fb47-utilities\") pod \"redhat-operators-mlp25\" (UID: \"7ea14af8-8fc6-44cf-9131-243a30d4fb47\") " pod="openshift-marketplace/redhat-operators-mlp25" Nov 22 09:29:17 crc kubenswrapper[4789]: I1122 09:29:17.316470 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ea14af8-8fc6-44cf-9131-243a30d4fb47-catalog-content\") pod \"redhat-operators-mlp25\" (UID: \"7ea14af8-8fc6-44cf-9131-243a30d4fb47\") " pod="openshift-marketplace/redhat-operators-mlp25" Nov 22 09:29:17 crc kubenswrapper[4789]: I1122 09:29:17.316522 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ea14af8-8fc6-44cf-9131-243a30d4fb47-utilities\") pod \"redhat-operators-mlp25\" (UID: \"7ea14af8-8fc6-44cf-9131-243a30d4fb47\") " pod="openshift-marketplace/redhat-operators-mlp25" Nov 22 09:29:17 crc kubenswrapper[4789]: I1122 09:29:17.340064 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4xdw5\" (UniqueName: \"kubernetes.io/projected/7ea14af8-8fc6-44cf-9131-243a30d4fb47-kube-api-access-4xdw5\") pod \"redhat-operators-mlp25\" (UID: \"7ea14af8-8fc6-44cf-9131-243a30d4fb47\") " pod="openshift-marketplace/redhat-operators-mlp25" Nov 22 09:29:17 crc kubenswrapper[4789]: I1122 09:29:17.428010 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mlp25" Nov 22 09:29:17 crc kubenswrapper[4789]: I1122 09:29:17.943506 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mlp25"] Nov 22 09:29:18 crc kubenswrapper[4789]: I1122 09:29:18.541282 4789 generic.go:334] "Generic (PLEG): container finished" podID="7ea14af8-8fc6-44cf-9131-243a30d4fb47" containerID="75d212407c49c963f896f6a3904bbdad11cea8fa450197d8177fef9a8bf6f524" exitCode=0 Nov 22 09:29:18 crc kubenswrapper[4789]: I1122 09:29:18.541361 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mlp25" event={"ID":"7ea14af8-8fc6-44cf-9131-243a30d4fb47","Type":"ContainerDied","Data":"75d212407c49c963f896f6a3904bbdad11cea8fa450197d8177fef9a8bf6f524"} Nov 22 09:29:18 crc kubenswrapper[4789]: I1122 09:29:18.541597 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mlp25" event={"ID":"7ea14af8-8fc6-44cf-9131-243a30d4fb47","Type":"ContainerStarted","Data":"d686f116d85acc85ce14cb328262477e62029d6a7c0f76ee6ea9cf321a0bfa4c"} Nov 22 09:29:18 crc kubenswrapper[4789]: I1122 09:29:18.548134 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n6xmf" event={"ID":"80835696-3e50-4565-a2b2-2d207248b37b","Type":"ContainerStarted","Data":"8a551ee59c4b2f7ef2ab55d72d0110dcda4a2bc49908ea57278b84d84e17087e"} Nov 22 09:29:21 crc kubenswrapper[4789]: I1122 09:29:21.586275 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mlp25" event={"ID":"7ea14af8-8fc6-44cf-9131-243a30d4fb47","Type":"ContainerStarted","Data":"9800df6a0dcf3cdb1273e026fb42b8a9af72550d0e0e73cc2e0b81ee11cab05b"} Nov 22 09:29:21 crc kubenswrapper[4789]: I1122 09:29:21.590529 4789 generic.go:334] "Generic (PLEG): container finished" podID="80835696-3e50-4565-a2b2-2d207248b37b" containerID="8a551ee59c4b2f7ef2ab55d72d0110dcda4a2bc49908ea57278b84d84e17087e" exitCode=0 Nov 22 09:29:21 crc kubenswrapper[4789]: I1122 09:29:21.590616 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n6xmf" event={"ID":"80835696-3e50-4565-a2b2-2d207248b37b","Type":"ContainerDied","Data":"8a551ee59c4b2f7ef2ab55d72d0110dcda4a2bc49908ea57278b84d84e17087e"} Nov 22 09:29:23 crc kubenswrapper[4789]: I1122 09:29:23.622281 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n6xmf" event={"ID":"80835696-3e50-4565-a2b2-2d207248b37b","Type":"ContainerStarted","Data":"95f35244a812bee2829602fccd60cd191ff2aa39d4a013cee9bfc9562096ab02"} Nov 22 09:29:23 crc kubenswrapper[4789]: I1122 09:29:23.645383 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-n6xmf" podStartSLOduration=3.297861174 podStartE2EDuration="9.645362149s" podCreationTimestamp="2025-11-22 09:29:14 +0000 UTC" firstStartedPulling="2025-11-22 09:29:16.523969217 +0000 UTC m=+5570.758369490" lastFinishedPulling="2025-11-22 09:29:22.871470192 +0000 UTC m=+5577.105870465" observedRunningTime="2025-11-22 09:29:23.643703484 +0000 UTC m=+5577.878103777" watchObservedRunningTime="2025-11-22 09:29:23.645362149 +0000 UTC m=+5577.879762422" Nov 22 09:29:25 crc kubenswrapper[4789]: I1122 09:29:25.041394 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-n6xmf" Nov 22 09:29:25 crc kubenswrapper[4789]: I1122 09:29:25.042508 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-n6xmf" Nov 22 09:29:26 crc kubenswrapper[4789]: I1122 09:29:26.098237 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-n6xmf" podUID="80835696-3e50-4565-a2b2-2d207248b37b" containerName="registry-server" probeResult="failure" output=< Nov 22 09:29:26 crc kubenswrapper[4789]: timeout: failed to connect service ":50051" within 1s Nov 22 09:29:26 crc kubenswrapper[4789]: > Nov 22 09:29:27 crc kubenswrapper[4789]: I1122 09:29:27.666663 4789 generic.go:334] "Generic (PLEG): container finished" podID="7ea14af8-8fc6-44cf-9131-243a30d4fb47" containerID="9800df6a0dcf3cdb1273e026fb42b8a9af72550d0e0e73cc2e0b81ee11cab05b" exitCode=0 Nov 22 09:29:27 crc kubenswrapper[4789]: I1122 09:29:27.666803 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mlp25" event={"ID":"7ea14af8-8fc6-44cf-9131-243a30d4fb47","Type":"ContainerDied","Data":"9800df6a0dcf3cdb1273e026fb42b8a9af72550d0e0e73cc2e0b81ee11cab05b"} Nov 22 09:29:28 crc kubenswrapper[4789]: I1122 09:29:28.677365 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mlp25" event={"ID":"7ea14af8-8fc6-44cf-9131-243a30d4fb47","Type":"ContainerStarted","Data":"31e6620ed59102d4b37df8485d4a4ca57509d0f0b964c46e5320087cd6a9e53f"} Nov 22 09:29:28 crc kubenswrapper[4789]: I1122 09:29:28.707265 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-mlp25" podStartSLOduration=1.966553658 podStartE2EDuration="11.70724361s" podCreationTimestamp="2025-11-22 09:29:17 +0000 UTC" firstStartedPulling="2025-11-22 09:29:18.544355958 +0000 UTC m=+5572.778756231" lastFinishedPulling="2025-11-22 09:29:28.28504591 +0000 UTC m=+5582.519446183" observedRunningTime="2025-11-22 09:29:28.694878522 +0000 UTC m=+5582.929278795" watchObservedRunningTime="2025-11-22 09:29:28.70724361 +0000 UTC m=+5582.941643883" Nov 22 09:29:36 crc kubenswrapper[4789]: I1122 09:29:36.098037 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-n6xmf" podUID="80835696-3e50-4565-a2b2-2d207248b37b" containerName="registry-server" probeResult="failure" output=< Nov 22 09:29:36 crc kubenswrapper[4789]: timeout: failed to connect service ":50051" within 1s Nov 22 09:29:36 crc kubenswrapper[4789]: > Nov 22 09:29:37 crc kubenswrapper[4789]: I1122 09:29:37.429508 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-mlp25" Nov 22 09:29:37 crc kubenswrapper[4789]: I1122 09:29:37.429581 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-mlp25" Nov 22 09:29:38 crc kubenswrapper[4789]: I1122 09:29:38.488832 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-mlp25" podUID="7ea14af8-8fc6-44cf-9131-243a30d4fb47" containerName="registry-server" probeResult="failure" output=< Nov 22 09:29:38 crc kubenswrapper[4789]: timeout: failed to connect service ":50051" within 1s Nov 22 09:29:38 crc kubenswrapper[4789]: > Nov 22 09:29:45 crc kubenswrapper[4789]: I1122 09:29:45.094127 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-n6xmf" Nov 22 09:29:45 crc kubenswrapper[4789]: I1122 09:29:45.150320 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-n6xmf" Nov 22 09:29:45 crc kubenswrapper[4789]: I1122 09:29:45.923207 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-n6xmf"] Nov 22 09:29:46 crc kubenswrapper[4789]: I1122 09:29:46.879019 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-n6xmf" podUID="80835696-3e50-4565-a2b2-2d207248b37b" containerName="registry-server" containerID="cri-o://95f35244a812bee2829602fccd60cd191ff2aa39d4a013cee9bfc9562096ab02" gracePeriod=2 Nov 22 09:29:47 crc kubenswrapper[4789]: I1122 09:29:47.452636 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-n6xmf" Nov 22 09:29:47 crc kubenswrapper[4789]: I1122 09:29:47.540470 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-mlp25" Nov 22 09:29:47 crc kubenswrapper[4789]: I1122 09:29:47.542003 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80835696-3e50-4565-a2b2-2d207248b37b-utilities\") pod \"80835696-3e50-4565-a2b2-2d207248b37b\" (UID: \"80835696-3e50-4565-a2b2-2d207248b37b\") " Nov 22 09:29:47 crc kubenswrapper[4789]: I1122 09:29:47.542318 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jbnkh\" (UniqueName: \"kubernetes.io/projected/80835696-3e50-4565-a2b2-2d207248b37b-kube-api-access-jbnkh\") pod \"80835696-3e50-4565-a2b2-2d207248b37b\" (UID: \"80835696-3e50-4565-a2b2-2d207248b37b\") " Nov 22 09:29:47 crc kubenswrapper[4789]: I1122 09:29:47.543031 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/80835696-3e50-4565-a2b2-2d207248b37b-utilities" (OuterVolumeSpecName: "utilities") pod "80835696-3e50-4565-a2b2-2d207248b37b" (UID: "80835696-3e50-4565-a2b2-2d207248b37b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 09:29:47 crc kubenswrapper[4789]: I1122 09:29:47.543280 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80835696-3e50-4565-a2b2-2d207248b37b-catalog-content\") pod \"80835696-3e50-4565-a2b2-2d207248b37b\" (UID: \"80835696-3e50-4565-a2b2-2d207248b37b\") " Nov 22 09:29:47 crc kubenswrapper[4789]: I1122 09:29:47.543801 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80835696-3e50-4565-a2b2-2d207248b37b-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 09:29:47 crc kubenswrapper[4789]: I1122 09:29:47.566594 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80835696-3e50-4565-a2b2-2d207248b37b-kube-api-access-jbnkh" (OuterVolumeSpecName: "kube-api-access-jbnkh") pod "80835696-3e50-4565-a2b2-2d207248b37b" (UID: "80835696-3e50-4565-a2b2-2d207248b37b"). InnerVolumeSpecName "kube-api-access-jbnkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 09:29:47 crc kubenswrapper[4789]: I1122 09:29:47.620060 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/80835696-3e50-4565-a2b2-2d207248b37b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "80835696-3e50-4565-a2b2-2d207248b37b" (UID: "80835696-3e50-4565-a2b2-2d207248b37b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 09:29:47 crc kubenswrapper[4789]: I1122 09:29:47.626716 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-mlp25" Nov 22 09:29:47 crc kubenswrapper[4789]: I1122 09:29:47.646354 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jbnkh\" (UniqueName: \"kubernetes.io/projected/80835696-3e50-4565-a2b2-2d207248b37b-kube-api-access-jbnkh\") on node \"crc\" DevicePath \"\"" Nov 22 09:29:47 crc kubenswrapper[4789]: I1122 09:29:47.646673 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80835696-3e50-4565-a2b2-2d207248b37b-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 09:29:47 crc kubenswrapper[4789]: I1122 09:29:47.893529 4789 generic.go:334] "Generic (PLEG): container finished" podID="80835696-3e50-4565-a2b2-2d207248b37b" containerID="95f35244a812bee2829602fccd60cd191ff2aa39d4a013cee9bfc9562096ab02" exitCode=0 Nov 22 09:29:47 crc kubenswrapper[4789]: I1122 09:29:47.893593 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n6xmf" event={"ID":"80835696-3e50-4565-a2b2-2d207248b37b","Type":"ContainerDied","Data":"95f35244a812bee2829602fccd60cd191ff2aa39d4a013cee9bfc9562096ab02"} Nov 22 09:29:47 crc kubenswrapper[4789]: I1122 09:29:47.893645 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n6xmf" event={"ID":"80835696-3e50-4565-a2b2-2d207248b37b","Type":"ContainerDied","Data":"392b9bbe080492370541abca1ffdbd92e7264a5c8a5e4e933a7dee516fdb720d"} Nov 22 09:29:47 crc kubenswrapper[4789]: I1122 09:29:47.893662 4789 scope.go:117] "RemoveContainer" containerID="95f35244a812bee2829602fccd60cd191ff2aa39d4a013cee9bfc9562096ab02" Nov 22 09:29:47 crc kubenswrapper[4789]: I1122 09:29:47.896582 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-n6xmf" Nov 22 09:29:47 crc kubenswrapper[4789]: I1122 09:29:47.945661 4789 scope.go:117] "RemoveContainer" containerID="8a551ee59c4b2f7ef2ab55d72d0110dcda4a2bc49908ea57278b84d84e17087e" Nov 22 09:29:47 crc kubenswrapper[4789]: I1122 09:29:47.964468 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-n6xmf"] Nov 22 09:29:47 crc kubenswrapper[4789]: I1122 09:29:47.979940 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-n6xmf"] Nov 22 09:29:47 crc kubenswrapper[4789]: I1122 09:29:47.987554 4789 scope.go:117] "RemoveContainer" containerID="5c52cee4baf23ed6ead07f4cbd81488a8e1d0413a2e055399e2ea18ee612bc44" Nov 22 09:29:48 crc kubenswrapper[4789]: I1122 09:29:48.033737 4789 scope.go:117] "RemoveContainer" containerID="95f35244a812bee2829602fccd60cd191ff2aa39d4a013cee9bfc9562096ab02" Nov 22 09:29:48 crc kubenswrapper[4789]: E1122 09:29:48.034295 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"95f35244a812bee2829602fccd60cd191ff2aa39d4a013cee9bfc9562096ab02\": container with ID starting with 95f35244a812bee2829602fccd60cd191ff2aa39d4a013cee9bfc9562096ab02 not found: ID does not exist" containerID="95f35244a812bee2829602fccd60cd191ff2aa39d4a013cee9bfc9562096ab02" Nov 22 09:29:48 crc kubenswrapper[4789]: I1122 09:29:48.034340 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"95f35244a812bee2829602fccd60cd191ff2aa39d4a013cee9bfc9562096ab02"} err="failed to get container status \"95f35244a812bee2829602fccd60cd191ff2aa39d4a013cee9bfc9562096ab02\": rpc error: code = NotFound desc = could not find container \"95f35244a812bee2829602fccd60cd191ff2aa39d4a013cee9bfc9562096ab02\": container with ID starting with 95f35244a812bee2829602fccd60cd191ff2aa39d4a013cee9bfc9562096ab02 not found: ID does not exist" Nov 22 09:29:48 crc kubenswrapper[4789]: I1122 09:29:48.034364 4789 scope.go:117] "RemoveContainer" containerID="8a551ee59c4b2f7ef2ab55d72d0110dcda4a2bc49908ea57278b84d84e17087e" Nov 22 09:29:48 crc kubenswrapper[4789]: E1122 09:29:48.034771 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8a551ee59c4b2f7ef2ab55d72d0110dcda4a2bc49908ea57278b84d84e17087e\": container with ID starting with 8a551ee59c4b2f7ef2ab55d72d0110dcda4a2bc49908ea57278b84d84e17087e not found: ID does not exist" containerID="8a551ee59c4b2f7ef2ab55d72d0110dcda4a2bc49908ea57278b84d84e17087e" Nov 22 09:29:48 crc kubenswrapper[4789]: I1122 09:29:48.034810 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a551ee59c4b2f7ef2ab55d72d0110dcda4a2bc49908ea57278b84d84e17087e"} err="failed to get container status \"8a551ee59c4b2f7ef2ab55d72d0110dcda4a2bc49908ea57278b84d84e17087e\": rpc error: code = NotFound desc = could not find container \"8a551ee59c4b2f7ef2ab55d72d0110dcda4a2bc49908ea57278b84d84e17087e\": container with ID starting with 8a551ee59c4b2f7ef2ab55d72d0110dcda4a2bc49908ea57278b84d84e17087e not found: ID does not exist" Nov 22 09:29:48 crc kubenswrapper[4789]: I1122 09:29:48.034863 4789 scope.go:117] "RemoveContainer" containerID="5c52cee4baf23ed6ead07f4cbd81488a8e1d0413a2e055399e2ea18ee612bc44" Nov 22 09:29:48 crc kubenswrapper[4789]: E1122 09:29:48.035340 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5c52cee4baf23ed6ead07f4cbd81488a8e1d0413a2e055399e2ea18ee612bc44\": container with ID starting with 5c52cee4baf23ed6ead07f4cbd81488a8e1d0413a2e055399e2ea18ee612bc44 not found: ID does not exist" containerID="5c52cee4baf23ed6ead07f4cbd81488a8e1d0413a2e055399e2ea18ee612bc44" Nov 22 09:29:48 crc kubenswrapper[4789]: I1122 09:29:48.035368 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c52cee4baf23ed6ead07f4cbd81488a8e1d0413a2e055399e2ea18ee612bc44"} err="failed to get container status \"5c52cee4baf23ed6ead07f4cbd81488a8e1d0413a2e055399e2ea18ee612bc44\": rpc error: code = NotFound desc = could not find container \"5c52cee4baf23ed6ead07f4cbd81488a8e1d0413a2e055399e2ea18ee612bc44\": container with ID starting with 5c52cee4baf23ed6ead07f4cbd81488a8e1d0413a2e055399e2ea18ee612bc44 not found: ID does not exist" Nov 22 09:29:49 crc kubenswrapper[4789]: I1122 09:29:49.117563 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-mlp25"] Nov 22 09:29:49 crc kubenswrapper[4789]: I1122 09:29:49.118097 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-mlp25" podUID="7ea14af8-8fc6-44cf-9131-243a30d4fb47" containerName="registry-server" containerID="cri-o://31e6620ed59102d4b37df8485d4a4ca57509d0f0b964c46e5320087cd6a9e53f" gracePeriod=2 Nov 22 09:29:49 crc kubenswrapper[4789]: I1122 09:29:49.703238 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mlp25" Nov 22 09:29:49 crc kubenswrapper[4789]: I1122 09:29:49.788535 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ea14af8-8fc6-44cf-9131-243a30d4fb47-utilities\") pod \"7ea14af8-8fc6-44cf-9131-243a30d4fb47\" (UID: \"7ea14af8-8fc6-44cf-9131-243a30d4fb47\") " Nov 22 09:29:49 crc kubenswrapper[4789]: I1122 09:29:49.788641 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4xdw5\" (UniqueName: \"kubernetes.io/projected/7ea14af8-8fc6-44cf-9131-243a30d4fb47-kube-api-access-4xdw5\") pod \"7ea14af8-8fc6-44cf-9131-243a30d4fb47\" (UID: \"7ea14af8-8fc6-44cf-9131-243a30d4fb47\") " Nov 22 09:29:49 crc kubenswrapper[4789]: I1122 09:29:49.788794 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ea14af8-8fc6-44cf-9131-243a30d4fb47-catalog-content\") pod \"7ea14af8-8fc6-44cf-9131-243a30d4fb47\" (UID: \"7ea14af8-8fc6-44cf-9131-243a30d4fb47\") " Nov 22 09:29:49 crc kubenswrapper[4789]: I1122 09:29:49.789634 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7ea14af8-8fc6-44cf-9131-243a30d4fb47-utilities" (OuterVolumeSpecName: "utilities") pod "7ea14af8-8fc6-44cf-9131-243a30d4fb47" (UID: "7ea14af8-8fc6-44cf-9131-243a30d4fb47"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 09:29:49 crc kubenswrapper[4789]: I1122 09:29:49.795324 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ea14af8-8fc6-44cf-9131-243a30d4fb47-kube-api-access-4xdw5" (OuterVolumeSpecName: "kube-api-access-4xdw5") pod "7ea14af8-8fc6-44cf-9131-243a30d4fb47" (UID: "7ea14af8-8fc6-44cf-9131-243a30d4fb47"). InnerVolumeSpecName "kube-api-access-4xdw5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 09:29:49 crc kubenswrapper[4789]: I1122 09:29:49.890929 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ea14af8-8fc6-44cf-9131-243a30d4fb47-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 09:29:49 crc kubenswrapper[4789]: I1122 09:29:49.891165 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4xdw5\" (UniqueName: \"kubernetes.io/projected/7ea14af8-8fc6-44cf-9131-243a30d4fb47-kube-api-access-4xdw5\") on node \"crc\" DevicePath \"\"" Nov 22 09:29:49 crc kubenswrapper[4789]: I1122 09:29:49.902828 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7ea14af8-8fc6-44cf-9131-243a30d4fb47-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7ea14af8-8fc6-44cf-9131-243a30d4fb47" (UID: "7ea14af8-8fc6-44cf-9131-243a30d4fb47"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 09:29:49 crc kubenswrapper[4789]: I1122 09:29:49.921830 4789 generic.go:334] "Generic (PLEG): container finished" podID="7ea14af8-8fc6-44cf-9131-243a30d4fb47" containerID="31e6620ed59102d4b37df8485d4a4ca57509d0f0b964c46e5320087cd6a9e53f" exitCode=0 Nov 22 09:29:49 crc kubenswrapper[4789]: I1122 09:29:49.921877 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mlp25" event={"ID":"7ea14af8-8fc6-44cf-9131-243a30d4fb47","Type":"ContainerDied","Data":"31e6620ed59102d4b37df8485d4a4ca57509d0f0b964c46e5320087cd6a9e53f"} Nov 22 09:29:49 crc kubenswrapper[4789]: I1122 09:29:49.921903 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mlp25" event={"ID":"7ea14af8-8fc6-44cf-9131-243a30d4fb47","Type":"ContainerDied","Data":"d686f116d85acc85ce14cb328262477e62029d6a7c0f76ee6ea9cf321a0bfa4c"} Nov 22 09:29:49 crc kubenswrapper[4789]: I1122 09:29:49.921919 4789 scope.go:117] "RemoveContainer" containerID="31e6620ed59102d4b37df8485d4a4ca57509d0f0b964c46e5320087cd6a9e53f" Nov 22 09:29:49 crc kubenswrapper[4789]: I1122 09:29:49.922036 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mlp25" Nov 22 09:29:49 crc kubenswrapper[4789]: I1122 09:29:49.952881 4789 scope.go:117] "RemoveContainer" containerID="9800df6a0dcf3cdb1273e026fb42b8a9af72550d0e0e73cc2e0b81ee11cab05b" Nov 22 09:29:49 crc kubenswrapper[4789]: I1122 09:29:49.984395 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="80835696-3e50-4565-a2b2-2d207248b37b" path="/var/lib/kubelet/pods/80835696-3e50-4565-a2b2-2d207248b37b/volumes" Nov 22 09:29:49 crc kubenswrapper[4789]: I1122 09:29:49.991786 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-mlp25"] Nov 22 09:29:49 crc kubenswrapper[4789]: I1122 09:29:49.992939 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ea14af8-8fc6-44cf-9131-243a30d4fb47-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 09:29:50 crc kubenswrapper[4789]: I1122 09:29:50.019996 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-mlp25"] Nov 22 09:29:50 crc kubenswrapper[4789]: I1122 09:29:50.024549 4789 scope.go:117] "RemoveContainer" containerID="75d212407c49c963f896f6a3904bbdad11cea8fa450197d8177fef9a8bf6f524" Nov 22 09:29:50 crc kubenswrapper[4789]: I1122 09:29:50.069698 4789 scope.go:117] "RemoveContainer" containerID="31e6620ed59102d4b37df8485d4a4ca57509d0f0b964c46e5320087cd6a9e53f" Nov 22 09:29:50 crc kubenswrapper[4789]: E1122 09:29:50.070253 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"31e6620ed59102d4b37df8485d4a4ca57509d0f0b964c46e5320087cd6a9e53f\": container with ID starting with 31e6620ed59102d4b37df8485d4a4ca57509d0f0b964c46e5320087cd6a9e53f not found: ID does not exist" containerID="31e6620ed59102d4b37df8485d4a4ca57509d0f0b964c46e5320087cd6a9e53f" Nov 22 09:29:50 crc kubenswrapper[4789]: I1122 09:29:50.070288 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"31e6620ed59102d4b37df8485d4a4ca57509d0f0b964c46e5320087cd6a9e53f"} err="failed to get container status \"31e6620ed59102d4b37df8485d4a4ca57509d0f0b964c46e5320087cd6a9e53f\": rpc error: code = NotFound desc = could not find container \"31e6620ed59102d4b37df8485d4a4ca57509d0f0b964c46e5320087cd6a9e53f\": container with ID starting with 31e6620ed59102d4b37df8485d4a4ca57509d0f0b964c46e5320087cd6a9e53f not found: ID does not exist" Nov 22 09:29:50 crc kubenswrapper[4789]: I1122 09:29:50.070317 4789 scope.go:117] "RemoveContainer" containerID="9800df6a0dcf3cdb1273e026fb42b8a9af72550d0e0e73cc2e0b81ee11cab05b" Nov 22 09:29:50 crc kubenswrapper[4789]: E1122 09:29:50.070637 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9800df6a0dcf3cdb1273e026fb42b8a9af72550d0e0e73cc2e0b81ee11cab05b\": container with ID starting with 9800df6a0dcf3cdb1273e026fb42b8a9af72550d0e0e73cc2e0b81ee11cab05b not found: ID does not exist" containerID="9800df6a0dcf3cdb1273e026fb42b8a9af72550d0e0e73cc2e0b81ee11cab05b" Nov 22 09:29:50 crc kubenswrapper[4789]: I1122 09:29:50.070694 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9800df6a0dcf3cdb1273e026fb42b8a9af72550d0e0e73cc2e0b81ee11cab05b"} err="failed to get container status \"9800df6a0dcf3cdb1273e026fb42b8a9af72550d0e0e73cc2e0b81ee11cab05b\": rpc error: code = NotFound desc = could not find container \"9800df6a0dcf3cdb1273e026fb42b8a9af72550d0e0e73cc2e0b81ee11cab05b\": container with ID starting with 9800df6a0dcf3cdb1273e026fb42b8a9af72550d0e0e73cc2e0b81ee11cab05b not found: ID does not exist" Nov 22 09:29:50 crc kubenswrapper[4789]: I1122 09:29:50.070726 4789 scope.go:117] "RemoveContainer" containerID="75d212407c49c963f896f6a3904bbdad11cea8fa450197d8177fef9a8bf6f524" Nov 22 09:29:50 crc kubenswrapper[4789]: E1122 09:29:50.071000 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"75d212407c49c963f896f6a3904bbdad11cea8fa450197d8177fef9a8bf6f524\": container with ID starting with 75d212407c49c963f896f6a3904bbdad11cea8fa450197d8177fef9a8bf6f524 not found: ID does not exist" containerID="75d212407c49c963f896f6a3904bbdad11cea8fa450197d8177fef9a8bf6f524" Nov 22 09:29:50 crc kubenswrapper[4789]: I1122 09:29:50.071020 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"75d212407c49c963f896f6a3904bbdad11cea8fa450197d8177fef9a8bf6f524"} err="failed to get container status \"75d212407c49c963f896f6a3904bbdad11cea8fa450197d8177fef9a8bf6f524\": rpc error: code = NotFound desc = could not find container \"75d212407c49c963f896f6a3904bbdad11cea8fa450197d8177fef9a8bf6f524\": container with ID starting with 75d212407c49c963f896f6a3904bbdad11cea8fa450197d8177fef9a8bf6f524 not found: ID does not exist" Nov 22 09:29:51 crc kubenswrapper[4789]: I1122 09:29:51.977727 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ea14af8-8fc6-44cf-9131-243a30d4fb47" path="/var/lib/kubelet/pods/7ea14af8-8fc6-44cf-9131-243a30d4fb47/volumes" Nov 22 09:30:01 crc kubenswrapper[4789]: I1122 09:30:01.851869 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396730-t8nxp"] Nov 22 09:30:01 crc kubenswrapper[4789]: E1122 09:30:01.852972 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ea14af8-8fc6-44cf-9131-243a30d4fb47" containerName="extract-content" Nov 22 09:30:01 crc kubenswrapper[4789]: I1122 09:30:01.852992 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ea14af8-8fc6-44cf-9131-243a30d4fb47" containerName="extract-content" Nov 22 09:30:01 crc kubenswrapper[4789]: E1122 09:30:01.853017 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80835696-3e50-4565-a2b2-2d207248b37b" containerName="extract-utilities" Nov 22 09:30:01 crc kubenswrapper[4789]: I1122 09:30:01.853026 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="80835696-3e50-4565-a2b2-2d207248b37b" containerName="extract-utilities" Nov 22 09:30:01 crc kubenswrapper[4789]: E1122 09:30:01.853040 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80835696-3e50-4565-a2b2-2d207248b37b" containerName="extract-content" Nov 22 09:30:01 crc kubenswrapper[4789]: I1122 09:30:01.853048 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="80835696-3e50-4565-a2b2-2d207248b37b" containerName="extract-content" Nov 22 09:30:01 crc kubenswrapper[4789]: E1122 09:30:01.853081 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ea14af8-8fc6-44cf-9131-243a30d4fb47" containerName="registry-server" Nov 22 09:30:01 crc kubenswrapper[4789]: I1122 09:30:01.853090 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ea14af8-8fc6-44cf-9131-243a30d4fb47" containerName="registry-server" Nov 22 09:30:01 crc kubenswrapper[4789]: E1122 09:30:01.853107 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80835696-3e50-4565-a2b2-2d207248b37b" containerName="registry-server" Nov 22 09:30:01 crc kubenswrapper[4789]: I1122 09:30:01.853115 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="80835696-3e50-4565-a2b2-2d207248b37b" containerName="registry-server" Nov 22 09:30:01 crc kubenswrapper[4789]: E1122 09:30:01.853138 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ea14af8-8fc6-44cf-9131-243a30d4fb47" containerName="extract-utilities" Nov 22 09:30:01 crc kubenswrapper[4789]: I1122 09:30:01.853148 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ea14af8-8fc6-44cf-9131-243a30d4fb47" containerName="extract-utilities" Nov 22 09:30:01 crc kubenswrapper[4789]: I1122 09:30:01.853367 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="80835696-3e50-4565-a2b2-2d207248b37b" containerName="registry-server" Nov 22 09:30:01 crc kubenswrapper[4789]: I1122 09:30:01.853409 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ea14af8-8fc6-44cf-9131-243a30d4fb47" containerName="registry-server" Nov 22 09:30:01 crc kubenswrapper[4789]: I1122 09:30:01.854231 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396730-t8nxp" Nov 22 09:30:01 crc kubenswrapper[4789]: I1122 09:30:01.856808 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 22 09:30:01 crc kubenswrapper[4789]: I1122 09:30:01.857056 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 22 09:30:01 crc kubenswrapper[4789]: I1122 09:30:01.878727 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396730-t8nxp"] Nov 22 09:30:01 crc kubenswrapper[4789]: I1122 09:30:01.903636 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2bf75c3b-b28f-4205-b47b-eb2b7bd7c70e-secret-volume\") pod \"collect-profiles-29396730-t8nxp\" (UID: \"2bf75c3b-b28f-4205-b47b-eb2b7bd7c70e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396730-t8nxp" Nov 22 09:30:01 crc kubenswrapper[4789]: I1122 09:30:01.903821 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dzz5p\" (UniqueName: \"kubernetes.io/projected/2bf75c3b-b28f-4205-b47b-eb2b7bd7c70e-kube-api-access-dzz5p\") pod \"collect-profiles-29396730-t8nxp\" (UID: \"2bf75c3b-b28f-4205-b47b-eb2b7bd7c70e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396730-t8nxp" Nov 22 09:30:01 crc kubenswrapper[4789]: I1122 09:30:01.903955 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2bf75c3b-b28f-4205-b47b-eb2b7bd7c70e-config-volume\") pod \"collect-profiles-29396730-t8nxp\" (UID: \"2bf75c3b-b28f-4205-b47b-eb2b7bd7c70e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396730-t8nxp" Nov 22 09:30:02 crc kubenswrapper[4789]: I1122 09:30:02.006301 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2bf75c3b-b28f-4205-b47b-eb2b7bd7c70e-config-volume\") pod \"collect-profiles-29396730-t8nxp\" (UID: \"2bf75c3b-b28f-4205-b47b-eb2b7bd7c70e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396730-t8nxp" Nov 22 09:30:02 crc kubenswrapper[4789]: I1122 09:30:02.006507 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2bf75c3b-b28f-4205-b47b-eb2b7bd7c70e-secret-volume\") pod \"collect-profiles-29396730-t8nxp\" (UID: \"2bf75c3b-b28f-4205-b47b-eb2b7bd7c70e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396730-t8nxp" Nov 22 09:30:02 crc kubenswrapper[4789]: I1122 09:30:02.006613 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dzz5p\" (UniqueName: \"kubernetes.io/projected/2bf75c3b-b28f-4205-b47b-eb2b7bd7c70e-kube-api-access-dzz5p\") pod \"collect-profiles-29396730-t8nxp\" (UID: \"2bf75c3b-b28f-4205-b47b-eb2b7bd7c70e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396730-t8nxp" Nov 22 09:30:02 crc kubenswrapper[4789]: I1122 09:30:02.007290 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2bf75c3b-b28f-4205-b47b-eb2b7bd7c70e-config-volume\") pod \"collect-profiles-29396730-t8nxp\" (UID: \"2bf75c3b-b28f-4205-b47b-eb2b7bd7c70e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396730-t8nxp" Nov 22 09:30:02 crc kubenswrapper[4789]: I1122 09:30:02.014252 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2bf75c3b-b28f-4205-b47b-eb2b7bd7c70e-secret-volume\") pod \"collect-profiles-29396730-t8nxp\" (UID: \"2bf75c3b-b28f-4205-b47b-eb2b7bd7c70e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396730-t8nxp" Nov 22 09:30:02 crc kubenswrapper[4789]: I1122 09:30:02.037993 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dzz5p\" (UniqueName: \"kubernetes.io/projected/2bf75c3b-b28f-4205-b47b-eb2b7bd7c70e-kube-api-access-dzz5p\") pod \"collect-profiles-29396730-t8nxp\" (UID: \"2bf75c3b-b28f-4205-b47b-eb2b7bd7c70e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396730-t8nxp" Nov 22 09:30:02 crc kubenswrapper[4789]: I1122 09:30:02.171630 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396730-t8nxp" Nov 22 09:30:02 crc kubenswrapper[4789]: I1122 09:30:02.663119 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396730-t8nxp"] Nov 22 09:30:02 crc kubenswrapper[4789]: W1122 09:30:02.666494 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2bf75c3b_b28f_4205_b47b_eb2b7bd7c70e.slice/crio-3cf9ac03c52038bf8eae376b8894e24e6485a4e8113faa1312c99d3029db9575 WatchSource:0}: Error finding container 3cf9ac03c52038bf8eae376b8894e24e6485a4e8113faa1312c99d3029db9575: Status 404 returned error can't find the container with id 3cf9ac03c52038bf8eae376b8894e24e6485a4e8113faa1312c99d3029db9575 Nov 22 09:30:03 crc kubenswrapper[4789]: I1122 09:30:03.110803 4789 generic.go:334] "Generic (PLEG): container finished" podID="2bf75c3b-b28f-4205-b47b-eb2b7bd7c70e" containerID="2dfa9b64332408b98d2c174d7d17576e468a64aec10ff9d95e11b756a14fa78d" exitCode=0 Nov 22 09:30:03 crc kubenswrapper[4789]: I1122 09:30:03.110921 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29396730-t8nxp" event={"ID":"2bf75c3b-b28f-4205-b47b-eb2b7bd7c70e","Type":"ContainerDied","Data":"2dfa9b64332408b98d2c174d7d17576e468a64aec10ff9d95e11b756a14fa78d"} Nov 22 09:30:03 crc kubenswrapper[4789]: I1122 09:30:03.111991 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29396730-t8nxp" event={"ID":"2bf75c3b-b28f-4205-b47b-eb2b7bd7c70e","Type":"ContainerStarted","Data":"3cf9ac03c52038bf8eae376b8894e24e6485a4e8113faa1312c99d3029db9575"} Nov 22 09:30:04 crc kubenswrapper[4789]: I1122 09:30:04.679784 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396730-t8nxp" Nov 22 09:30:04 crc kubenswrapper[4789]: I1122 09:30:04.761902 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dzz5p\" (UniqueName: \"kubernetes.io/projected/2bf75c3b-b28f-4205-b47b-eb2b7bd7c70e-kube-api-access-dzz5p\") pod \"2bf75c3b-b28f-4205-b47b-eb2b7bd7c70e\" (UID: \"2bf75c3b-b28f-4205-b47b-eb2b7bd7c70e\") " Nov 22 09:30:04 crc kubenswrapper[4789]: I1122 09:30:04.762006 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2bf75c3b-b28f-4205-b47b-eb2b7bd7c70e-config-volume\") pod \"2bf75c3b-b28f-4205-b47b-eb2b7bd7c70e\" (UID: \"2bf75c3b-b28f-4205-b47b-eb2b7bd7c70e\") " Nov 22 09:30:04 crc kubenswrapper[4789]: I1122 09:30:04.762339 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2bf75c3b-b28f-4205-b47b-eb2b7bd7c70e-secret-volume\") pod \"2bf75c3b-b28f-4205-b47b-eb2b7bd7c70e\" (UID: \"2bf75c3b-b28f-4205-b47b-eb2b7bd7c70e\") " Nov 22 09:30:04 crc kubenswrapper[4789]: I1122 09:30:04.763595 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2bf75c3b-b28f-4205-b47b-eb2b7bd7c70e-config-volume" (OuterVolumeSpecName: "config-volume") pod "2bf75c3b-b28f-4205-b47b-eb2b7bd7c70e" (UID: "2bf75c3b-b28f-4205-b47b-eb2b7bd7c70e"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 09:30:04 crc kubenswrapper[4789]: I1122 09:30:04.779345 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2bf75c3b-b28f-4205-b47b-eb2b7bd7c70e-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "2bf75c3b-b28f-4205-b47b-eb2b7bd7c70e" (UID: "2bf75c3b-b28f-4205-b47b-eb2b7bd7c70e"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 09:30:04 crc kubenswrapper[4789]: I1122 09:30:04.785001 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2bf75c3b-b28f-4205-b47b-eb2b7bd7c70e-kube-api-access-dzz5p" (OuterVolumeSpecName: "kube-api-access-dzz5p") pod "2bf75c3b-b28f-4205-b47b-eb2b7bd7c70e" (UID: "2bf75c3b-b28f-4205-b47b-eb2b7bd7c70e"). InnerVolumeSpecName "kube-api-access-dzz5p". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 09:30:04 crc kubenswrapper[4789]: I1122 09:30:04.864597 4789 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2bf75c3b-b28f-4205-b47b-eb2b7bd7c70e-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 22 09:30:04 crc kubenswrapper[4789]: I1122 09:30:04.864633 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dzz5p\" (UniqueName: \"kubernetes.io/projected/2bf75c3b-b28f-4205-b47b-eb2b7bd7c70e-kube-api-access-dzz5p\") on node \"crc\" DevicePath \"\"" Nov 22 09:30:04 crc kubenswrapper[4789]: I1122 09:30:04.864643 4789 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2bf75c3b-b28f-4205-b47b-eb2b7bd7c70e-config-volume\") on node \"crc\" DevicePath \"\"" Nov 22 09:30:05 crc kubenswrapper[4789]: I1122 09:30:05.136921 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29396730-t8nxp" event={"ID":"2bf75c3b-b28f-4205-b47b-eb2b7bd7c70e","Type":"ContainerDied","Data":"3cf9ac03c52038bf8eae376b8894e24e6485a4e8113faa1312c99d3029db9575"} Nov 22 09:30:05 crc kubenswrapper[4789]: I1122 09:30:05.136960 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3cf9ac03c52038bf8eae376b8894e24e6485a4e8113faa1312c99d3029db9575" Nov 22 09:30:05 crc kubenswrapper[4789]: I1122 09:30:05.137036 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396730-t8nxp" Nov 22 09:30:05 crc kubenswrapper[4789]: I1122 09:30:05.782361 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396685-jt9jd"] Nov 22 09:30:05 crc kubenswrapper[4789]: I1122 09:30:05.797678 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396685-jt9jd"] Nov 22 09:30:05 crc kubenswrapper[4789]: I1122 09:30:05.980931 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="625a9eb2-dfa9-4b93-b6b9-17402117c6a2" path="/var/lib/kubelet/pods/625a9eb2-dfa9-4b93-b6b9-17402117c6a2/volumes" Nov 22 09:30:29 crc kubenswrapper[4789]: I1122 09:30:29.264337 4789 scope.go:117] "RemoveContainer" containerID="6ce056936c6c9b5253e11a773ae44c4f000ac7e44793b4a172a4c739489b03ac" Nov 22 09:30:35 crc kubenswrapper[4789]: I1122 09:30:35.372451 4789 patch_prober.go:28] interesting pod/machine-config-daemon-pxvrm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 09:30:35 crc kubenswrapper[4789]: I1122 09:30:35.373042 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 09:31:05 crc kubenswrapper[4789]: I1122 09:31:05.371824 4789 patch_prober.go:28] interesting pod/machine-config-daemon-pxvrm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 09:31:05 crc kubenswrapper[4789]: I1122 09:31:05.372441 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 09:31:35 crc kubenswrapper[4789]: I1122 09:31:35.372017 4789 patch_prober.go:28] interesting pod/machine-config-daemon-pxvrm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 09:31:35 crc kubenswrapper[4789]: I1122 09:31:35.372547 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 09:31:35 crc kubenswrapper[4789]: I1122 09:31:35.372589 4789 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" Nov 22 09:31:35 crc kubenswrapper[4789]: I1122 09:31:35.373300 4789 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a2b76708b26cd8ba98b6d3c6d8e8d18687de1639e9e73083d0f8694b338937ca"} pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 22 09:31:35 crc kubenswrapper[4789]: I1122 09:31:35.373352 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" containerID="cri-o://a2b76708b26cd8ba98b6d3c6d8e8d18687de1639e9e73083d0f8694b338937ca" gracePeriod=600 Nov 22 09:31:35 crc kubenswrapper[4789]: E1122 09:31:35.496614 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:31:36 crc kubenswrapper[4789]: I1122 09:31:36.018728 4789 generic.go:334] "Generic (PLEG): container finished" podID="58aa8071-7ecc-4692-8789-2db04ec70647" containerID="a2b76708b26cd8ba98b6d3c6d8e8d18687de1639e9e73083d0f8694b338937ca" exitCode=0 Nov 22 09:31:36 crc kubenswrapper[4789]: I1122 09:31:36.018776 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" event={"ID":"58aa8071-7ecc-4692-8789-2db04ec70647","Type":"ContainerDied","Data":"a2b76708b26cd8ba98b6d3c6d8e8d18687de1639e9e73083d0f8694b338937ca"} Nov 22 09:31:36 crc kubenswrapper[4789]: I1122 09:31:36.018829 4789 scope.go:117] "RemoveContainer" containerID="ed4b8b74d74a5feda58881cf2c15dfd3369874753b4acfeffa0f8739aacacab0" Nov 22 09:31:36 crc kubenswrapper[4789]: I1122 09:31:36.019386 4789 scope.go:117] "RemoveContainer" containerID="a2b76708b26cd8ba98b6d3c6d8e8d18687de1639e9e73083d0f8694b338937ca" Nov 22 09:31:36 crc kubenswrapper[4789]: E1122 09:31:36.019623 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:31:48 crc kubenswrapper[4789]: I1122 09:31:48.965537 4789 scope.go:117] "RemoveContainer" containerID="a2b76708b26cd8ba98b6d3c6d8e8d18687de1639e9e73083d0f8694b338937ca" Nov 22 09:31:48 crc kubenswrapper[4789]: E1122 09:31:48.966362 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:32:03 crc kubenswrapper[4789]: I1122 09:32:03.965654 4789 scope.go:117] "RemoveContainer" containerID="a2b76708b26cd8ba98b6d3c6d8e8d18687de1639e9e73083d0f8694b338937ca" Nov 22 09:32:03 crc kubenswrapper[4789]: E1122 09:32:03.966639 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:32:18 crc kubenswrapper[4789]: I1122 09:32:18.965597 4789 scope.go:117] "RemoveContainer" containerID="a2b76708b26cd8ba98b6d3c6d8e8d18687de1639e9e73083d0f8694b338937ca" Nov 22 09:32:18 crc kubenswrapper[4789]: E1122 09:32:18.966326 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:32:30 crc kubenswrapper[4789]: I1122 09:32:30.965187 4789 scope.go:117] "RemoveContainer" containerID="a2b76708b26cd8ba98b6d3c6d8e8d18687de1639e9e73083d0f8694b338937ca" Nov 22 09:32:30 crc kubenswrapper[4789]: E1122 09:32:30.966133 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:32:41 crc kubenswrapper[4789]: I1122 09:32:41.965907 4789 scope.go:117] "RemoveContainer" containerID="a2b76708b26cd8ba98b6d3c6d8e8d18687de1639e9e73083d0f8694b338937ca" Nov 22 09:32:41 crc kubenswrapper[4789]: E1122 09:32:41.966846 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:32:49 crc kubenswrapper[4789]: I1122 09:32:49.801154 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="deee1bbe-0f19-4f57-bebc-de5b01ccac7f" containerName="ceilometer-notification-agent" probeResult="failure" output="command timed out" Nov 22 09:32:49 crc kubenswrapper[4789]: I1122 09:32:49.801151 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="deee1bbe-0f19-4f57-bebc-de5b01ccac7f" containerName="ceilometer-central-agent" probeResult="failure" output="command timed out" Nov 22 09:32:52 crc kubenswrapper[4789]: I1122 09:32:52.966651 4789 scope.go:117] "RemoveContainer" containerID="a2b76708b26cd8ba98b6d3c6d8e8d18687de1639e9e73083d0f8694b338937ca" Nov 22 09:32:52 crc kubenswrapper[4789]: E1122 09:32:52.968058 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:33:06 crc kubenswrapper[4789]: I1122 09:33:06.965205 4789 scope.go:117] "RemoveContainer" containerID="a2b76708b26cd8ba98b6d3c6d8e8d18687de1639e9e73083d0f8694b338937ca" Nov 22 09:33:06 crc kubenswrapper[4789]: E1122 09:33:06.966025 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:33:20 crc kubenswrapper[4789]: I1122 09:33:20.965784 4789 scope.go:117] "RemoveContainer" containerID="a2b76708b26cd8ba98b6d3c6d8e8d18687de1639e9e73083d0f8694b338937ca" Nov 22 09:33:20 crc kubenswrapper[4789]: E1122 09:33:20.966651 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:33:22 crc kubenswrapper[4789]: I1122 09:33:22.273763 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-kgfsm"] Nov 22 09:33:22 crc kubenswrapper[4789]: E1122 09:33:22.274701 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bf75c3b-b28f-4205-b47b-eb2b7bd7c70e" containerName="collect-profiles" Nov 22 09:33:22 crc kubenswrapper[4789]: I1122 09:33:22.274717 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bf75c3b-b28f-4205-b47b-eb2b7bd7c70e" containerName="collect-profiles" Nov 22 09:33:22 crc kubenswrapper[4789]: I1122 09:33:22.274949 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="2bf75c3b-b28f-4205-b47b-eb2b7bd7c70e" containerName="collect-profiles" Nov 22 09:33:22 crc kubenswrapper[4789]: I1122 09:33:22.276549 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kgfsm" Nov 22 09:33:22 crc kubenswrapper[4789]: I1122 09:33:22.309848 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kgfsm"] Nov 22 09:33:22 crc kubenswrapper[4789]: I1122 09:33:22.399345 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kr2r5\" (UniqueName: \"kubernetes.io/projected/5f84d005-b7bd-4062-997e-e30019d0233c-kube-api-access-kr2r5\") pod \"redhat-marketplace-kgfsm\" (UID: \"5f84d005-b7bd-4062-997e-e30019d0233c\") " pod="openshift-marketplace/redhat-marketplace-kgfsm" Nov 22 09:33:22 crc kubenswrapper[4789]: I1122 09:33:22.399762 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f84d005-b7bd-4062-997e-e30019d0233c-utilities\") pod \"redhat-marketplace-kgfsm\" (UID: \"5f84d005-b7bd-4062-997e-e30019d0233c\") " pod="openshift-marketplace/redhat-marketplace-kgfsm" Nov 22 09:33:22 crc kubenswrapper[4789]: I1122 09:33:22.399838 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f84d005-b7bd-4062-997e-e30019d0233c-catalog-content\") pod \"redhat-marketplace-kgfsm\" (UID: \"5f84d005-b7bd-4062-997e-e30019d0233c\") " pod="openshift-marketplace/redhat-marketplace-kgfsm" Nov 22 09:33:22 crc kubenswrapper[4789]: I1122 09:33:22.501158 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f84d005-b7bd-4062-997e-e30019d0233c-catalog-content\") pod \"redhat-marketplace-kgfsm\" (UID: \"5f84d005-b7bd-4062-997e-e30019d0233c\") " pod="openshift-marketplace/redhat-marketplace-kgfsm" Nov 22 09:33:22 crc kubenswrapper[4789]: I1122 09:33:22.501267 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kr2r5\" (UniqueName: \"kubernetes.io/projected/5f84d005-b7bd-4062-997e-e30019d0233c-kube-api-access-kr2r5\") pod \"redhat-marketplace-kgfsm\" (UID: \"5f84d005-b7bd-4062-997e-e30019d0233c\") " pod="openshift-marketplace/redhat-marketplace-kgfsm" Nov 22 09:33:22 crc kubenswrapper[4789]: I1122 09:33:22.501340 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f84d005-b7bd-4062-997e-e30019d0233c-utilities\") pod \"redhat-marketplace-kgfsm\" (UID: \"5f84d005-b7bd-4062-997e-e30019d0233c\") " pod="openshift-marketplace/redhat-marketplace-kgfsm" Nov 22 09:33:22 crc kubenswrapper[4789]: I1122 09:33:22.501901 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f84d005-b7bd-4062-997e-e30019d0233c-catalog-content\") pod \"redhat-marketplace-kgfsm\" (UID: \"5f84d005-b7bd-4062-997e-e30019d0233c\") " pod="openshift-marketplace/redhat-marketplace-kgfsm" Nov 22 09:33:22 crc kubenswrapper[4789]: I1122 09:33:22.501973 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f84d005-b7bd-4062-997e-e30019d0233c-utilities\") pod \"redhat-marketplace-kgfsm\" (UID: \"5f84d005-b7bd-4062-997e-e30019d0233c\") " pod="openshift-marketplace/redhat-marketplace-kgfsm" Nov 22 09:33:22 crc kubenswrapper[4789]: I1122 09:33:22.520570 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kr2r5\" (UniqueName: \"kubernetes.io/projected/5f84d005-b7bd-4062-997e-e30019d0233c-kube-api-access-kr2r5\") pod \"redhat-marketplace-kgfsm\" (UID: \"5f84d005-b7bd-4062-997e-e30019d0233c\") " pod="openshift-marketplace/redhat-marketplace-kgfsm" Nov 22 09:33:22 crc kubenswrapper[4789]: I1122 09:33:22.617078 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kgfsm" Nov 22 09:33:23 crc kubenswrapper[4789]: I1122 09:33:23.067449 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kgfsm"] Nov 22 09:33:23 crc kubenswrapper[4789]: W1122 09:33:23.070699 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5f84d005_b7bd_4062_997e_e30019d0233c.slice/crio-2bfc651c7ab146c972d3d710c98b7685a906fb7fb9a034bc3455ddff44c8aa53 WatchSource:0}: Error finding container 2bfc651c7ab146c972d3d710c98b7685a906fb7fb9a034bc3455ddff44c8aa53: Status 404 returned error can't find the container with id 2bfc651c7ab146c972d3d710c98b7685a906fb7fb9a034bc3455ddff44c8aa53 Nov 22 09:33:23 crc kubenswrapper[4789]: I1122 09:33:23.113249 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kgfsm" event={"ID":"5f84d005-b7bd-4062-997e-e30019d0233c","Type":"ContainerStarted","Data":"2bfc651c7ab146c972d3d710c98b7685a906fb7fb9a034bc3455ddff44c8aa53"} Nov 22 09:33:24 crc kubenswrapper[4789]: I1122 09:33:24.121515 4789 generic.go:334] "Generic (PLEG): container finished" podID="5f84d005-b7bd-4062-997e-e30019d0233c" containerID="da6d19b59583f367dcaee81085c2b5d2366752cf837f7c02be6908ba3188012c" exitCode=0 Nov 22 09:33:24 crc kubenswrapper[4789]: I1122 09:33:24.123908 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kgfsm" event={"ID":"5f84d005-b7bd-4062-997e-e30019d0233c","Type":"ContainerDied","Data":"da6d19b59583f367dcaee81085c2b5d2366752cf837f7c02be6908ba3188012c"} Nov 22 09:33:25 crc kubenswrapper[4789]: I1122 09:33:25.135022 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kgfsm" event={"ID":"5f84d005-b7bd-4062-997e-e30019d0233c","Type":"ContainerStarted","Data":"db4568b67635e08defa4bdb6e6170bdc8f4905e2d5a7d6684ea454f7fd045206"} Nov 22 09:33:26 crc kubenswrapper[4789]: I1122 09:33:26.145709 4789 generic.go:334] "Generic (PLEG): container finished" podID="5f84d005-b7bd-4062-997e-e30019d0233c" containerID="db4568b67635e08defa4bdb6e6170bdc8f4905e2d5a7d6684ea454f7fd045206" exitCode=0 Nov 22 09:33:26 crc kubenswrapper[4789]: I1122 09:33:26.145791 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kgfsm" event={"ID":"5f84d005-b7bd-4062-997e-e30019d0233c","Type":"ContainerDied","Data":"db4568b67635e08defa4bdb6e6170bdc8f4905e2d5a7d6684ea454f7fd045206"} Nov 22 09:33:27 crc kubenswrapper[4789]: I1122 09:33:27.159924 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kgfsm" event={"ID":"5f84d005-b7bd-4062-997e-e30019d0233c","Type":"ContainerStarted","Data":"b786fcaab043ce9f82405121bf1ab32978f777a194edb3d584aafe070823ddd8"} Nov 22 09:33:32 crc kubenswrapper[4789]: I1122 09:33:32.617659 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-kgfsm" Nov 22 09:33:32 crc kubenswrapper[4789]: I1122 09:33:32.618294 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-kgfsm" Nov 22 09:33:32 crc kubenswrapper[4789]: I1122 09:33:32.668501 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-kgfsm" Nov 22 09:33:32 crc kubenswrapper[4789]: I1122 09:33:32.704544 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-kgfsm" podStartSLOduration=8.260990403 podStartE2EDuration="10.704521357s" podCreationTimestamp="2025-11-22 09:33:22 +0000 UTC" firstStartedPulling="2025-11-22 09:33:24.125017554 +0000 UTC m=+5818.359417827" lastFinishedPulling="2025-11-22 09:33:26.568548508 +0000 UTC m=+5820.802948781" observedRunningTime="2025-11-22 09:33:27.181476342 +0000 UTC m=+5821.415876625" watchObservedRunningTime="2025-11-22 09:33:32.704521357 +0000 UTC m=+5826.938921640" Nov 22 09:33:33 crc kubenswrapper[4789]: I1122 09:33:33.294391 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-kgfsm" Nov 22 09:33:33 crc kubenswrapper[4789]: I1122 09:33:33.362587 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-kgfsm"] Nov 22 09:33:33 crc kubenswrapper[4789]: I1122 09:33:33.967038 4789 scope.go:117] "RemoveContainer" containerID="a2b76708b26cd8ba98b6d3c6d8e8d18687de1639e9e73083d0f8694b338937ca" Nov 22 09:33:33 crc kubenswrapper[4789]: E1122 09:33:33.967465 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:33:35 crc kubenswrapper[4789]: I1122 09:33:35.251529 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-kgfsm" podUID="5f84d005-b7bd-4062-997e-e30019d0233c" containerName="registry-server" containerID="cri-o://b786fcaab043ce9f82405121bf1ab32978f777a194edb3d584aafe070823ddd8" gracePeriod=2 Nov 22 09:33:35 crc kubenswrapper[4789]: I1122 09:33:35.900252 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kgfsm" Nov 22 09:33:35 crc kubenswrapper[4789]: I1122 09:33:35.995495 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f84d005-b7bd-4062-997e-e30019d0233c-catalog-content\") pod \"5f84d005-b7bd-4062-997e-e30019d0233c\" (UID: \"5f84d005-b7bd-4062-997e-e30019d0233c\") " Nov 22 09:33:35 crc kubenswrapper[4789]: I1122 09:33:35.995595 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kr2r5\" (UniqueName: \"kubernetes.io/projected/5f84d005-b7bd-4062-997e-e30019d0233c-kube-api-access-kr2r5\") pod \"5f84d005-b7bd-4062-997e-e30019d0233c\" (UID: \"5f84d005-b7bd-4062-997e-e30019d0233c\") " Nov 22 09:33:35 crc kubenswrapper[4789]: I1122 09:33:35.995794 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f84d005-b7bd-4062-997e-e30019d0233c-utilities\") pod \"5f84d005-b7bd-4062-997e-e30019d0233c\" (UID: \"5f84d005-b7bd-4062-997e-e30019d0233c\") " Nov 22 09:33:35 crc kubenswrapper[4789]: I1122 09:33:35.996898 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5f84d005-b7bd-4062-997e-e30019d0233c-utilities" (OuterVolumeSpecName: "utilities") pod "5f84d005-b7bd-4062-997e-e30019d0233c" (UID: "5f84d005-b7bd-4062-997e-e30019d0233c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 09:33:36 crc kubenswrapper[4789]: I1122 09:33:36.003423 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f84d005-b7bd-4062-997e-e30019d0233c-kube-api-access-kr2r5" (OuterVolumeSpecName: "kube-api-access-kr2r5") pod "5f84d005-b7bd-4062-997e-e30019d0233c" (UID: "5f84d005-b7bd-4062-997e-e30019d0233c"). InnerVolumeSpecName "kube-api-access-kr2r5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 09:33:36 crc kubenswrapper[4789]: I1122 09:33:36.024181 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5f84d005-b7bd-4062-997e-e30019d0233c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5f84d005-b7bd-4062-997e-e30019d0233c" (UID: "5f84d005-b7bd-4062-997e-e30019d0233c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 09:33:36 crc kubenswrapper[4789]: I1122 09:33:36.099474 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f84d005-b7bd-4062-997e-e30019d0233c-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 09:33:36 crc kubenswrapper[4789]: I1122 09:33:36.099536 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kr2r5\" (UniqueName: \"kubernetes.io/projected/5f84d005-b7bd-4062-997e-e30019d0233c-kube-api-access-kr2r5\") on node \"crc\" DevicePath \"\"" Nov 22 09:33:36 crc kubenswrapper[4789]: I1122 09:33:36.099557 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f84d005-b7bd-4062-997e-e30019d0233c-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 09:33:36 crc kubenswrapper[4789]: I1122 09:33:36.262633 4789 generic.go:334] "Generic (PLEG): container finished" podID="5f84d005-b7bd-4062-997e-e30019d0233c" containerID="b786fcaab043ce9f82405121bf1ab32978f777a194edb3d584aafe070823ddd8" exitCode=0 Nov 22 09:33:36 crc kubenswrapper[4789]: I1122 09:33:36.262692 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kgfsm" event={"ID":"5f84d005-b7bd-4062-997e-e30019d0233c","Type":"ContainerDied","Data":"b786fcaab043ce9f82405121bf1ab32978f777a194edb3d584aafe070823ddd8"} Nov 22 09:33:36 crc kubenswrapper[4789]: I1122 09:33:36.262692 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kgfsm" Nov 22 09:33:36 crc kubenswrapper[4789]: I1122 09:33:36.262744 4789 scope.go:117] "RemoveContainer" containerID="b786fcaab043ce9f82405121bf1ab32978f777a194edb3d584aafe070823ddd8" Nov 22 09:33:36 crc kubenswrapper[4789]: I1122 09:33:36.262728 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kgfsm" event={"ID":"5f84d005-b7bd-4062-997e-e30019d0233c","Type":"ContainerDied","Data":"2bfc651c7ab146c972d3d710c98b7685a906fb7fb9a034bc3455ddff44c8aa53"} Nov 22 09:33:36 crc kubenswrapper[4789]: I1122 09:33:36.286497 4789 scope.go:117] "RemoveContainer" containerID="db4568b67635e08defa4bdb6e6170bdc8f4905e2d5a7d6684ea454f7fd045206" Nov 22 09:33:36 crc kubenswrapper[4789]: I1122 09:33:36.314360 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-kgfsm"] Nov 22 09:33:36 crc kubenswrapper[4789]: I1122 09:33:36.321352 4789 scope.go:117] "RemoveContainer" containerID="da6d19b59583f367dcaee81085c2b5d2366752cf837f7c02be6908ba3188012c" Nov 22 09:33:36 crc kubenswrapper[4789]: I1122 09:33:36.323584 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-kgfsm"] Nov 22 09:33:36 crc kubenswrapper[4789]: I1122 09:33:36.361658 4789 scope.go:117] "RemoveContainer" containerID="b786fcaab043ce9f82405121bf1ab32978f777a194edb3d584aafe070823ddd8" Nov 22 09:33:36 crc kubenswrapper[4789]: E1122 09:33:36.362189 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b786fcaab043ce9f82405121bf1ab32978f777a194edb3d584aafe070823ddd8\": container with ID starting with b786fcaab043ce9f82405121bf1ab32978f777a194edb3d584aafe070823ddd8 not found: ID does not exist" containerID="b786fcaab043ce9f82405121bf1ab32978f777a194edb3d584aafe070823ddd8" Nov 22 09:33:36 crc kubenswrapper[4789]: I1122 09:33:36.362246 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b786fcaab043ce9f82405121bf1ab32978f777a194edb3d584aafe070823ddd8"} err="failed to get container status \"b786fcaab043ce9f82405121bf1ab32978f777a194edb3d584aafe070823ddd8\": rpc error: code = NotFound desc = could not find container \"b786fcaab043ce9f82405121bf1ab32978f777a194edb3d584aafe070823ddd8\": container with ID starting with b786fcaab043ce9f82405121bf1ab32978f777a194edb3d584aafe070823ddd8 not found: ID does not exist" Nov 22 09:33:36 crc kubenswrapper[4789]: I1122 09:33:36.362278 4789 scope.go:117] "RemoveContainer" containerID="db4568b67635e08defa4bdb6e6170bdc8f4905e2d5a7d6684ea454f7fd045206" Nov 22 09:33:36 crc kubenswrapper[4789]: E1122 09:33:36.362839 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db4568b67635e08defa4bdb6e6170bdc8f4905e2d5a7d6684ea454f7fd045206\": container with ID starting with db4568b67635e08defa4bdb6e6170bdc8f4905e2d5a7d6684ea454f7fd045206 not found: ID does not exist" containerID="db4568b67635e08defa4bdb6e6170bdc8f4905e2d5a7d6684ea454f7fd045206" Nov 22 09:33:36 crc kubenswrapper[4789]: I1122 09:33:36.362891 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db4568b67635e08defa4bdb6e6170bdc8f4905e2d5a7d6684ea454f7fd045206"} err="failed to get container status \"db4568b67635e08defa4bdb6e6170bdc8f4905e2d5a7d6684ea454f7fd045206\": rpc error: code = NotFound desc = could not find container \"db4568b67635e08defa4bdb6e6170bdc8f4905e2d5a7d6684ea454f7fd045206\": container with ID starting with db4568b67635e08defa4bdb6e6170bdc8f4905e2d5a7d6684ea454f7fd045206 not found: ID does not exist" Nov 22 09:33:36 crc kubenswrapper[4789]: I1122 09:33:36.362924 4789 scope.go:117] "RemoveContainer" containerID="da6d19b59583f367dcaee81085c2b5d2366752cf837f7c02be6908ba3188012c" Nov 22 09:33:36 crc kubenswrapper[4789]: E1122 09:33:36.363256 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da6d19b59583f367dcaee81085c2b5d2366752cf837f7c02be6908ba3188012c\": container with ID starting with da6d19b59583f367dcaee81085c2b5d2366752cf837f7c02be6908ba3188012c not found: ID does not exist" containerID="da6d19b59583f367dcaee81085c2b5d2366752cf837f7c02be6908ba3188012c" Nov 22 09:33:36 crc kubenswrapper[4789]: I1122 09:33:36.363293 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da6d19b59583f367dcaee81085c2b5d2366752cf837f7c02be6908ba3188012c"} err="failed to get container status \"da6d19b59583f367dcaee81085c2b5d2366752cf837f7c02be6908ba3188012c\": rpc error: code = NotFound desc = could not find container \"da6d19b59583f367dcaee81085c2b5d2366752cf837f7c02be6908ba3188012c\": container with ID starting with da6d19b59583f367dcaee81085c2b5d2366752cf837f7c02be6908ba3188012c not found: ID does not exist" Nov 22 09:33:37 crc kubenswrapper[4789]: I1122 09:33:37.975895 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5f84d005-b7bd-4062-997e-e30019d0233c" path="/var/lib/kubelet/pods/5f84d005-b7bd-4062-997e-e30019d0233c/volumes" Nov 22 09:33:46 crc kubenswrapper[4789]: I1122 09:33:46.965224 4789 scope.go:117] "RemoveContainer" containerID="a2b76708b26cd8ba98b6d3c6d8e8d18687de1639e9e73083d0f8694b338937ca" Nov 22 09:33:46 crc kubenswrapper[4789]: E1122 09:33:46.966009 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:34:00 crc kubenswrapper[4789]: I1122 09:34:00.964831 4789 scope.go:117] "RemoveContainer" containerID="a2b76708b26cd8ba98b6d3c6d8e8d18687de1639e9e73083d0f8694b338937ca" Nov 22 09:34:00 crc kubenswrapper[4789]: E1122 09:34:00.965470 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:34:14 crc kubenswrapper[4789]: I1122 09:34:14.965624 4789 scope.go:117] "RemoveContainer" containerID="a2b76708b26cd8ba98b6d3c6d8e8d18687de1639e9e73083d0f8694b338937ca" Nov 22 09:34:14 crc kubenswrapper[4789]: E1122 09:34:14.966600 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:34:25 crc kubenswrapper[4789]: I1122 09:34:25.965418 4789 scope.go:117] "RemoveContainer" containerID="a2b76708b26cd8ba98b6d3c6d8e8d18687de1639e9e73083d0f8694b338937ca" Nov 22 09:34:25 crc kubenswrapper[4789]: E1122 09:34:25.966935 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:34:39 crc kubenswrapper[4789]: I1122 09:34:39.965937 4789 scope.go:117] "RemoveContainer" containerID="a2b76708b26cd8ba98b6d3c6d8e8d18687de1639e9e73083d0f8694b338937ca" Nov 22 09:34:39 crc kubenswrapper[4789]: E1122 09:34:39.966937 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:34:52 crc kubenswrapper[4789]: I1122 09:34:52.964792 4789 scope.go:117] "RemoveContainer" containerID="a2b76708b26cd8ba98b6d3c6d8e8d18687de1639e9e73083d0f8694b338937ca" Nov 22 09:34:52 crc kubenswrapper[4789]: E1122 09:34:52.965578 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:35:04 crc kubenswrapper[4789]: I1122 09:35:04.965783 4789 scope.go:117] "RemoveContainer" containerID="a2b76708b26cd8ba98b6d3c6d8e8d18687de1639e9e73083d0f8694b338937ca" Nov 22 09:35:04 crc kubenswrapper[4789]: E1122 09:35:04.966900 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:35:16 crc kubenswrapper[4789]: I1122 09:35:16.967949 4789 scope.go:117] "RemoveContainer" containerID="a2b76708b26cd8ba98b6d3c6d8e8d18687de1639e9e73083d0f8694b338937ca" Nov 22 09:35:16 crc kubenswrapper[4789]: E1122 09:35:16.969040 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:35:27 crc kubenswrapper[4789]: I1122 09:35:27.977265 4789 scope.go:117] "RemoveContainer" containerID="a2b76708b26cd8ba98b6d3c6d8e8d18687de1639e9e73083d0f8694b338937ca" Nov 22 09:35:27 crc kubenswrapper[4789]: E1122 09:35:27.978032 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:35:40 crc kubenswrapper[4789]: I1122 09:35:40.964990 4789 scope.go:117] "RemoveContainer" containerID="a2b76708b26cd8ba98b6d3c6d8e8d18687de1639e9e73083d0f8694b338937ca" Nov 22 09:35:40 crc kubenswrapper[4789]: E1122 09:35:40.965849 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:35:53 crc kubenswrapper[4789]: I1122 09:35:53.966194 4789 scope.go:117] "RemoveContainer" containerID="a2b76708b26cd8ba98b6d3c6d8e8d18687de1639e9e73083d0f8694b338937ca" Nov 22 09:35:53 crc kubenswrapper[4789]: E1122 09:35:53.967569 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:36:07 crc kubenswrapper[4789]: I1122 09:36:07.975282 4789 scope.go:117] "RemoveContainer" containerID="a2b76708b26cd8ba98b6d3c6d8e8d18687de1639e9e73083d0f8694b338937ca" Nov 22 09:36:07 crc kubenswrapper[4789]: E1122 09:36:07.976023 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:36:20 crc kubenswrapper[4789]: I1122 09:36:20.965135 4789 scope.go:117] "RemoveContainer" containerID="a2b76708b26cd8ba98b6d3c6d8e8d18687de1639e9e73083d0f8694b338937ca" Nov 22 09:36:20 crc kubenswrapper[4789]: E1122 09:36:20.965817 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:36:35 crc kubenswrapper[4789]: I1122 09:36:35.965391 4789 scope.go:117] "RemoveContainer" containerID="a2b76708b26cd8ba98b6d3c6d8e8d18687de1639e9e73083d0f8694b338937ca" Nov 22 09:36:37 crc kubenswrapper[4789]: I1122 09:36:37.141243 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" event={"ID":"58aa8071-7ecc-4692-8789-2db04ec70647","Type":"ContainerStarted","Data":"26197f33465241392fa4dd1556c3f62bf6a1720878d713e1c6c6d88744c59810"} Nov 22 09:39:05 crc kubenswrapper[4789]: I1122 09:39:05.372218 4789 patch_prober.go:28] interesting pod/machine-config-daemon-pxvrm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 09:39:05 crc kubenswrapper[4789]: I1122 09:39:05.373180 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 09:39:25 crc kubenswrapper[4789]: I1122 09:39:25.664945 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-pclcd"] Nov 22 09:39:25 crc kubenswrapper[4789]: E1122 09:39:25.665920 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f84d005-b7bd-4062-997e-e30019d0233c" containerName="extract-content" Nov 22 09:39:25 crc kubenswrapper[4789]: I1122 09:39:25.665938 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f84d005-b7bd-4062-997e-e30019d0233c" containerName="extract-content" Nov 22 09:39:25 crc kubenswrapper[4789]: E1122 09:39:25.665977 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f84d005-b7bd-4062-997e-e30019d0233c" containerName="registry-server" Nov 22 09:39:25 crc kubenswrapper[4789]: I1122 09:39:25.665983 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f84d005-b7bd-4062-997e-e30019d0233c" containerName="registry-server" Nov 22 09:39:25 crc kubenswrapper[4789]: E1122 09:39:25.665999 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f84d005-b7bd-4062-997e-e30019d0233c" containerName="extract-utilities" Nov 22 09:39:25 crc kubenswrapper[4789]: I1122 09:39:25.666006 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f84d005-b7bd-4062-997e-e30019d0233c" containerName="extract-utilities" Nov 22 09:39:25 crc kubenswrapper[4789]: I1122 09:39:25.666232 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f84d005-b7bd-4062-997e-e30019d0233c" containerName="registry-server" Nov 22 09:39:25 crc kubenswrapper[4789]: I1122 09:39:25.667912 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pclcd" Nov 22 09:39:25 crc kubenswrapper[4789]: I1122 09:39:25.681957 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-pclcd"] Nov 22 09:39:25 crc kubenswrapper[4789]: I1122 09:39:25.775892 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7wspf\" (UniqueName: \"kubernetes.io/projected/5990ca71-48e3-464d-88f1-101649a06d49-kube-api-access-7wspf\") pod \"certified-operators-pclcd\" (UID: \"5990ca71-48e3-464d-88f1-101649a06d49\") " pod="openshift-marketplace/certified-operators-pclcd" Nov 22 09:39:25 crc kubenswrapper[4789]: I1122 09:39:25.775961 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5990ca71-48e3-464d-88f1-101649a06d49-utilities\") pod \"certified-operators-pclcd\" (UID: \"5990ca71-48e3-464d-88f1-101649a06d49\") " pod="openshift-marketplace/certified-operators-pclcd" Nov 22 09:39:25 crc kubenswrapper[4789]: I1122 09:39:25.776043 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5990ca71-48e3-464d-88f1-101649a06d49-catalog-content\") pod \"certified-operators-pclcd\" (UID: \"5990ca71-48e3-464d-88f1-101649a06d49\") " pod="openshift-marketplace/certified-operators-pclcd" Nov 22 09:39:25 crc kubenswrapper[4789]: I1122 09:39:25.877355 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5990ca71-48e3-464d-88f1-101649a06d49-catalog-content\") pod \"certified-operators-pclcd\" (UID: \"5990ca71-48e3-464d-88f1-101649a06d49\") " pod="openshift-marketplace/certified-operators-pclcd" Nov 22 09:39:25 crc kubenswrapper[4789]: I1122 09:39:25.877463 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7wspf\" (UniqueName: \"kubernetes.io/projected/5990ca71-48e3-464d-88f1-101649a06d49-kube-api-access-7wspf\") pod \"certified-operators-pclcd\" (UID: \"5990ca71-48e3-464d-88f1-101649a06d49\") " pod="openshift-marketplace/certified-operators-pclcd" Nov 22 09:39:25 crc kubenswrapper[4789]: I1122 09:39:25.877517 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5990ca71-48e3-464d-88f1-101649a06d49-utilities\") pod \"certified-operators-pclcd\" (UID: \"5990ca71-48e3-464d-88f1-101649a06d49\") " pod="openshift-marketplace/certified-operators-pclcd" Nov 22 09:39:25 crc kubenswrapper[4789]: I1122 09:39:25.878059 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5990ca71-48e3-464d-88f1-101649a06d49-catalog-content\") pod \"certified-operators-pclcd\" (UID: \"5990ca71-48e3-464d-88f1-101649a06d49\") " pod="openshift-marketplace/certified-operators-pclcd" Nov 22 09:39:25 crc kubenswrapper[4789]: I1122 09:39:25.878114 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5990ca71-48e3-464d-88f1-101649a06d49-utilities\") pod \"certified-operators-pclcd\" (UID: \"5990ca71-48e3-464d-88f1-101649a06d49\") " pod="openshift-marketplace/certified-operators-pclcd" Nov 22 09:39:25 crc kubenswrapper[4789]: I1122 09:39:25.904309 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7wspf\" (UniqueName: \"kubernetes.io/projected/5990ca71-48e3-464d-88f1-101649a06d49-kube-api-access-7wspf\") pod \"certified-operators-pclcd\" (UID: \"5990ca71-48e3-464d-88f1-101649a06d49\") " pod="openshift-marketplace/certified-operators-pclcd" Nov 22 09:39:25 crc kubenswrapper[4789]: I1122 09:39:25.987150 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pclcd" Nov 22 09:39:26 crc kubenswrapper[4789]: I1122 09:39:26.608085 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-pclcd"] Nov 22 09:39:26 crc kubenswrapper[4789]: I1122 09:39:26.818715 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pclcd" event={"ID":"5990ca71-48e3-464d-88f1-101649a06d49","Type":"ContainerStarted","Data":"eda6119524011443d1394e810ae5d635967fb43a731801392e2ec6bf66786ecb"} Nov 22 09:39:27 crc kubenswrapper[4789]: I1122 09:39:27.832663 4789 generic.go:334] "Generic (PLEG): container finished" podID="5990ca71-48e3-464d-88f1-101649a06d49" containerID="50fd19daf8be24fd648f9b6e3263b1e6783735c777f049a0059eb7bad815ccb9" exitCode=0 Nov 22 09:39:27 crc kubenswrapper[4789]: I1122 09:39:27.833094 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pclcd" event={"ID":"5990ca71-48e3-464d-88f1-101649a06d49","Type":"ContainerDied","Data":"50fd19daf8be24fd648f9b6e3263b1e6783735c777f049a0059eb7bad815ccb9"} Nov 22 09:39:27 crc kubenswrapper[4789]: I1122 09:39:27.835551 4789 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 22 09:39:31 crc kubenswrapper[4789]: I1122 09:39:31.870388 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pclcd" event={"ID":"5990ca71-48e3-464d-88f1-101649a06d49","Type":"ContainerStarted","Data":"ab1c0d51a1b58f81de797bec2ec2edfdffe0c60bce72c0541974e79ed7522707"} Nov 22 09:39:34 crc kubenswrapper[4789]: I1122 09:39:34.901408 4789 generic.go:334] "Generic (PLEG): container finished" podID="5990ca71-48e3-464d-88f1-101649a06d49" containerID="ab1c0d51a1b58f81de797bec2ec2edfdffe0c60bce72c0541974e79ed7522707" exitCode=0 Nov 22 09:39:34 crc kubenswrapper[4789]: I1122 09:39:34.901517 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pclcd" event={"ID":"5990ca71-48e3-464d-88f1-101649a06d49","Type":"ContainerDied","Data":"ab1c0d51a1b58f81de797bec2ec2edfdffe0c60bce72c0541974e79ed7522707"} Nov 22 09:39:35 crc kubenswrapper[4789]: I1122 09:39:35.372495 4789 patch_prober.go:28] interesting pod/machine-config-daemon-pxvrm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 09:39:35 crc kubenswrapper[4789]: I1122 09:39:35.372565 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 09:39:35 crc kubenswrapper[4789]: I1122 09:39:35.401892 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-sdt4b"] Nov 22 09:39:35 crc kubenswrapper[4789]: I1122 09:39:35.405115 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sdt4b" Nov 22 09:39:35 crc kubenswrapper[4789]: I1122 09:39:35.425924 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-sdt4b"] Nov 22 09:39:35 crc kubenswrapper[4789]: I1122 09:39:35.485152 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8709d41d-a670-4827-b0d1-bbe20e10a3fc-utilities\") pod \"redhat-operators-sdt4b\" (UID: \"8709d41d-a670-4827-b0d1-bbe20e10a3fc\") " pod="openshift-marketplace/redhat-operators-sdt4b" Nov 22 09:39:35 crc kubenswrapper[4789]: I1122 09:39:35.485441 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6f8wc\" (UniqueName: \"kubernetes.io/projected/8709d41d-a670-4827-b0d1-bbe20e10a3fc-kube-api-access-6f8wc\") pod \"redhat-operators-sdt4b\" (UID: \"8709d41d-a670-4827-b0d1-bbe20e10a3fc\") " pod="openshift-marketplace/redhat-operators-sdt4b" Nov 22 09:39:35 crc kubenswrapper[4789]: I1122 09:39:35.485462 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8709d41d-a670-4827-b0d1-bbe20e10a3fc-catalog-content\") pod \"redhat-operators-sdt4b\" (UID: \"8709d41d-a670-4827-b0d1-bbe20e10a3fc\") " pod="openshift-marketplace/redhat-operators-sdt4b" Nov 22 09:39:35 crc kubenswrapper[4789]: I1122 09:39:35.586809 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8709d41d-a670-4827-b0d1-bbe20e10a3fc-utilities\") pod \"redhat-operators-sdt4b\" (UID: \"8709d41d-a670-4827-b0d1-bbe20e10a3fc\") " pod="openshift-marketplace/redhat-operators-sdt4b" Nov 22 09:39:35 crc kubenswrapper[4789]: I1122 09:39:35.586861 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6f8wc\" (UniqueName: \"kubernetes.io/projected/8709d41d-a670-4827-b0d1-bbe20e10a3fc-kube-api-access-6f8wc\") pod \"redhat-operators-sdt4b\" (UID: \"8709d41d-a670-4827-b0d1-bbe20e10a3fc\") " pod="openshift-marketplace/redhat-operators-sdt4b" Nov 22 09:39:35 crc kubenswrapper[4789]: I1122 09:39:35.586880 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8709d41d-a670-4827-b0d1-bbe20e10a3fc-catalog-content\") pod \"redhat-operators-sdt4b\" (UID: \"8709d41d-a670-4827-b0d1-bbe20e10a3fc\") " pod="openshift-marketplace/redhat-operators-sdt4b" Nov 22 09:39:35 crc kubenswrapper[4789]: I1122 09:39:35.587446 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8709d41d-a670-4827-b0d1-bbe20e10a3fc-utilities\") pod \"redhat-operators-sdt4b\" (UID: \"8709d41d-a670-4827-b0d1-bbe20e10a3fc\") " pod="openshift-marketplace/redhat-operators-sdt4b" Nov 22 09:39:35 crc kubenswrapper[4789]: I1122 09:39:35.587463 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8709d41d-a670-4827-b0d1-bbe20e10a3fc-catalog-content\") pod \"redhat-operators-sdt4b\" (UID: \"8709d41d-a670-4827-b0d1-bbe20e10a3fc\") " pod="openshift-marketplace/redhat-operators-sdt4b" Nov 22 09:39:35 crc kubenswrapper[4789]: I1122 09:39:35.630556 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6f8wc\" (UniqueName: \"kubernetes.io/projected/8709d41d-a670-4827-b0d1-bbe20e10a3fc-kube-api-access-6f8wc\") pod \"redhat-operators-sdt4b\" (UID: \"8709d41d-a670-4827-b0d1-bbe20e10a3fc\") " pod="openshift-marketplace/redhat-operators-sdt4b" Nov 22 09:39:35 crc kubenswrapper[4789]: I1122 09:39:35.765166 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sdt4b" Nov 22 09:39:35 crc kubenswrapper[4789]: I1122 09:39:35.919516 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pclcd" event={"ID":"5990ca71-48e3-464d-88f1-101649a06d49","Type":"ContainerStarted","Data":"15f783a60175f794ddeaeb9a79380c30108df207acc31260584132fca568b0b3"} Nov 22 09:39:35 crc kubenswrapper[4789]: I1122 09:39:35.949675 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-pclcd" podStartSLOduration=3.402363946 podStartE2EDuration="10.949659159s" podCreationTimestamp="2025-11-22 09:39:25 +0000 UTC" firstStartedPulling="2025-11-22 09:39:27.835273035 +0000 UTC m=+6182.069673308" lastFinishedPulling="2025-11-22 09:39:35.382568238 +0000 UTC m=+6189.616968521" observedRunningTime="2025-11-22 09:39:35.942306272 +0000 UTC m=+6190.176706565" watchObservedRunningTime="2025-11-22 09:39:35.949659159 +0000 UTC m=+6190.184059432" Nov 22 09:39:35 crc kubenswrapper[4789]: I1122 09:39:35.987581 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-pclcd" Nov 22 09:39:35 crc kubenswrapper[4789]: I1122 09:39:35.987788 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-pclcd" Nov 22 09:39:36 crc kubenswrapper[4789]: I1122 09:39:36.221107 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-sdt4b"] Nov 22 09:39:36 crc kubenswrapper[4789]: I1122 09:39:36.933128 4789 generic.go:334] "Generic (PLEG): container finished" podID="8709d41d-a670-4827-b0d1-bbe20e10a3fc" containerID="860f4201ccc9413f644816910aee66dd12cfa561a9a2f750d0a03fa7800bd46a" exitCode=0 Nov 22 09:39:36 crc kubenswrapper[4789]: I1122 09:39:36.933210 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sdt4b" event={"ID":"8709d41d-a670-4827-b0d1-bbe20e10a3fc","Type":"ContainerDied","Data":"860f4201ccc9413f644816910aee66dd12cfa561a9a2f750d0a03fa7800bd46a"} Nov 22 09:39:36 crc kubenswrapper[4789]: I1122 09:39:36.933675 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sdt4b" event={"ID":"8709d41d-a670-4827-b0d1-bbe20e10a3fc","Type":"ContainerStarted","Data":"9df5d44969215388f6c0381b4e63d1f206ea767fc6d4c2c04b8c266e670ba66b"} Nov 22 09:39:37 crc kubenswrapper[4789]: I1122 09:39:37.045103 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-pclcd" podUID="5990ca71-48e3-464d-88f1-101649a06d49" containerName="registry-server" probeResult="failure" output=< Nov 22 09:39:37 crc kubenswrapper[4789]: timeout: failed to connect service ":50051" within 1s Nov 22 09:39:37 crc kubenswrapper[4789]: > Nov 22 09:39:37 crc kubenswrapper[4789]: I1122 09:39:37.945135 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sdt4b" event={"ID":"8709d41d-a670-4827-b0d1-bbe20e10a3fc","Type":"ContainerStarted","Data":"4ed1d27ce9f7de3dfd91f7cfa6fca1abdedf493da6a76009db12e8cb69c16540"} Nov 22 09:39:45 crc kubenswrapper[4789]: I1122 09:39:45.036125 4789 generic.go:334] "Generic (PLEG): container finished" podID="8709d41d-a670-4827-b0d1-bbe20e10a3fc" containerID="4ed1d27ce9f7de3dfd91f7cfa6fca1abdedf493da6a76009db12e8cb69c16540" exitCode=0 Nov 22 09:39:45 crc kubenswrapper[4789]: I1122 09:39:45.036191 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sdt4b" event={"ID":"8709d41d-a670-4827-b0d1-bbe20e10a3fc","Type":"ContainerDied","Data":"4ed1d27ce9f7de3dfd91f7cfa6fca1abdedf493da6a76009db12e8cb69c16540"} Nov 22 09:39:47 crc kubenswrapper[4789]: I1122 09:39:47.049373 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-pclcd" podUID="5990ca71-48e3-464d-88f1-101649a06d49" containerName="registry-server" probeResult="failure" output=< Nov 22 09:39:47 crc kubenswrapper[4789]: timeout: failed to connect service ":50051" within 1s Nov 22 09:39:47 crc kubenswrapper[4789]: > Nov 22 09:39:47 crc kubenswrapper[4789]: I1122 09:39:47.059854 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sdt4b" event={"ID":"8709d41d-a670-4827-b0d1-bbe20e10a3fc","Type":"ContainerStarted","Data":"ae96530c553cd5343a6ec13c544566ed3de6894c0e1b29c36c390f2e684cc047"} Nov 22 09:39:47 crc kubenswrapper[4789]: I1122 09:39:47.076390 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-sdt4b" podStartSLOduration=3.258220027 podStartE2EDuration="12.076372318s" podCreationTimestamp="2025-11-22 09:39:35 +0000 UTC" firstStartedPulling="2025-11-22 09:39:36.935086821 +0000 UTC m=+6191.169487094" lastFinishedPulling="2025-11-22 09:39:45.753239072 +0000 UTC m=+6199.987639385" observedRunningTime="2025-11-22 09:39:47.074193469 +0000 UTC m=+6201.308593742" watchObservedRunningTime="2025-11-22 09:39:47.076372318 +0000 UTC m=+6201.310772591" Nov 22 09:39:55 crc kubenswrapper[4789]: I1122 09:39:55.766044 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-sdt4b" Nov 22 09:39:55 crc kubenswrapper[4789]: I1122 09:39:55.767068 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-sdt4b" Nov 22 09:39:56 crc kubenswrapper[4789]: I1122 09:39:56.044456 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-pclcd" Nov 22 09:39:56 crc kubenswrapper[4789]: I1122 09:39:56.108779 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-pclcd" Nov 22 09:39:56 crc kubenswrapper[4789]: I1122 09:39:56.829989 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-sdt4b" podUID="8709d41d-a670-4827-b0d1-bbe20e10a3fc" containerName="registry-server" probeResult="failure" output=< Nov 22 09:39:56 crc kubenswrapper[4789]: timeout: failed to connect service ":50051" within 1s Nov 22 09:39:56 crc kubenswrapper[4789]: > Nov 22 09:39:56 crc kubenswrapper[4789]: I1122 09:39:56.863860 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-pclcd"] Nov 22 09:39:57 crc kubenswrapper[4789]: I1122 09:39:57.148323 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-pclcd" podUID="5990ca71-48e3-464d-88f1-101649a06d49" containerName="registry-server" containerID="cri-o://15f783a60175f794ddeaeb9a79380c30108df207acc31260584132fca568b0b3" gracePeriod=2 Nov 22 09:39:57 crc kubenswrapper[4789]: I1122 09:39:57.714093 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pclcd" Nov 22 09:39:57 crc kubenswrapper[4789]: I1122 09:39:57.867624 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5990ca71-48e3-464d-88f1-101649a06d49-utilities\") pod \"5990ca71-48e3-464d-88f1-101649a06d49\" (UID: \"5990ca71-48e3-464d-88f1-101649a06d49\") " Nov 22 09:39:57 crc kubenswrapper[4789]: I1122 09:39:57.867712 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7wspf\" (UniqueName: \"kubernetes.io/projected/5990ca71-48e3-464d-88f1-101649a06d49-kube-api-access-7wspf\") pod \"5990ca71-48e3-464d-88f1-101649a06d49\" (UID: \"5990ca71-48e3-464d-88f1-101649a06d49\") " Nov 22 09:39:57 crc kubenswrapper[4789]: I1122 09:39:57.867990 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5990ca71-48e3-464d-88f1-101649a06d49-catalog-content\") pod \"5990ca71-48e3-464d-88f1-101649a06d49\" (UID: \"5990ca71-48e3-464d-88f1-101649a06d49\") " Nov 22 09:39:57 crc kubenswrapper[4789]: I1122 09:39:57.869028 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5990ca71-48e3-464d-88f1-101649a06d49-utilities" (OuterVolumeSpecName: "utilities") pod "5990ca71-48e3-464d-88f1-101649a06d49" (UID: "5990ca71-48e3-464d-88f1-101649a06d49"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 09:39:57 crc kubenswrapper[4789]: I1122 09:39:57.890534 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5990ca71-48e3-464d-88f1-101649a06d49-kube-api-access-7wspf" (OuterVolumeSpecName: "kube-api-access-7wspf") pod "5990ca71-48e3-464d-88f1-101649a06d49" (UID: "5990ca71-48e3-464d-88f1-101649a06d49"). InnerVolumeSpecName "kube-api-access-7wspf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 09:39:57 crc kubenswrapper[4789]: I1122 09:39:57.918573 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5990ca71-48e3-464d-88f1-101649a06d49-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5990ca71-48e3-464d-88f1-101649a06d49" (UID: "5990ca71-48e3-464d-88f1-101649a06d49"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 09:39:57 crc kubenswrapper[4789]: I1122 09:39:57.971054 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5990ca71-48e3-464d-88f1-101649a06d49-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 09:39:57 crc kubenswrapper[4789]: I1122 09:39:57.971375 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7wspf\" (UniqueName: \"kubernetes.io/projected/5990ca71-48e3-464d-88f1-101649a06d49-kube-api-access-7wspf\") on node \"crc\" DevicePath \"\"" Nov 22 09:39:57 crc kubenswrapper[4789]: I1122 09:39:57.971454 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5990ca71-48e3-464d-88f1-101649a06d49-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 09:39:58 crc kubenswrapper[4789]: I1122 09:39:58.158414 4789 generic.go:334] "Generic (PLEG): container finished" podID="5990ca71-48e3-464d-88f1-101649a06d49" containerID="15f783a60175f794ddeaeb9a79380c30108df207acc31260584132fca568b0b3" exitCode=0 Nov 22 09:39:58 crc kubenswrapper[4789]: I1122 09:39:58.158463 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pclcd" Nov 22 09:39:58 crc kubenswrapper[4789]: I1122 09:39:58.158466 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pclcd" event={"ID":"5990ca71-48e3-464d-88f1-101649a06d49","Type":"ContainerDied","Data":"15f783a60175f794ddeaeb9a79380c30108df207acc31260584132fca568b0b3"} Nov 22 09:39:58 crc kubenswrapper[4789]: I1122 09:39:58.158494 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pclcd" event={"ID":"5990ca71-48e3-464d-88f1-101649a06d49","Type":"ContainerDied","Data":"eda6119524011443d1394e810ae5d635967fb43a731801392e2ec6bf66786ecb"} Nov 22 09:39:58 crc kubenswrapper[4789]: I1122 09:39:58.158513 4789 scope.go:117] "RemoveContainer" containerID="15f783a60175f794ddeaeb9a79380c30108df207acc31260584132fca568b0b3" Nov 22 09:39:58 crc kubenswrapper[4789]: I1122 09:39:58.180840 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-pclcd"] Nov 22 09:39:58 crc kubenswrapper[4789]: I1122 09:39:58.182327 4789 scope.go:117] "RemoveContainer" containerID="ab1c0d51a1b58f81de797bec2ec2edfdffe0c60bce72c0541974e79ed7522707" Nov 22 09:39:58 crc kubenswrapper[4789]: I1122 09:39:58.190286 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-pclcd"] Nov 22 09:39:58 crc kubenswrapper[4789]: I1122 09:39:58.202655 4789 scope.go:117] "RemoveContainer" containerID="50fd19daf8be24fd648f9b6e3263b1e6783735c777f049a0059eb7bad815ccb9" Nov 22 09:39:58 crc kubenswrapper[4789]: I1122 09:39:58.249140 4789 scope.go:117] "RemoveContainer" containerID="15f783a60175f794ddeaeb9a79380c30108df207acc31260584132fca568b0b3" Nov 22 09:39:58 crc kubenswrapper[4789]: E1122 09:39:58.250042 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"15f783a60175f794ddeaeb9a79380c30108df207acc31260584132fca568b0b3\": container with ID starting with 15f783a60175f794ddeaeb9a79380c30108df207acc31260584132fca568b0b3 not found: ID does not exist" containerID="15f783a60175f794ddeaeb9a79380c30108df207acc31260584132fca568b0b3" Nov 22 09:39:58 crc kubenswrapper[4789]: I1122 09:39:58.250088 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"15f783a60175f794ddeaeb9a79380c30108df207acc31260584132fca568b0b3"} err="failed to get container status \"15f783a60175f794ddeaeb9a79380c30108df207acc31260584132fca568b0b3\": rpc error: code = NotFound desc = could not find container \"15f783a60175f794ddeaeb9a79380c30108df207acc31260584132fca568b0b3\": container with ID starting with 15f783a60175f794ddeaeb9a79380c30108df207acc31260584132fca568b0b3 not found: ID does not exist" Nov 22 09:39:58 crc kubenswrapper[4789]: I1122 09:39:58.250119 4789 scope.go:117] "RemoveContainer" containerID="ab1c0d51a1b58f81de797bec2ec2edfdffe0c60bce72c0541974e79ed7522707" Nov 22 09:39:58 crc kubenswrapper[4789]: E1122 09:39:58.250493 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ab1c0d51a1b58f81de797bec2ec2edfdffe0c60bce72c0541974e79ed7522707\": container with ID starting with ab1c0d51a1b58f81de797bec2ec2edfdffe0c60bce72c0541974e79ed7522707 not found: ID does not exist" containerID="ab1c0d51a1b58f81de797bec2ec2edfdffe0c60bce72c0541974e79ed7522707" Nov 22 09:39:58 crc kubenswrapper[4789]: I1122 09:39:58.250550 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ab1c0d51a1b58f81de797bec2ec2edfdffe0c60bce72c0541974e79ed7522707"} err="failed to get container status \"ab1c0d51a1b58f81de797bec2ec2edfdffe0c60bce72c0541974e79ed7522707\": rpc error: code = NotFound desc = could not find container \"ab1c0d51a1b58f81de797bec2ec2edfdffe0c60bce72c0541974e79ed7522707\": container with ID starting with ab1c0d51a1b58f81de797bec2ec2edfdffe0c60bce72c0541974e79ed7522707 not found: ID does not exist" Nov 22 09:39:58 crc kubenswrapper[4789]: I1122 09:39:58.250578 4789 scope.go:117] "RemoveContainer" containerID="50fd19daf8be24fd648f9b6e3263b1e6783735c777f049a0059eb7bad815ccb9" Nov 22 09:39:58 crc kubenswrapper[4789]: E1122 09:39:58.251172 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"50fd19daf8be24fd648f9b6e3263b1e6783735c777f049a0059eb7bad815ccb9\": container with ID starting with 50fd19daf8be24fd648f9b6e3263b1e6783735c777f049a0059eb7bad815ccb9 not found: ID does not exist" containerID="50fd19daf8be24fd648f9b6e3263b1e6783735c777f049a0059eb7bad815ccb9" Nov 22 09:39:58 crc kubenswrapper[4789]: I1122 09:39:58.251226 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"50fd19daf8be24fd648f9b6e3263b1e6783735c777f049a0059eb7bad815ccb9"} err="failed to get container status \"50fd19daf8be24fd648f9b6e3263b1e6783735c777f049a0059eb7bad815ccb9\": rpc error: code = NotFound desc = could not find container \"50fd19daf8be24fd648f9b6e3263b1e6783735c777f049a0059eb7bad815ccb9\": container with ID starting with 50fd19daf8be24fd648f9b6e3263b1e6783735c777f049a0059eb7bad815ccb9 not found: ID does not exist" Nov 22 09:39:59 crc kubenswrapper[4789]: I1122 09:39:59.982883 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5990ca71-48e3-464d-88f1-101649a06d49" path="/var/lib/kubelet/pods/5990ca71-48e3-464d-88f1-101649a06d49/volumes" Nov 22 09:40:05 crc kubenswrapper[4789]: I1122 09:40:05.372385 4789 patch_prober.go:28] interesting pod/machine-config-daemon-pxvrm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 09:40:05 crc kubenswrapper[4789]: I1122 09:40:05.373213 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 09:40:05 crc kubenswrapper[4789]: I1122 09:40:05.373279 4789 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" Nov 22 09:40:05 crc kubenswrapper[4789]: I1122 09:40:05.374501 4789 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"26197f33465241392fa4dd1556c3f62bf6a1720878d713e1c6c6d88744c59810"} pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 22 09:40:05 crc kubenswrapper[4789]: I1122 09:40:05.374639 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" containerID="cri-o://26197f33465241392fa4dd1556c3f62bf6a1720878d713e1c6c6d88744c59810" gracePeriod=600 Nov 22 09:40:06 crc kubenswrapper[4789]: I1122 09:40:06.288474 4789 generic.go:334] "Generic (PLEG): container finished" podID="58aa8071-7ecc-4692-8789-2db04ec70647" containerID="26197f33465241392fa4dd1556c3f62bf6a1720878d713e1c6c6d88744c59810" exitCode=0 Nov 22 09:40:06 crc kubenswrapper[4789]: I1122 09:40:06.288814 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" event={"ID":"58aa8071-7ecc-4692-8789-2db04ec70647","Type":"ContainerDied","Data":"26197f33465241392fa4dd1556c3f62bf6a1720878d713e1c6c6d88744c59810"} Nov 22 09:40:06 crc kubenswrapper[4789]: I1122 09:40:06.288851 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" event={"ID":"58aa8071-7ecc-4692-8789-2db04ec70647","Type":"ContainerStarted","Data":"4ce6f99fe0532d0774ed3a238dd72bf64bbc780ed6b623e2fc3116d00f8d5b2f"} Nov 22 09:40:06 crc kubenswrapper[4789]: I1122 09:40:06.288872 4789 scope.go:117] "RemoveContainer" containerID="a2b76708b26cd8ba98b6d3c6d8e8d18687de1639e9e73083d0f8694b338937ca" Nov 22 09:40:06 crc kubenswrapper[4789]: I1122 09:40:06.831590 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-sdt4b" podUID="8709d41d-a670-4827-b0d1-bbe20e10a3fc" containerName="registry-server" probeResult="failure" output=< Nov 22 09:40:06 crc kubenswrapper[4789]: timeout: failed to connect service ":50051" within 1s Nov 22 09:40:06 crc kubenswrapper[4789]: > Nov 22 09:40:16 crc kubenswrapper[4789]: I1122 09:40:16.806410 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-sdt4b" podUID="8709d41d-a670-4827-b0d1-bbe20e10a3fc" containerName="registry-server" probeResult="failure" output=< Nov 22 09:40:16 crc kubenswrapper[4789]: timeout: failed to connect service ":50051" within 1s Nov 22 09:40:16 crc kubenswrapper[4789]: > Nov 22 09:40:25 crc kubenswrapper[4789]: I1122 09:40:25.838039 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-sdt4b" Nov 22 09:40:25 crc kubenswrapper[4789]: I1122 09:40:25.909331 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-sdt4b" Nov 22 09:40:26 crc kubenswrapper[4789]: I1122 09:40:26.898068 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-sdt4b"] Nov 22 09:40:27 crc kubenswrapper[4789]: I1122 09:40:27.514027 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-sdt4b" podUID="8709d41d-a670-4827-b0d1-bbe20e10a3fc" containerName="registry-server" containerID="cri-o://ae96530c553cd5343a6ec13c544566ed3de6894c0e1b29c36c390f2e684cc047" gracePeriod=2 Nov 22 09:40:28 crc kubenswrapper[4789]: I1122 09:40:28.200583 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sdt4b" Nov 22 09:40:28 crc kubenswrapper[4789]: I1122 09:40:28.355709 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6f8wc\" (UniqueName: \"kubernetes.io/projected/8709d41d-a670-4827-b0d1-bbe20e10a3fc-kube-api-access-6f8wc\") pod \"8709d41d-a670-4827-b0d1-bbe20e10a3fc\" (UID: \"8709d41d-a670-4827-b0d1-bbe20e10a3fc\") " Nov 22 09:40:28 crc kubenswrapper[4789]: I1122 09:40:28.356618 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8709d41d-a670-4827-b0d1-bbe20e10a3fc-utilities\") pod \"8709d41d-a670-4827-b0d1-bbe20e10a3fc\" (UID: \"8709d41d-a670-4827-b0d1-bbe20e10a3fc\") " Nov 22 09:40:28 crc kubenswrapper[4789]: I1122 09:40:28.356993 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8709d41d-a670-4827-b0d1-bbe20e10a3fc-catalog-content\") pod \"8709d41d-a670-4827-b0d1-bbe20e10a3fc\" (UID: \"8709d41d-a670-4827-b0d1-bbe20e10a3fc\") " Nov 22 09:40:28 crc kubenswrapper[4789]: I1122 09:40:28.357377 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8709d41d-a670-4827-b0d1-bbe20e10a3fc-utilities" (OuterVolumeSpecName: "utilities") pod "8709d41d-a670-4827-b0d1-bbe20e10a3fc" (UID: "8709d41d-a670-4827-b0d1-bbe20e10a3fc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 09:40:28 crc kubenswrapper[4789]: I1122 09:40:28.358361 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8709d41d-a670-4827-b0d1-bbe20e10a3fc-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 09:40:28 crc kubenswrapper[4789]: I1122 09:40:28.361366 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8709d41d-a670-4827-b0d1-bbe20e10a3fc-kube-api-access-6f8wc" (OuterVolumeSpecName: "kube-api-access-6f8wc") pod "8709d41d-a670-4827-b0d1-bbe20e10a3fc" (UID: "8709d41d-a670-4827-b0d1-bbe20e10a3fc"). InnerVolumeSpecName "kube-api-access-6f8wc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 09:40:28 crc kubenswrapper[4789]: I1122 09:40:28.452901 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8709d41d-a670-4827-b0d1-bbe20e10a3fc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8709d41d-a670-4827-b0d1-bbe20e10a3fc" (UID: "8709d41d-a670-4827-b0d1-bbe20e10a3fc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 09:40:28 crc kubenswrapper[4789]: I1122 09:40:28.460511 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8709d41d-a670-4827-b0d1-bbe20e10a3fc-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 09:40:28 crc kubenswrapper[4789]: I1122 09:40:28.460539 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6f8wc\" (UniqueName: \"kubernetes.io/projected/8709d41d-a670-4827-b0d1-bbe20e10a3fc-kube-api-access-6f8wc\") on node \"crc\" DevicePath \"\"" Nov 22 09:40:28 crc kubenswrapper[4789]: I1122 09:40:28.530394 4789 generic.go:334] "Generic (PLEG): container finished" podID="8709d41d-a670-4827-b0d1-bbe20e10a3fc" containerID="ae96530c553cd5343a6ec13c544566ed3de6894c0e1b29c36c390f2e684cc047" exitCode=0 Nov 22 09:40:28 crc kubenswrapper[4789]: I1122 09:40:28.530437 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sdt4b" event={"ID":"8709d41d-a670-4827-b0d1-bbe20e10a3fc","Type":"ContainerDied","Data":"ae96530c553cd5343a6ec13c544566ed3de6894c0e1b29c36c390f2e684cc047"} Nov 22 09:40:28 crc kubenswrapper[4789]: I1122 09:40:28.530461 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sdt4b" event={"ID":"8709d41d-a670-4827-b0d1-bbe20e10a3fc","Type":"ContainerDied","Data":"9df5d44969215388f6c0381b4e63d1f206ea767fc6d4c2c04b8c266e670ba66b"} Nov 22 09:40:28 crc kubenswrapper[4789]: I1122 09:40:28.530477 4789 scope.go:117] "RemoveContainer" containerID="ae96530c553cd5343a6ec13c544566ed3de6894c0e1b29c36c390f2e684cc047" Nov 22 09:40:28 crc kubenswrapper[4789]: I1122 09:40:28.530594 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sdt4b" Nov 22 09:40:28 crc kubenswrapper[4789]: I1122 09:40:28.567594 4789 scope.go:117] "RemoveContainer" containerID="4ed1d27ce9f7de3dfd91f7cfa6fca1abdedf493da6a76009db12e8cb69c16540" Nov 22 09:40:28 crc kubenswrapper[4789]: I1122 09:40:28.595204 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-sdt4b"] Nov 22 09:40:28 crc kubenswrapper[4789]: I1122 09:40:28.607951 4789 scope.go:117] "RemoveContainer" containerID="860f4201ccc9413f644816910aee66dd12cfa561a9a2f750d0a03fa7800bd46a" Nov 22 09:40:28 crc kubenswrapper[4789]: I1122 09:40:28.621140 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-sdt4b"] Nov 22 09:40:28 crc kubenswrapper[4789]: I1122 09:40:28.658056 4789 scope.go:117] "RemoveContainer" containerID="ae96530c553cd5343a6ec13c544566ed3de6894c0e1b29c36c390f2e684cc047" Nov 22 09:40:28 crc kubenswrapper[4789]: E1122 09:40:28.658488 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ae96530c553cd5343a6ec13c544566ed3de6894c0e1b29c36c390f2e684cc047\": container with ID starting with ae96530c553cd5343a6ec13c544566ed3de6894c0e1b29c36c390f2e684cc047 not found: ID does not exist" containerID="ae96530c553cd5343a6ec13c544566ed3de6894c0e1b29c36c390f2e684cc047" Nov 22 09:40:28 crc kubenswrapper[4789]: I1122 09:40:28.658535 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae96530c553cd5343a6ec13c544566ed3de6894c0e1b29c36c390f2e684cc047"} err="failed to get container status \"ae96530c553cd5343a6ec13c544566ed3de6894c0e1b29c36c390f2e684cc047\": rpc error: code = NotFound desc = could not find container \"ae96530c553cd5343a6ec13c544566ed3de6894c0e1b29c36c390f2e684cc047\": container with ID starting with ae96530c553cd5343a6ec13c544566ed3de6894c0e1b29c36c390f2e684cc047 not found: ID does not exist" Nov 22 09:40:28 crc kubenswrapper[4789]: I1122 09:40:28.658582 4789 scope.go:117] "RemoveContainer" containerID="4ed1d27ce9f7de3dfd91f7cfa6fca1abdedf493da6a76009db12e8cb69c16540" Nov 22 09:40:28 crc kubenswrapper[4789]: E1122 09:40:28.658925 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4ed1d27ce9f7de3dfd91f7cfa6fca1abdedf493da6a76009db12e8cb69c16540\": container with ID starting with 4ed1d27ce9f7de3dfd91f7cfa6fca1abdedf493da6a76009db12e8cb69c16540 not found: ID does not exist" containerID="4ed1d27ce9f7de3dfd91f7cfa6fca1abdedf493da6a76009db12e8cb69c16540" Nov 22 09:40:28 crc kubenswrapper[4789]: I1122 09:40:28.658982 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ed1d27ce9f7de3dfd91f7cfa6fca1abdedf493da6a76009db12e8cb69c16540"} err="failed to get container status \"4ed1d27ce9f7de3dfd91f7cfa6fca1abdedf493da6a76009db12e8cb69c16540\": rpc error: code = NotFound desc = could not find container \"4ed1d27ce9f7de3dfd91f7cfa6fca1abdedf493da6a76009db12e8cb69c16540\": container with ID starting with 4ed1d27ce9f7de3dfd91f7cfa6fca1abdedf493da6a76009db12e8cb69c16540 not found: ID does not exist" Nov 22 09:40:28 crc kubenswrapper[4789]: I1122 09:40:28.659011 4789 scope.go:117] "RemoveContainer" containerID="860f4201ccc9413f644816910aee66dd12cfa561a9a2f750d0a03fa7800bd46a" Nov 22 09:40:28 crc kubenswrapper[4789]: E1122 09:40:28.659311 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"860f4201ccc9413f644816910aee66dd12cfa561a9a2f750d0a03fa7800bd46a\": container with ID starting with 860f4201ccc9413f644816910aee66dd12cfa561a9a2f750d0a03fa7800bd46a not found: ID does not exist" containerID="860f4201ccc9413f644816910aee66dd12cfa561a9a2f750d0a03fa7800bd46a" Nov 22 09:40:28 crc kubenswrapper[4789]: I1122 09:40:28.659352 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"860f4201ccc9413f644816910aee66dd12cfa561a9a2f750d0a03fa7800bd46a"} err="failed to get container status \"860f4201ccc9413f644816910aee66dd12cfa561a9a2f750d0a03fa7800bd46a\": rpc error: code = NotFound desc = could not find container \"860f4201ccc9413f644816910aee66dd12cfa561a9a2f750d0a03fa7800bd46a\": container with ID starting with 860f4201ccc9413f644816910aee66dd12cfa561a9a2f750d0a03fa7800bd46a not found: ID does not exist" Nov 22 09:40:29 crc kubenswrapper[4789]: I1122 09:40:29.981525 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8709d41d-a670-4827-b0d1-bbe20e10a3fc" path="/var/lib/kubelet/pods/8709d41d-a670-4827-b0d1-bbe20e10a3fc/volumes" Nov 22 09:42:05 crc kubenswrapper[4789]: I1122 09:42:05.371934 4789 patch_prober.go:28] interesting pod/machine-config-daemon-pxvrm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 09:42:05 crc kubenswrapper[4789]: I1122 09:42:05.372587 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 09:42:21 crc kubenswrapper[4789]: I1122 09:42:21.082360 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-ccfs5"] Nov 22 09:42:21 crc kubenswrapper[4789]: E1122 09:42:21.084146 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5990ca71-48e3-464d-88f1-101649a06d49" containerName="extract-utilities" Nov 22 09:42:21 crc kubenswrapper[4789]: I1122 09:42:21.084181 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="5990ca71-48e3-464d-88f1-101649a06d49" containerName="extract-utilities" Nov 22 09:42:21 crc kubenswrapper[4789]: E1122 09:42:21.084229 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8709d41d-a670-4827-b0d1-bbe20e10a3fc" containerName="extract-utilities" Nov 22 09:42:21 crc kubenswrapper[4789]: I1122 09:42:21.084248 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="8709d41d-a670-4827-b0d1-bbe20e10a3fc" containerName="extract-utilities" Nov 22 09:42:21 crc kubenswrapper[4789]: E1122 09:42:21.084288 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8709d41d-a670-4827-b0d1-bbe20e10a3fc" containerName="registry-server" Nov 22 09:42:21 crc kubenswrapper[4789]: I1122 09:42:21.084307 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="8709d41d-a670-4827-b0d1-bbe20e10a3fc" containerName="registry-server" Nov 22 09:42:21 crc kubenswrapper[4789]: E1122 09:42:21.084339 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8709d41d-a670-4827-b0d1-bbe20e10a3fc" containerName="extract-content" Nov 22 09:42:21 crc kubenswrapper[4789]: I1122 09:42:21.084360 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="8709d41d-a670-4827-b0d1-bbe20e10a3fc" containerName="extract-content" Nov 22 09:42:21 crc kubenswrapper[4789]: E1122 09:42:21.084410 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5990ca71-48e3-464d-88f1-101649a06d49" containerName="extract-content" Nov 22 09:42:21 crc kubenswrapper[4789]: I1122 09:42:21.084428 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="5990ca71-48e3-464d-88f1-101649a06d49" containerName="extract-content" Nov 22 09:42:21 crc kubenswrapper[4789]: E1122 09:42:21.084460 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5990ca71-48e3-464d-88f1-101649a06d49" containerName="registry-server" Nov 22 09:42:21 crc kubenswrapper[4789]: I1122 09:42:21.084477 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="5990ca71-48e3-464d-88f1-101649a06d49" containerName="registry-server" Nov 22 09:42:21 crc kubenswrapper[4789]: I1122 09:42:21.084927 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="5990ca71-48e3-464d-88f1-101649a06d49" containerName="registry-server" Nov 22 09:42:21 crc kubenswrapper[4789]: I1122 09:42:21.084960 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="8709d41d-a670-4827-b0d1-bbe20e10a3fc" containerName="registry-server" Nov 22 09:42:21 crc kubenswrapper[4789]: I1122 09:42:21.087101 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ccfs5" Nov 22 09:42:21 crc kubenswrapper[4789]: I1122 09:42:21.095187 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ccfs5"] Nov 22 09:42:21 crc kubenswrapper[4789]: I1122 09:42:21.226575 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34c15871-61bb-4607-bb17-a30c7e319c09-catalog-content\") pod \"community-operators-ccfs5\" (UID: \"34c15871-61bb-4607-bb17-a30c7e319c09\") " pod="openshift-marketplace/community-operators-ccfs5" Nov 22 09:42:21 crc kubenswrapper[4789]: I1122 09:42:21.226638 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n5547\" (UniqueName: \"kubernetes.io/projected/34c15871-61bb-4607-bb17-a30c7e319c09-kube-api-access-n5547\") pod \"community-operators-ccfs5\" (UID: \"34c15871-61bb-4607-bb17-a30c7e319c09\") " pod="openshift-marketplace/community-operators-ccfs5" Nov 22 09:42:21 crc kubenswrapper[4789]: I1122 09:42:21.227013 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34c15871-61bb-4607-bb17-a30c7e319c09-utilities\") pod \"community-operators-ccfs5\" (UID: \"34c15871-61bb-4607-bb17-a30c7e319c09\") " pod="openshift-marketplace/community-operators-ccfs5" Nov 22 09:42:21 crc kubenswrapper[4789]: I1122 09:42:21.329834 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n5547\" (UniqueName: \"kubernetes.io/projected/34c15871-61bb-4607-bb17-a30c7e319c09-kube-api-access-n5547\") pod \"community-operators-ccfs5\" (UID: \"34c15871-61bb-4607-bb17-a30c7e319c09\") " pod="openshift-marketplace/community-operators-ccfs5" Nov 22 09:42:21 crc kubenswrapper[4789]: I1122 09:42:21.329987 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34c15871-61bb-4607-bb17-a30c7e319c09-utilities\") pod \"community-operators-ccfs5\" (UID: \"34c15871-61bb-4607-bb17-a30c7e319c09\") " pod="openshift-marketplace/community-operators-ccfs5" Nov 22 09:42:21 crc kubenswrapper[4789]: I1122 09:42:21.330175 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34c15871-61bb-4607-bb17-a30c7e319c09-catalog-content\") pod \"community-operators-ccfs5\" (UID: \"34c15871-61bb-4607-bb17-a30c7e319c09\") " pod="openshift-marketplace/community-operators-ccfs5" Nov 22 09:42:21 crc kubenswrapper[4789]: I1122 09:42:21.330923 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34c15871-61bb-4607-bb17-a30c7e319c09-catalog-content\") pod \"community-operators-ccfs5\" (UID: \"34c15871-61bb-4607-bb17-a30c7e319c09\") " pod="openshift-marketplace/community-operators-ccfs5" Nov 22 09:42:21 crc kubenswrapper[4789]: I1122 09:42:21.331325 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34c15871-61bb-4607-bb17-a30c7e319c09-utilities\") pod \"community-operators-ccfs5\" (UID: \"34c15871-61bb-4607-bb17-a30c7e319c09\") " pod="openshift-marketplace/community-operators-ccfs5" Nov 22 09:42:21 crc kubenswrapper[4789]: I1122 09:42:21.355499 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n5547\" (UniqueName: \"kubernetes.io/projected/34c15871-61bb-4607-bb17-a30c7e319c09-kube-api-access-n5547\") pod \"community-operators-ccfs5\" (UID: \"34c15871-61bb-4607-bb17-a30c7e319c09\") " pod="openshift-marketplace/community-operators-ccfs5" Nov 22 09:42:21 crc kubenswrapper[4789]: I1122 09:42:21.414982 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ccfs5" Nov 22 09:42:21 crc kubenswrapper[4789]: I1122 09:42:21.919385 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ccfs5"] Nov 22 09:42:21 crc kubenswrapper[4789]: W1122 09:42:21.934608 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod34c15871_61bb_4607_bb17_a30c7e319c09.slice/crio-d346fbbce0841fb51d3b78d42906de4e2594337a7d965cc6b00bce717c9ef528 WatchSource:0}: Error finding container d346fbbce0841fb51d3b78d42906de4e2594337a7d965cc6b00bce717c9ef528: Status 404 returned error can't find the container with id d346fbbce0841fb51d3b78d42906de4e2594337a7d965cc6b00bce717c9ef528 Nov 22 09:42:22 crc kubenswrapper[4789]: I1122 09:42:22.552056 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ccfs5" event={"ID":"34c15871-61bb-4607-bb17-a30c7e319c09","Type":"ContainerStarted","Data":"2af3e191819576f81d8c755a64dc310445598be0fd1ef2a09928e9c946710086"} Nov 22 09:42:22 crc kubenswrapper[4789]: I1122 09:42:22.552387 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ccfs5" event={"ID":"34c15871-61bb-4607-bb17-a30c7e319c09","Type":"ContainerStarted","Data":"d346fbbce0841fb51d3b78d42906de4e2594337a7d965cc6b00bce717c9ef528"} Nov 22 09:42:23 crc kubenswrapper[4789]: I1122 09:42:23.568977 4789 generic.go:334] "Generic (PLEG): container finished" podID="34c15871-61bb-4607-bb17-a30c7e319c09" containerID="2af3e191819576f81d8c755a64dc310445598be0fd1ef2a09928e9c946710086" exitCode=0 Nov 22 09:42:23 crc kubenswrapper[4789]: I1122 09:42:23.569106 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ccfs5" event={"ID":"34c15871-61bb-4607-bb17-a30c7e319c09","Type":"ContainerDied","Data":"2af3e191819576f81d8c755a64dc310445598be0fd1ef2a09928e9c946710086"} Nov 22 09:42:32 crc kubenswrapper[4789]: I1122 09:42:32.664264 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ccfs5" event={"ID":"34c15871-61bb-4607-bb17-a30c7e319c09","Type":"ContainerStarted","Data":"8ea121f2d5144ed3229522ca0fd704a565990608076119a8c491ca06c54a7e85"} Nov 22 09:42:35 crc kubenswrapper[4789]: I1122 09:42:35.372375 4789 patch_prober.go:28] interesting pod/machine-config-daemon-pxvrm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 09:42:35 crc kubenswrapper[4789]: I1122 09:42:35.372972 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 09:42:38 crc kubenswrapper[4789]: I1122 09:42:38.726892 4789 generic.go:334] "Generic (PLEG): container finished" podID="34c15871-61bb-4607-bb17-a30c7e319c09" containerID="8ea121f2d5144ed3229522ca0fd704a565990608076119a8c491ca06c54a7e85" exitCode=0 Nov 22 09:42:38 crc kubenswrapper[4789]: I1122 09:42:38.726973 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ccfs5" event={"ID":"34c15871-61bb-4607-bb17-a30c7e319c09","Type":"ContainerDied","Data":"8ea121f2d5144ed3229522ca0fd704a565990608076119a8c491ca06c54a7e85"} Nov 22 09:42:40 crc kubenswrapper[4789]: I1122 09:42:40.747874 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ccfs5" event={"ID":"34c15871-61bb-4607-bb17-a30c7e319c09","Type":"ContainerStarted","Data":"d71b3c5da9f5ec2a056eb49a24452552190bb8a527d1d9002f9b436e39cd9156"} Nov 22 09:42:40 crc kubenswrapper[4789]: I1122 09:42:40.772391 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-ccfs5" podStartSLOduration=3.49776131 podStartE2EDuration="19.772375634s" podCreationTimestamp="2025-11-22 09:42:21 +0000 UTC" firstStartedPulling="2025-11-22 09:42:23.571746266 +0000 UTC m=+6357.806146579" lastFinishedPulling="2025-11-22 09:42:39.84636063 +0000 UTC m=+6374.080760903" observedRunningTime="2025-11-22 09:42:40.769060394 +0000 UTC m=+6375.003460667" watchObservedRunningTime="2025-11-22 09:42:40.772375634 +0000 UTC m=+6375.006775907" Nov 22 09:42:41 crc kubenswrapper[4789]: I1122 09:42:41.415433 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-ccfs5" Nov 22 09:42:41 crc kubenswrapper[4789]: I1122 09:42:41.415788 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-ccfs5" Nov 22 09:42:42 crc kubenswrapper[4789]: I1122 09:42:42.465250 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-ccfs5" podUID="34c15871-61bb-4607-bb17-a30c7e319c09" containerName="registry-server" probeResult="failure" output=< Nov 22 09:42:42 crc kubenswrapper[4789]: timeout: failed to connect service ":50051" within 1s Nov 22 09:42:42 crc kubenswrapper[4789]: > Nov 22 09:42:51 crc kubenswrapper[4789]: I1122 09:42:51.460809 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-ccfs5" Nov 22 09:42:51 crc kubenswrapper[4789]: I1122 09:42:51.519196 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-ccfs5" Nov 22 09:42:52 crc kubenswrapper[4789]: I1122 09:42:52.261542 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ccfs5"] Nov 22 09:42:52 crc kubenswrapper[4789]: I1122 09:42:52.856960 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-ccfs5" podUID="34c15871-61bb-4607-bb17-a30c7e319c09" containerName="registry-server" containerID="cri-o://d71b3c5da9f5ec2a056eb49a24452552190bb8a527d1d9002f9b436e39cd9156" gracePeriod=2 Nov 22 09:42:53 crc kubenswrapper[4789]: I1122 09:42:53.535206 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ccfs5" Nov 22 09:42:53 crc kubenswrapper[4789]: I1122 09:42:53.650336 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n5547\" (UniqueName: \"kubernetes.io/projected/34c15871-61bb-4607-bb17-a30c7e319c09-kube-api-access-n5547\") pod \"34c15871-61bb-4607-bb17-a30c7e319c09\" (UID: \"34c15871-61bb-4607-bb17-a30c7e319c09\") " Nov 22 09:42:53 crc kubenswrapper[4789]: I1122 09:42:53.650406 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34c15871-61bb-4607-bb17-a30c7e319c09-catalog-content\") pod \"34c15871-61bb-4607-bb17-a30c7e319c09\" (UID: \"34c15871-61bb-4607-bb17-a30c7e319c09\") " Nov 22 09:42:53 crc kubenswrapper[4789]: I1122 09:42:53.650497 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34c15871-61bb-4607-bb17-a30c7e319c09-utilities\") pod \"34c15871-61bb-4607-bb17-a30c7e319c09\" (UID: \"34c15871-61bb-4607-bb17-a30c7e319c09\") " Nov 22 09:42:53 crc kubenswrapper[4789]: I1122 09:42:53.651557 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/34c15871-61bb-4607-bb17-a30c7e319c09-utilities" (OuterVolumeSpecName: "utilities") pod "34c15871-61bb-4607-bb17-a30c7e319c09" (UID: "34c15871-61bb-4607-bb17-a30c7e319c09"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 09:42:53 crc kubenswrapper[4789]: I1122 09:42:53.656354 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34c15871-61bb-4607-bb17-a30c7e319c09-kube-api-access-n5547" (OuterVolumeSpecName: "kube-api-access-n5547") pod "34c15871-61bb-4607-bb17-a30c7e319c09" (UID: "34c15871-61bb-4607-bb17-a30c7e319c09"). InnerVolumeSpecName "kube-api-access-n5547". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 09:42:53 crc kubenswrapper[4789]: I1122 09:42:53.715737 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/34c15871-61bb-4607-bb17-a30c7e319c09-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "34c15871-61bb-4607-bb17-a30c7e319c09" (UID: "34c15871-61bb-4607-bb17-a30c7e319c09"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 09:42:53 crc kubenswrapper[4789]: I1122 09:42:53.752999 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34c15871-61bb-4607-bb17-a30c7e319c09-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 09:42:53 crc kubenswrapper[4789]: I1122 09:42:53.753031 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n5547\" (UniqueName: \"kubernetes.io/projected/34c15871-61bb-4607-bb17-a30c7e319c09-kube-api-access-n5547\") on node \"crc\" DevicePath \"\"" Nov 22 09:42:53 crc kubenswrapper[4789]: I1122 09:42:53.753042 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34c15871-61bb-4607-bb17-a30c7e319c09-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 09:42:53 crc kubenswrapper[4789]: I1122 09:42:53.868716 4789 generic.go:334] "Generic (PLEG): container finished" podID="34c15871-61bb-4607-bb17-a30c7e319c09" containerID="d71b3c5da9f5ec2a056eb49a24452552190bb8a527d1d9002f9b436e39cd9156" exitCode=0 Nov 22 09:42:53 crc kubenswrapper[4789]: I1122 09:42:53.868783 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ccfs5" event={"ID":"34c15871-61bb-4607-bb17-a30c7e319c09","Type":"ContainerDied","Data":"d71b3c5da9f5ec2a056eb49a24452552190bb8a527d1d9002f9b436e39cd9156"} Nov 22 09:42:53 crc kubenswrapper[4789]: I1122 09:42:53.868811 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ccfs5" event={"ID":"34c15871-61bb-4607-bb17-a30c7e319c09","Type":"ContainerDied","Data":"d346fbbce0841fb51d3b78d42906de4e2594337a7d965cc6b00bce717c9ef528"} Nov 22 09:42:53 crc kubenswrapper[4789]: I1122 09:42:53.868817 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ccfs5" Nov 22 09:42:53 crc kubenswrapper[4789]: I1122 09:42:53.868830 4789 scope.go:117] "RemoveContainer" containerID="d71b3c5da9f5ec2a056eb49a24452552190bb8a527d1d9002f9b436e39cd9156" Nov 22 09:42:53 crc kubenswrapper[4789]: I1122 09:42:53.903371 4789 scope.go:117] "RemoveContainer" containerID="8ea121f2d5144ed3229522ca0fd704a565990608076119a8c491ca06c54a7e85" Nov 22 09:42:53 crc kubenswrapper[4789]: I1122 09:42:53.920223 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ccfs5"] Nov 22 09:42:53 crc kubenswrapper[4789]: I1122 09:42:53.928842 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-ccfs5"] Nov 22 09:42:53 crc kubenswrapper[4789]: I1122 09:42:53.932975 4789 scope.go:117] "RemoveContainer" containerID="2af3e191819576f81d8c755a64dc310445598be0fd1ef2a09928e9c946710086" Nov 22 09:42:53 crc kubenswrapper[4789]: I1122 09:42:53.977744 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="34c15871-61bb-4607-bb17-a30c7e319c09" path="/var/lib/kubelet/pods/34c15871-61bb-4607-bb17-a30c7e319c09/volumes" Nov 22 09:42:53 crc kubenswrapper[4789]: I1122 09:42:53.986429 4789 scope.go:117] "RemoveContainer" containerID="d71b3c5da9f5ec2a056eb49a24452552190bb8a527d1d9002f9b436e39cd9156" Nov 22 09:42:53 crc kubenswrapper[4789]: E1122 09:42:53.986959 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d71b3c5da9f5ec2a056eb49a24452552190bb8a527d1d9002f9b436e39cd9156\": container with ID starting with d71b3c5da9f5ec2a056eb49a24452552190bb8a527d1d9002f9b436e39cd9156 not found: ID does not exist" containerID="d71b3c5da9f5ec2a056eb49a24452552190bb8a527d1d9002f9b436e39cd9156" Nov 22 09:42:53 crc kubenswrapper[4789]: I1122 09:42:53.987001 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d71b3c5da9f5ec2a056eb49a24452552190bb8a527d1d9002f9b436e39cd9156"} err="failed to get container status \"d71b3c5da9f5ec2a056eb49a24452552190bb8a527d1d9002f9b436e39cd9156\": rpc error: code = NotFound desc = could not find container \"d71b3c5da9f5ec2a056eb49a24452552190bb8a527d1d9002f9b436e39cd9156\": container with ID starting with d71b3c5da9f5ec2a056eb49a24452552190bb8a527d1d9002f9b436e39cd9156 not found: ID does not exist" Nov 22 09:42:53 crc kubenswrapper[4789]: I1122 09:42:53.987031 4789 scope.go:117] "RemoveContainer" containerID="8ea121f2d5144ed3229522ca0fd704a565990608076119a8c491ca06c54a7e85" Nov 22 09:42:53 crc kubenswrapper[4789]: E1122 09:42:53.987317 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8ea121f2d5144ed3229522ca0fd704a565990608076119a8c491ca06c54a7e85\": container with ID starting with 8ea121f2d5144ed3229522ca0fd704a565990608076119a8c491ca06c54a7e85 not found: ID does not exist" containerID="8ea121f2d5144ed3229522ca0fd704a565990608076119a8c491ca06c54a7e85" Nov 22 09:42:53 crc kubenswrapper[4789]: I1122 09:42:53.987351 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8ea121f2d5144ed3229522ca0fd704a565990608076119a8c491ca06c54a7e85"} err="failed to get container status \"8ea121f2d5144ed3229522ca0fd704a565990608076119a8c491ca06c54a7e85\": rpc error: code = NotFound desc = could not find container \"8ea121f2d5144ed3229522ca0fd704a565990608076119a8c491ca06c54a7e85\": container with ID starting with 8ea121f2d5144ed3229522ca0fd704a565990608076119a8c491ca06c54a7e85 not found: ID does not exist" Nov 22 09:42:53 crc kubenswrapper[4789]: I1122 09:42:53.987376 4789 scope.go:117] "RemoveContainer" containerID="2af3e191819576f81d8c755a64dc310445598be0fd1ef2a09928e9c946710086" Nov 22 09:42:53 crc kubenswrapper[4789]: E1122 09:42:53.987580 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2af3e191819576f81d8c755a64dc310445598be0fd1ef2a09928e9c946710086\": container with ID starting with 2af3e191819576f81d8c755a64dc310445598be0fd1ef2a09928e9c946710086 not found: ID does not exist" containerID="2af3e191819576f81d8c755a64dc310445598be0fd1ef2a09928e9c946710086" Nov 22 09:42:53 crc kubenswrapper[4789]: I1122 09:42:53.987609 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2af3e191819576f81d8c755a64dc310445598be0fd1ef2a09928e9c946710086"} err="failed to get container status \"2af3e191819576f81d8c755a64dc310445598be0fd1ef2a09928e9c946710086\": rpc error: code = NotFound desc = could not find container \"2af3e191819576f81d8c755a64dc310445598be0fd1ef2a09928e9c946710086\": container with ID starting with 2af3e191819576f81d8c755a64dc310445598be0fd1ef2a09928e9c946710086 not found: ID does not exist" Nov 22 09:43:05 crc kubenswrapper[4789]: I1122 09:43:05.372202 4789 patch_prober.go:28] interesting pod/machine-config-daemon-pxvrm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 09:43:05 crc kubenswrapper[4789]: I1122 09:43:05.374031 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 09:43:05 crc kubenswrapper[4789]: I1122 09:43:05.374128 4789 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" Nov 22 09:43:05 crc kubenswrapper[4789]: I1122 09:43:05.375318 4789 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4ce6f99fe0532d0774ed3a238dd72bf64bbc780ed6b623e2fc3116d00f8d5b2f"} pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 22 09:43:05 crc kubenswrapper[4789]: I1122 09:43:05.375455 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" containerID="cri-o://4ce6f99fe0532d0774ed3a238dd72bf64bbc780ed6b623e2fc3116d00f8d5b2f" gracePeriod=600 Nov 22 09:43:05 crc kubenswrapper[4789]: E1122 09:43:05.563959 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:43:05 crc kubenswrapper[4789]: I1122 09:43:05.974814 4789 generic.go:334] "Generic (PLEG): container finished" podID="58aa8071-7ecc-4692-8789-2db04ec70647" containerID="4ce6f99fe0532d0774ed3a238dd72bf64bbc780ed6b623e2fc3116d00f8d5b2f" exitCode=0 Nov 22 09:43:05 crc kubenswrapper[4789]: I1122 09:43:05.976549 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" event={"ID":"58aa8071-7ecc-4692-8789-2db04ec70647","Type":"ContainerDied","Data":"4ce6f99fe0532d0774ed3a238dd72bf64bbc780ed6b623e2fc3116d00f8d5b2f"} Nov 22 09:43:05 crc kubenswrapper[4789]: I1122 09:43:05.976604 4789 scope.go:117] "RemoveContainer" containerID="26197f33465241392fa4dd1556c3f62bf6a1720878d713e1c6c6d88744c59810" Nov 22 09:43:05 crc kubenswrapper[4789]: I1122 09:43:05.977291 4789 scope.go:117] "RemoveContainer" containerID="4ce6f99fe0532d0774ed3a238dd72bf64bbc780ed6b623e2fc3116d00f8d5b2f" Nov 22 09:43:05 crc kubenswrapper[4789]: E1122 09:43:05.977612 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:43:17 crc kubenswrapper[4789]: I1122 09:43:17.975933 4789 scope.go:117] "RemoveContainer" containerID="4ce6f99fe0532d0774ed3a238dd72bf64bbc780ed6b623e2fc3116d00f8d5b2f" Nov 22 09:43:17 crc kubenswrapper[4789]: E1122 09:43:17.976797 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:43:32 crc kubenswrapper[4789]: I1122 09:43:32.965567 4789 scope.go:117] "RemoveContainer" containerID="4ce6f99fe0532d0774ed3a238dd72bf64bbc780ed6b623e2fc3116d00f8d5b2f" Nov 22 09:43:32 crc kubenswrapper[4789]: E1122 09:43:32.966401 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:43:45 crc kubenswrapper[4789]: I1122 09:43:45.964672 4789 scope.go:117] "RemoveContainer" containerID="4ce6f99fe0532d0774ed3a238dd72bf64bbc780ed6b623e2fc3116d00f8d5b2f" Nov 22 09:43:45 crc kubenswrapper[4789]: E1122 09:43:45.965633 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:43:56 crc kubenswrapper[4789]: I1122 09:43:56.965235 4789 scope.go:117] "RemoveContainer" containerID="4ce6f99fe0532d0774ed3a238dd72bf64bbc780ed6b623e2fc3116d00f8d5b2f" Nov 22 09:43:56 crc kubenswrapper[4789]: E1122 09:43:56.966104 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:44:11 crc kubenswrapper[4789]: I1122 09:44:11.965974 4789 scope.go:117] "RemoveContainer" containerID="4ce6f99fe0532d0774ed3a238dd72bf64bbc780ed6b623e2fc3116d00f8d5b2f" Nov 22 09:44:11 crc kubenswrapper[4789]: E1122 09:44:11.966711 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:44:23 crc kubenswrapper[4789]: I1122 09:44:23.965839 4789 scope.go:117] "RemoveContainer" containerID="4ce6f99fe0532d0774ed3a238dd72bf64bbc780ed6b623e2fc3116d00f8d5b2f" Nov 22 09:44:23 crc kubenswrapper[4789]: E1122 09:44:23.966636 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:44:38 crc kubenswrapper[4789]: I1122 09:44:38.965231 4789 scope.go:117] "RemoveContainer" containerID="4ce6f99fe0532d0774ed3a238dd72bf64bbc780ed6b623e2fc3116d00f8d5b2f" Nov 22 09:44:38 crc kubenswrapper[4789]: E1122 09:44:38.966169 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:44:49 crc kubenswrapper[4789]: I1122 09:44:49.966142 4789 scope.go:117] "RemoveContainer" containerID="4ce6f99fe0532d0774ed3a238dd72bf64bbc780ed6b623e2fc3116d00f8d5b2f" Nov 22 09:44:49 crc kubenswrapper[4789]: E1122 09:44:49.967093 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:45:00 crc kubenswrapper[4789]: I1122 09:45:00.178194 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396745-q52jz"] Nov 22 09:45:00 crc kubenswrapper[4789]: E1122 09:45:00.179135 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34c15871-61bb-4607-bb17-a30c7e319c09" containerName="extract-content" Nov 22 09:45:00 crc kubenswrapper[4789]: I1122 09:45:00.179154 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="34c15871-61bb-4607-bb17-a30c7e319c09" containerName="extract-content" Nov 22 09:45:00 crc kubenswrapper[4789]: E1122 09:45:00.179179 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34c15871-61bb-4607-bb17-a30c7e319c09" containerName="extract-utilities" Nov 22 09:45:00 crc kubenswrapper[4789]: I1122 09:45:00.179187 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="34c15871-61bb-4607-bb17-a30c7e319c09" containerName="extract-utilities" Nov 22 09:45:00 crc kubenswrapper[4789]: E1122 09:45:00.179227 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34c15871-61bb-4607-bb17-a30c7e319c09" containerName="registry-server" Nov 22 09:45:00 crc kubenswrapper[4789]: I1122 09:45:00.179236 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="34c15871-61bb-4607-bb17-a30c7e319c09" containerName="registry-server" Nov 22 09:45:00 crc kubenswrapper[4789]: I1122 09:45:00.179435 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="34c15871-61bb-4607-bb17-a30c7e319c09" containerName="registry-server" Nov 22 09:45:00 crc kubenswrapper[4789]: I1122 09:45:00.180202 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396745-q52jz" Nov 22 09:45:00 crc kubenswrapper[4789]: I1122 09:45:00.182112 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 22 09:45:00 crc kubenswrapper[4789]: I1122 09:45:00.182586 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 22 09:45:00 crc kubenswrapper[4789]: I1122 09:45:00.189396 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396745-q52jz"] Nov 22 09:45:00 crc kubenswrapper[4789]: I1122 09:45:00.299519 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fxbjj\" (UniqueName: \"kubernetes.io/projected/adfe0251-8ae5-48bc-a0b2-23399ddaf4d8-kube-api-access-fxbjj\") pod \"collect-profiles-29396745-q52jz\" (UID: \"adfe0251-8ae5-48bc-a0b2-23399ddaf4d8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396745-q52jz" Nov 22 09:45:00 crc kubenswrapper[4789]: I1122 09:45:00.299575 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/adfe0251-8ae5-48bc-a0b2-23399ddaf4d8-secret-volume\") pod \"collect-profiles-29396745-q52jz\" (UID: \"adfe0251-8ae5-48bc-a0b2-23399ddaf4d8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396745-q52jz" Nov 22 09:45:00 crc kubenswrapper[4789]: I1122 09:45:00.299623 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/adfe0251-8ae5-48bc-a0b2-23399ddaf4d8-config-volume\") pod \"collect-profiles-29396745-q52jz\" (UID: \"adfe0251-8ae5-48bc-a0b2-23399ddaf4d8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396745-q52jz" Nov 22 09:45:00 crc kubenswrapper[4789]: I1122 09:45:00.400898 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fxbjj\" (UniqueName: \"kubernetes.io/projected/adfe0251-8ae5-48bc-a0b2-23399ddaf4d8-kube-api-access-fxbjj\") pod \"collect-profiles-29396745-q52jz\" (UID: \"adfe0251-8ae5-48bc-a0b2-23399ddaf4d8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396745-q52jz" Nov 22 09:45:00 crc kubenswrapper[4789]: I1122 09:45:00.400966 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/adfe0251-8ae5-48bc-a0b2-23399ddaf4d8-secret-volume\") pod \"collect-profiles-29396745-q52jz\" (UID: \"adfe0251-8ae5-48bc-a0b2-23399ddaf4d8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396745-q52jz" Nov 22 09:45:00 crc kubenswrapper[4789]: I1122 09:45:00.401032 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/adfe0251-8ae5-48bc-a0b2-23399ddaf4d8-config-volume\") pod \"collect-profiles-29396745-q52jz\" (UID: \"adfe0251-8ae5-48bc-a0b2-23399ddaf4d8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396745-q52jz" Nov 22 09:45:00 crc kubenswrapper[4789]: I1122 09:45:00.402618 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/adfe0251-8ae5-48bc-a0b2-23399ddaf4d8-config-volume\") pod \"collect-profiles-29396745-q52jz\" (UID: \"adfe0251-8ae5-48bc-a0b2-23399ddaf4d8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396745-q52jz" Nov 22 09:45:00 crc kubenswrapper[4789]: I1122 09:45:00.407066 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/adfe0251-8ae5-48bc-a0b2-23399ddaf4d8-secret-volume\") pod \"collect-profiles-29396745-q52jz\" (UID: \"adfe0251-8ae5-48bc-a0b2-23399ddaf4d8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396745-q52jz" Nov 22 09:45:00 crc kubenswrapper[4789]: I1122 09:45:00.417592 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fxbjj\" (UniqueName: \"kubernetes.io/projected/adfe0251-8ae5-48bc-a0b2-23399ddaf4d8-kube-api-access-fxbjj\") pod \"collect-profiles-29396745-q52jz\" (UID: \"adfe0251-8ae5-48bc-a0b2-23399ddaf4d8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396745-q52jz" Nov 22 09:45:00 crc kubenswrapper[4789]: I1122 09:45:00.501998 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396745-q52jz" Nov 22 09:45:00 crc kubenswrapper[4789]: I1122 09:45:00.997209 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396745-q52jz"] Nov 22 09:45:01 crc kubenswrapper[4789]: I1122 09:45:01.157924 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29396745-q52jz" event={"ID":"adfe0251-8ae5-48bc-a0b2-23399ddaf4d8","Type":"ContainerStarted","Data":"f1b374ffcfce15861e136d23cd8e0e42362ade6bfebbcc441e687cca1f988985"} Nov 22 09:45:02 crc kubenswrapper[4789]: I1122 09:45:02.181159 4789 generic.go:334] "Generic (PLEG): container finished" podID="adfe0251-8ae5-48bc-a0b2-23399ddaf4d8" containerID="7a17bd5bc847af1fd9e52cbfc735adb9c04cdb9695e08dd9e2e0b0eee77a160f" exitCode=0 Nov 22 09:45:02 crc kubenswrapper[4789]: I1122 09:45:02.181241 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29396745-q52jz" event={"ID":"adfe0251-8ae5-48bc-a0b2-23399ddaf4d8","Type":"ContainerDied","Data":"7a17bd5bc847af1fd9e52cbfc735adb9c04cdb9695e08dd9e2e0b0eee77a160f"} Nov 22 09:45:02 crc kubenswrapper[4789]: I1122 09:45:02.965244 4789 scope.go:117] "RemoveContainer" containerID="4ce6f99fe0532d0774ed3a238dd72bf64bbc780ed6b623e2fc3116d00f8d5b2f" Nov 22 09:45:02 crc kubenswrapper[4789]: E1122 09:45:02.965701 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:45:03 crc kubenswrapper[4789]: I1122 09:45:03.706453 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396745-q52jz" Nov 22 09:45:03 crc kubenswrapper[4789]: I1122 09:45:03.769234 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/adfe0251-8ae5-48bc-a0b2-23399ddaf4d8-config-volume\") pod \"adfe0251-8ae5-48bc-a0b2-23399ddaf4d8\" (UID: \"adfe0251-8ae5-48bc-a0b2-23399ddaf4d8\") " Nov 22 09:45:03 crc kubenswrapper[4789]: I1122 09:45:03.769554 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fxbjj\" (UniqueName: \"kubernetes.io/projected/adfe0251-8ae5-48bc-a0b2-23399ddaf4d8-kube-api-access-fxbjj\") pod \"adfe0251-8ae5-48bc-a0b2-23399ddaf4d8\" (UID: \"adfe0251-8ae5-48bc-a0b2-23399ddaf4d8\") " Nov 22 09:45:03 crc kubenswrapper[4789]: I1122 09:45:03.769659 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/adfe0251-8ae5-48bc-a0b2-23399ddaf4d8-secret-volume\") pod \"adfe0251-8ae5-48bc-a0b2-23399ddaf4d8\" (UID: \"adfe0251-8ae5-48bc-a0b2-23399ddaf4d8\") " Nov 22 09:45:03 crc kubenswrapper[4789]: I1122 09:45:03.769979 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/adfe0251-8ae5-48bc-a0b2-23399ddaf4d8-config-volume" (OuterVolumeSpecName: "config-volume") pod "adfe0251-8ae5-48bc-a0b2-23399ddaf4d8" (UID: "adfe0251-8ae5-48bc-a0b2-23399ddaf4d8"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 09:45:03 crc kubenswrapper[4789]: I1122 09:45:03.770285 4789 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/adfe0251-8ae5-48bc-a0b2-23399ddaf4d8-config-volume\") on node \"crc\" DevicePath \"\"" Nov 22 09:45:03 crc kubenswrapper[4789]: I1122 09:45:03.781115 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/adfe0251-8ae5-48bc-a0b2-23399ddaf4d8-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "adfe0251-8ae5-48bc-a0b2-23399ddaf4d8" (UID: "adfe0251-8ae5-48bc-a0b2-23399ddaf4d8"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 09:45:03 crc kubenswrapper[4789]: I1122 09:45:03.781188 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/adfe0251-8ae5-48bc-a0b2-23399ddaf4d8-kube-api-access-fxbjj" (OuterVolumeSpecName: "kube-api-access-fxbjj") pod "adfe0251-8ae5-48bc-a0b2-23399ddaf4d8" (UID: "adfe0251-8ae5-48bc-a0b2-23399ddaf4d8"). InnerVolumeSpecName "kube-api-access-fxbjj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 09:45:03 crc kubenswrapper[4789]: I1122 09:45:03.872029 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fxbjj\" (UniqueName: \"kubernetes.io/projected/adfe0251-8ae5-48bc-a0b2-23399ddaf4d8-kube-api-access-fxbjj\") on node \"crc\" DevicePath \"\"" Nov 22 09:45:03 crc kubenswrapper[4789]: I1122 09:45:03.872074 4789 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/adfe0251-8ae5-48bc-a0b2-23399ddaf4d8-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 22 09:45:04 crc kubenswrapper[4789]: I1122 09:45:04.200235 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29396745-q52jz" event={"ID":"adfe0251-8ae5-48bc-a0b2-23399ddaf4d8","Type":"ContainerDied","Data":"f1b374ffcfce15861e136d23cd8e0e42362ade6bfebbcc441e687cca1f988985"} Nov 22 09:45:04 crc kubenswrapper[4789]: I1122 09:45:04.200291 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f1b374ffcfce15861e136d23cd8e0e42362ade6bfebbcc441e687cca1f988985" Nov 22 09:45:04 crc kubenswrapper[4789]: I1122 09:45:04.200367 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396745-q52jz" Nov 22 09:45:04 crc kubenswrapper[4789]: I1122 09:45:04.785950 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396700-5m7wf"] Nov 22 09:45:04 crc kubenswrapper[4789]: I1122 09:45:04.796537 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396700-5m7wf"] Nov 22 09:45:05 crc kubenswrapper[4789]: I1122 09:45:05.979156 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b530e47e-9170-41a3-9135-19da688d2a1e" path="/var/lib/kubelet/pods/b530e47e-9170-41a3-9135-19da688d2a1e/volumes" Nov 22 09:45:17 crc kubenswrapper[4789]: I1122 09:45:17.978405 4789 scope.go:117] "RemoveContainer" containerID="4ce6f99fe0532d0774ed3a238dd72bf64bbc780ed6b623e2fc3116d00f8d5b2f" Nov 22 09:45:17 crc kubenswrapper[4789]: E1122 09:45:17.979461 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:45:29 crc kubenswrapper[4789]: I1122 09:45:29.781032 4789 scope.go:117] "RemoveContainer" containerID="099aa6447c7eb2f7abee9517700a8ad369c692c48bd9eb28ee3bb2d679b3a02e" Nov 22 09:45:30 crc kubenswrapper[4789]: I1122 09:45:30.964882 4789 scope.go:117] "RemoveContainer" containerID="4ce6f99fe0532d0774ed3a238dd72bf64bbc780ed6b623e2fc3116d00f8d5b2f" Nov 22 09:45:30 crc kubenswrapper[4789]: E1122 09:45:30.965428 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:45:41 crc kubenswrapper[4789]: I1122 09:45:41.965794 4789 scope.go:117] "RemoveContainer" containerID="4ce6f99fe0532d0774ed3a238dd72bf64bbc780ed6b623e2fc3116d00f8d5b2f" Nov 22 09:45:41 crc kubenswrapper[4789]: E1122 09:45:41.966577 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:45:53 crc kubenswrapper[4789]: I1122 09:45:53.970008 4789 scope.go:117] "RemoveContainer" containerID="4ce6f99fe0532d0774ed3a238dd72bf64bbc780ed6b623e2fc3116d00f8d5b2f" Nov 22 09:45:53 crc kubenswrapper[4789]: E1122 09:45:53.970684 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:46:07 crc kubenswrapper[4789]: I1122 09:46:07.979809 4789 scope.go:117] "RemoveContainer" containerID="4ce6f99fe0532d0774ed3a238dd72bf64bbc780ed6b623e2fc3116d00f8d5b2f" Nov 22 09:46:07 crc kubenswrapper[4789]: E1122 09:46:07.980878 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:46:22 crc kubenswrapper[4789]: I1122 09:46:15.624107 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/octavia-operator-controller-manager-6fdc856c5d-mccpp" podUID="bc429681-3e2a-4318-870d-ef62aef7d55b" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.83:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 22 09:46:22 crc kubenswrapper[4789]: I1122 09:46:15.624169 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/octavia-operator-controller-manager-6fdc856c5d-mccpp" podUID="bc429681-3e2a-4318-870d-ef62aef7d55b" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.83:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 22 09:46:22 crc kubenswrapper[4789]: I1122 09:46:19.375980 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/cinder-volume-volume1-0" podUID="2b1ad6b9-9f67-4533-bafd-5933ade7f938" containerName="cinder-volume" probeResult="failure" output="Get \"http://10.217.0.238:8080/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 22 09:46:22 crc kubenswrapper[4789]: I1122 09:46:19.423992 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/cinder-backup-0" podUID="62901ed9-929f-443e-ae1b-dc75cb2e1514" containerName="cinder-backup" probeResult="failure" output="Get \"http://10.217.0.239:8080/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 22 09:46:22 crc kubenswrapper[4789]: I1122 09:46:19.799018 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="deee1bbe-0f19-4f57-bebc-de5b01ccac7f" containerName="ceilometer-central-agent" probeResult="failure" output="command timed out" Nov 22 09:46:22 crc kubenswrapper[4789]: I1122 09:46:19.801409 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="deee1bbe-0f19-4f57-bebc-de5b01ccac7f" containerName="ceilometer-notification-agent" probeResult="failure" output="command timed out" Nov 22 09:46:22 crc kubenswrapper[4789]: I1122 09:46:20.442083 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/cinder-scheduler-0" podUID="b24f838e-1881-4921-8f51-8bb32afc4177" containerName="cinder-scheduler" probeResult="failure" output="Get \"http://10.217.0.158:8080/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 22 09:46:22 crc kubenswrapper[4789]: I1122 09:46:20.965885 4789 scope.go:117] "RemoveContainer" containerID="4ce6f99fe0532d0774ed3a238dd72bf64bbc780ed6b623e2fc3116d00f8d5b2f" Nov 22 09:46:22 crc kubenswrapper[4789]: E1122 09:46:20.966396 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:46:22 crc kubenswrapper[4789]: I1122 09:46:22.258329 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="deee1bbe-0f19-4f57-bebc-de5b01ccac7f" containerName="ceilometer-central-agent" probeResult="failure" output=< Nov 22 09:46:22 crc kubenswrapper[4789]: Unkown error: Expecting value: line 1 column 1 (char 0) Nov 22 09:46:22 crc kubenswrapper[4789]: > Nov 22 09:46:25 crc kubenswrapper[4789]: I1122 09:46:25.057819 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="deee1bbe-0f19-4f57-bebc-de5b01ccac7f" containerName="ceilometer-central-agent" probeResult="failure" output=< Nov 22 09:46:25 crc kubenswrapper[4789]: Unkown error: Expecting value: line 1 column 1 (char 0) Nov 22 09:46:25 crc kubenswrapper[4789]: > Nov 22 09:46:25 crc kubenswrapper[4789]: I1122 09:46:25.059408 4789 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/ceilometer-0" Nov 22 09:46:25 crc kubenswrapper[4789]: I1122 09:46:25.060738 4789 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="ceilometer-central-agent" containerStatusID={"Type":"cri-o","ID":"156c27da864eebfb12372068b839ebbd34f238ba528e1e5a41b839fc5378b2a2"} pod="openstack/ceilometer-0" containerMessage="Container ceilometer-central-agent failed liveness probe, will be restarted" Nov 22 09:46:25 crc kubenswrapper[4789]: I1122 09:46:25.061141 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="deee1bbe-0f19-4f57-bebc-de5b01ccac7f" containerName="ceilometer-central-agent" containerID="cri-o://156c27da864eebfb12372068b839ebbd34f238ba528e1e5a41b839fc5378b2a2" gracePeriod=30 Nov 22 09:46:25 crc kubenswrapper[4789]: I1122 09:46:25.505262 4789 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 22 09:46:25 crc kubenswrapper[4789]: I1122 09:46:25.536370 4789 generic.go:334] "Generic (PLEG): container finished" podID="deee1bbe-0f19-4f57-bebc-de5b01ccac7f" containerID="156c27da864eebfb12372068b839ebbd34f238ba528e1e5a41b839fc5378b2a2" exitCode=0 Nov 22 09:46:25 crc kubenswrapper[4789]: I1122 09:46:25.536406 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"deee1bbe-0f19-4f57-bebc-de5b01ccac7f","Type":"ContainerDied","Data":"156c27da864eebfb12372068b839ebbd34f238ba528e1e5a41b839fc5378b2a2"} Nov 22 09:46:27 crc kubenswrapper[4789]: I1122 09:46:27.557823 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"deee1bbe-0f19-4f57-bebc-de5b01ccac7f","Type":"ContainerStarted","Data":"71d9d32a49efdf8d857033a51fef42c4eb54719f30871bbe8f30ed0529efda72"} Nov 22 09:46:35 crc kubenswrapper[4789]: I1122 09:46:35.966308 4789 scope.go:117] "RemoveContainer" containerID="4ce6f99fe0532d0774ed3a238dd72bf64bbc780ed6b623e2fc3116d00f8d5b2f" Nov 22 09:46:35 crc kubenswrapper[4789]: E1122 09:46:35.967315 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:46:50 crc kubenswrapper[4789]: I1122 09:46:50.965230 4789 scope.go:117] "RemoveContainer" containerID="4ce6f99fe0532d0774ed3a238dd72bf64bbc780ed6b623e2fc3116d00f8d5b2f" Nov 22 09:46:50 crc kubenswrapper[4789]: E1122 09:46:50.966083 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:46:58 crc kubenswrapper[4789]: I1122 09:46:58.403765 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-flcts"] Nov 22 09:46:58 crc kubenswrapper[4789]: E1122 09:46:58.404785 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="adfe0251-8ae5-48bc-a0b2-23399ddaf4d8" containerName="collect-profiles" Nov 22 09:46:58 crc kubenswrapper[4789]: I1122 09:46:58.404801 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="adfe0251-8ae5-48bc-a0b2-23399ddaf4d8" containerName="collect-profiles" Nov 22 09:46:58 crc kubenswrapper[4789]: I1122 09:46:58.405013 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="adfe0251-8ae5-48bc-a0b2-23399ddaf4d8" containerName="collect-profiles" Nov 22 09:46:58 crc kubenswrapper[4789]: I1122 09:46:58.406570 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-flcts" Nov 22 09:46:58 crc kubenswrapper[4789]: I1122 09:46:58.414411 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-flcts"] Nov 22 09:46:58 crc kubenswrapper[4789]: I1122 09:46:58.502885 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1776d674-3207-4d31-934f-ce7f68685d0b-utilities\") pod \"redhat-marketplace-flcts\" (UID: \"1776d674-3207-4d31-934f-ce7f68685d0b\") " pod="openshift-marketplace/redhat-marketplace-flcts" Nov 22 09:46:58 crc kubenswrapper[4789]: I1122 09:46:58.502949 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1776d674-3207-4d31-934f-ce7f68685d0b-catalog-content\") pod \"redhat-marketplace-flcts\" (UID: \"1776d674-3207-4d31-934f-ce7f68685d0b\") " pod="openshift-marketplace/redhat-marketplace-flcts" Nov 22 09:46:58 crc kubenswrapper[4789]: I1122 09:46:58.503114 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bsjgn\" (UniqueName: \"kubernetes.io/projected/1776d674-3207-4d31-934f-ce7f68685d0b-kube-api-access-bsjgn\") pod \"redhat-marketplace-flcts\" (UID: \"1776d674-3207-4d31-934f-ce7f68685d0b\") " pod="openshift-marketplace/redhat-marketplace-flcts" Nov 22 09:46:58 crc kubenswrapper[4789]: I1122 09:46:58.604646 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bsjgn\" (UniqueName: \"kubernetes.io/projected/1776d674-3207-4d31-934f-ce7f68685d0b-kube-api-access-bsjgn\") pod \"redhat-marketplace-flcts\" (UID: \"1776d674-3207-4d31-934f-ce7f68685d0b\") " pod="openshift-marketplace/redhat-marketplace-flcts" Nov 22 09:46:58 crc kubenswrapper[4789]: I1122 09:46:58.604730 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1776d674-3207-4d31-934f-ce7f68685d0b-utilities\") pod \"redhat-marketplace-flcts\" (UID: \"1776d674-3207-4d31-934f-ce7f68685d0b\") " pod="openshift-marketplace/redhat-marketplace-flcts" Nov 22 09:46:58 crc kubenswrapper[4789]: I1122 09:46:58.604772 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1776d674-3207-4d31-934f-ce7f68685d0b-catalog-content\") pod \"redhat-marketplace-flcts\" (UID: \"1776d674-3207-4d31-934f-ce7f68685d0b\") " pod="openshift-marketplace/redhat-marketplace-flcts" Nov 22 09:46:58 crc kubenswrapper[4789]: I1122 09:46:58.605238 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1776d674-3207-4d31-934f-ce7f68685d0b-catalog-content\") pod \"redhat-marketplace-flcts\" (UID: \"1776d674-3207-4d31-934f-ce7f68685d0b\") " pod="openshift-marketplace/redhat-marketplace-flcts" Nov 22 09:46:58 crc kubenswrapper[4789]: I1122 09:46:58.605727 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1776d674-3207-4d31-934f-ce7f68685d0b-utilities\") pod \"redhat-marketplace-flcts\" (UID: \"1776d674-3207-4d31-934f-ce7f68685d0b\") " pod="openshift-marketplace/redhat-marketplace-flcts" Nov 22 09:46:58 crc kubenswrapper[4789]: I1122 09:46:58.624325 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bsjgn\" (UniqueName: \"kubernetes.io/projected/1776d674-3207-4d31-934f-ce7f68685d0b-kube-api-access-bsjgn\") pod \"redhat-marketplace-flcts\" (UID: \"1776d674-3207-4d31-934f-ce7f68685d0b\") " pod="openshift-marketplace/redhat-marketplace-flcts" Nov 22 09:46:58 crc kubenswrapper[4789]: I1122 09:46:58.774685 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-flcts" Nov 22 09:46:59 crc kubenswrapper[4789]: I1122 09:46:59.225462 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-flcts"] Nov 22 09:46:59 crc kubenswrapper[4789]: I1122 09:46:59.887500 4789 generic.go:334] "Generic (PLEG): container finished" podID="1776d674-3207-4d31-934f-ce7f68685d0b" containerID="54721876b9ef89602b3727ae24f8700d14790424b0cb77b5552eb3fa212ff403" exitCode=0 Nov 22 09:46:59 crc kubenswrapper[4789]: I1122 09:46:59.887585 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-flcts" event={"ID":"1776d674-3207-4d31-934f-ce7f68685d0b","Type":"ContainerDied","Data":"54721876b9ef89602b3727ae24f8700d14790424b0cb77b5552eb3fa212ff403"} Nov 22 09:46:59 crc kubenswrapper[4789]: I1122 09:46:59.887807 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-flcts" event={"ID":"1776d674-3207-4d31-934f-ce7f68685d0b","Type":"ContainerStarted","Data":"750ee7b86ab83358f15b73b8883caf93806cad3315354b20ccb502d9878e9f61"} Nov 22 09:47:00 crc kubenswrapper[4789]: I1122 09:47:00.899405 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-flcts" event={"ID":"1776d674-3207-4d31-934f-ce7f68685d0b","Type":"ContainerStarted","Data":"3b367d2477c16b710ff8c1672b800da2f6c730265311debd7cdf1b6eed53c0e6"} Nov 22 09:47:02 crc kubenswrapper[4789]: I1122 09:47:02.920544 4789 generic.go:334] "Generic (PLEG): container finished" podID="1776d674-3207-4d31-934f-ce7f68685d0b" containerID="3b367d2477c16b710ff8c1672b800da2f6c730265311debd7cdf1b6eed53c0e6" exitCode=0 Nov 22 09:47:02 crc kubenswrapper[4789]: I1122 09:47:02.920623 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-flcts" event={"ID":"1776d674-3207-4d31-934f-ce7f68685d0b","Type":"ContainerDied","Data":"3b367d2477c16b710ff8c1672b800da2f6c730265311debd7cdf1b6eed53c0e6"} Nov 22 09:47:03 crc kubenswrapper[4789]: I1122 09:47:03.936303 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-flcts" event={"ID":"1776d674-3207-4d31-934f-ce7f68685d0b","Type":"ContainerStarted","Data":"a91f9e1ac30d062a20fd46dc5bd8ce5f1f82d27687eacc14eab040e4d4bfdb81"} Nov 22 09:47:03 crc kubenswrapper[4789]: I1122 09:47:03.963493 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-flcts" podStartSLOduration=2.503507634 podStartE2EDuration="5.963469874s" podCreationTimestamp="2025-11-22 09:46:58 +0000 UTC" firstStartedPulling="2025-11-22 09:46:59.893438793 +0000 UTC m=+6634.127839066" lastFinishedPulling="2025-11-22 09:47:03.353401033 +0000 UTC m=+6637.587801306" observedRunningTime="2025-11-22 09:47:03.961366137 +0000 UTC m=+6638.195766420" watchObservedRunningTime="2025-11-22 09:47:03.963469874 +0000 UTC m=+6638.197870157" Nov 22 09:47:04 crc kubenswrapper[4789]: I1122 09:47:04.965912 4789 scope.go:117] "RemoveContainer" containerID="4ce6f99fe0532d0774ed3a238dd72bf64bbc780ed6b623e2fc3116d00f8d5b2f" Nov 22 09:47:04 crc kubenswrapper[4789]: E1122 09:47:04.966297 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:47:08 crc kubenswrapper[4789]: I1122 09:47:08.775302 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-flcts" Nov 22 09:47:08 crc kubenswrapper[4789]: I1122 09:47:08.775981 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-flcts" Nov 22 09:47:08 crc kubenswrapper[4789]: I1122 09:47:08.852942 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-flcts" Nov 22 09:47:09 crc kubenswrapper[4789]: I1122 09:47:09.057678 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-flcts" Nov 22 09:47:09 crc kubenswrapper[4789]: I1122 09:47:09.119820 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-flcts"] Nov 22 09:47:11 crc kubenswrapper[4789]: I1122 09:47:11.008339 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-flcts" podUID="1776d674-3207-4d31-934f-ce7f68685d0b" containerName="registry-server" containerID="cri-o://a91f9e1ac30d062a20fd46dc5bd8ce5f1f82d27687eacc14eab040e4d4bfdb81" gracePeriod=2 Nov 22 09:47:11 crc kubenswrapper[4789]: I1122 09:47:11.594174 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-flcts" Nov 22 09:47:11 crc kubenswrapper[4789]: I1122 09:47:11.688665 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1776d674-3207-4d31-934f-ce7f68685d0b-catalog-content\") pod \"1776d674-3207-4d31-934f-ce7f68685d0b\" (UID: \"1776d674-3207-4d31-934f-ce7f68685d0b\") " Nov 22 09:47:11 crc kubenswrapper[4789]: I1122 09:47:11.689053 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bsjgn\" (UniqueName: \"kubernetes.io/projected/1776d674-3207-4d31-934f-ce7f68685d0b-kube-api-access-bsjgn\") pod \"1776d674-3207-4d31-934f-ce7f68685d0b\" (UID: \"1776d674-3207-4d31-934f-ce7f68685d0b\") " Nov 22 09:47:11 crc kubenswrapper[4789]: I1122 09:47:11.689116 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1776d674-3207-4d31-934f-ce7f68685d0b-utilities\") pod \"1776d674-3207-4d31-934f-ce7f68685d0b\" (UID: \"1776d674-3207-4d31-934f-ce7f68685d0b\") " Nov 22 09:47:11 crc kubenswrapper[4789]: I1122 09:47:11.690536 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1776d674-3207-4d31-934f-ce7f68685d0b-utilities" (OuterVolumeSpecName: "utilities") pod "1776d674-3207-4d31-934f-ce7f68685d0b" (UID: "1776d674-3207-4d31-934f-ce7f68685d0b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 09:47:11 crc kubenswrapper[4789]: I1122 09:47:11.694699 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1776d674-3207-4d31-934f-ce7f68685d0b-kube-api-access-bsjgn" (OuterVolumeSpecName: "kube-api-access-bsjgn") pod "1776d674-3207-4d31-934f-ce7f68685d0b" (UID: "1776d674-3207-4d31-934f-ce7f68685d0b"). InnerVolumeSpecName "kube-api-access-bsjgn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 09:47:11 crc kubenswrapper[4789]: I1122 09:47:11.706737 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1776d674-3207-4d31-934f-ce7f68685d0b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1776d674-3207-4d31-934f-ce7f68685d0b" (UID: "1776d674-3207-4d31-934f-ce7f68685d0b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 09:47:11 crc kubenswrapper[4789]: I1122 09:47:11.791947 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1776d674-3207-4d31-934f-ce7f68685d0b-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 09:47:11 crc kubenswrapper[4789]: I1122 09:47:11.791998 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bsjgn\" (UniqueName: \"kubernetes.io/projected/1776d674-3207-4d31-934f-ce7f68685d0b-kube-api-access-bsjgn\") on node \"crc\" DevicePath \"\"" Nov 22 09:47:11 crc kubenswrapper[4789]: I1122 09:47:11.792008 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1776d674-3207-4d31-934f-ce7f68685d0b-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 09:47:12 crc kubenswrapper[4789]: I1122 09:47:12.026326 4789 generic.go:334] "Generic (PLEG): container finished" podID="1776d674-3207-4d31-934f-ce7f68685d0b" containerID="a91f9e1ac30d062a20fd46dc5bd8ce5f1f82d27687eacc14eab040e4d4bfdb81" exitCode=0 Nov 22 09:47:12 crc kubenswrapper[4789]: I1122 09:47:12.026400 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-flcts" event={"ID":"1776d674-3207-4d31-934f-ce7f68685d0b","Type":"ContainerDied","Data":"a91f9e1ac30d062a20fd46dc5bd8ce5f1f82d27687eacc14eab040e4d4bfdb81"} Nov 22 09:47:12 crc kubenswrapper[4789]: I1122 09:47:12.026452 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-flcts" event={"ID":"1776d674-3207-4d31-934f-ce7f68685d0b","Type":"ContainerDied","Data":"750ee7b86ab83358f15b73b8883caf93806cad3315354b20ccb502d9878e9f61"} Nov 22 09:47:12 crc kubenswrapper[4789]: I1122 09:47:12.026473 4789 scope.go:117] "RemoveContainer" containerID="a91f9e1ac30d062a20fd46dc5bd8ce5f1f82d27687eacc14eab040e4d4bfdb81" Nov 22 09:47:12 crc kubenswrapper[4789]: I1122 09:47:12.026541 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-flcts" Nov 22 09:47:12 crc kubenswrapper[4789]: I1122 09:47:12.062228 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-flcts"] Nov 22 09:47:12 crc kubenswrapper[4789]: I1122 09:47:12.069154 4789 scope.go:117] "RemoveContainer" containerID="3b367d2477c16b710ff8c1672b800da2f6c730265311debd7cdf1b6eed53c0e6" Nov 22 09:47:12 crc kubenswrapper[4789]: I1122 09:47:12.076494 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-flcts"] Nov 22 09:47:12 crc kubenswrapper[4789]: I1122 09:47:12.104610 4789 scope.go:117] "RemoveContainer" containerID="54721876b9ef89602b3727ae24f8700d14790424b0cb77b5552eb3fa212ff403" Nov 22 09:47:12 crc kubenswrapper[4789]: I1122 09:47:12.168715 4789 scope.go:117] "RemoveContainer" containerID="a91f9e1ac30d062a20fd46dc5bd8ce5f1f82d27687eacc14eab040e4d4bfdb81" Nov 22 09:47:12 crc kubenswrapper[4789]: E1122 09:47:12.169283 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a91f9e1ac30d062a20fd46dc5bd8ce5f1f82d27687eacc14eab040e4d4bfdb81\": container with ID starting with a91f9e1ac30d062a20fd46dc5bd8ce5f1f82d27687eacc14eab040e4d4bfdb81 not found: ID does not exist" containerID="a91f9e1ac30d062a20fd46dc5bd8ce5f1f82d27687eacc14eab040e4d4bfdb81" Nov 22 09:47:12 crc kubenswrapper[4789]: I1122 09:47:12.169331 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a91f9e1ac30d062a20fd46dc5bd8ce5f1f82d27687eacc14eab040e4d4bfdb81"} err="failed to get container status \"a91f9e1ac30d062a20fd46dc5bd8ce5f1f82d27687eacc14eab040e4d4bfdb81\": rpc error: code = NotFound desc = could not find container \"a91f9e1ac30d062a20fd46dc5bd8ce5f1f82d27687eacc14eab040e4d4bfdb81\": container with ID starting with a91f9e1ac30d062a20fd46dc5bd8ce5f1f82d27687eacc14eab040e4d4bfdb81 not found: ID does not exist" Nov 22 09:47:12 crc kubenswrapper[4789]: I1122 09:47:12.169365 4789 scope.go:117] "RemoveContainer" containerID="3b367d2477c16b710ff8c1672b800da2f6c730265311debd7cdf1b6eed53c0e6" Nov 22 09:47:12 crc kubenswrapper[4789]: E1122 09:47:12.169723 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3b367d2477c16b710ff8c1672b800da2f6c730265311debd7cdf1b6eed53c0e6\": container with ID starting with 3b367d2477c16b710ff8c1672b800da2f6c730265311debd7cdf1b6eed53c0e6 not found: ID does not exist" containerID="3b367d2477c16b710ff8c1672b800da2f6c730265311debd7cdf1b6eed53c0e6" Nov 22 09:47:12 crc kubenswrapper[4789]: I1122 09:47:12.169781 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b367d2477c16b710ff8c1672b800da2f6c730265311debd7cdf1b6eed53c0e6"} err="failed to get container status \"3b367d2477c16b710ff8c1672b800da2f6c730265311debd7cdf1b6eed53c0e6\": rpc error: code = NotFound desc = could not find container \"3b367d2477c16b710ff8c1672b800da2f6c730265311debd7cdf1b6eed53c0e6\": container with ID starting with 3b367d2477c16b710ff8c1672b800da2f6c730265311debd7cdf1b6eed53c0e6 not found: ID does not exist" Nov 22 09:47:12 crc kubenswrapper[4789]: I1122 09:47:12.169800 4789 scope.go:117] "RemoveContainer" containerID="54721876b9ef89602b3727ae24f8700d14790424b0cb77b5552eb3fa212ff403" Nov 22 09:47:12 crc kubenswrapper[4789]: E1122 09:47:12.170103 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"54721876b9ef89602b3727ae24f8700d14790424b0cb77b5552eb3fa212ff403\": container with ID starting with 54721876b9ef89602b3727ae24f8700d14790424b0cb77b5552eb3fa212ff403 not found: ID does not exist" containerID="54721876b9ef89602b3727ae24f8700d14790424b0cb77b5552eb3fa212ff403" Nov 22 09:47:12 crc kubenswrapper[4789]: I1122 09:47:12.170133 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"54721876b9ef89602b3727ae24f8700d14790424b0cb77b5552eb3fa212ff403"} err="failed to get container status \"54721876b9ef89602b3727ae24f8700d14790424b0cb77b5552eb3fa212ff403\": rpc error: code = NotFound desc = could not find container \"54721876b9ef89602b3727ae24f8700d14790424b0cb77b5552eb3fa212ff403\": container with ID starting with 54721876b9ef89602b3727ae24f8700d14790424b0cb77b5552eb3fa212ff403 not found: ID does not exist" Nov 22 09:47:13 crc kubenswrapper[4789]: I1122 09:47:13.985017 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1776d674-3207-4d31-934f-ce7f68685d0b" path="/var/lib/kubelet/pods/1776d674-3207-4d31-934f-ce7f68685d0b/volumes" Nov 22 09:47:18 crc kubenswrapper[4789]: I1122 09:47:18.965739 4789 scope.go:117] "RemoveContainer" containerID="4ce6f99fe0532d0774ed3a238dd72bf64bbc780ed6b623e2fc3116d00f8d5b2f" Nov 22 09:47:18 crc kubenswrapper[4789]: E1122 09:47:18.967084 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:47:33 crc kubenswrapper[4789]: I1122 09:47:33.966149 4789 scope.go:117] "RemoveContainer" containerID="4ce6f99fe0532d0774ed3a238dd72bf64bbc780ed6b623e2fc3116d00f8d5b2f" Nov 22 09:47:33 crc kubenswrapper[4789]: E1122 09:47:33.967068 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:47:35 crc kubenswrapper[4789]: I1122 09:47:35.227243 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/heat-operator-controller-manager-bf4c6585d-6shgh" podUID="0cf958d5-222e-4598-a1a7-fe887644cd24" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.74:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 22 09:47:36 crc kubenswrapper[4789]: I1122 09:47:36.798124 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-fzmzt" podUID="03652378-7433-48c7-a601-aea6d67f1a14" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.91:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 22 09:47:36 crc kubenswrapper[4789]: I1122 09:47:36.839129 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-fzmzt" podUID="03652378-7433-48c7-a601-aea6d67f1a14" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.91:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 22 09:47:45 crc kubenswrapper[4789]: I1122 09:47:45.227999 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/heat-operator-controller-manager-bf4c6585d-6shgh" podUID="0cf958d5-222e-4598-a1a7-fe887644cd24" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.74:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 22 09:47:46 crc kubenswrapper[4789]: I1122 09:47:46.965699 4789 scope.go:117] "RemoveContainer" containerID="4ce6f99fe0532d0774ed3a238dd72bf64bbc780ed6b623e2fc3116d00f8d5b2f" Nov 22 09:47:46 crc kubenswrapper[4789]: E1122 09:47:46.966267 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:47:47 crc kubenswrapper[4789]: I1122 09:47:47.441975 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/cinder-scheduler-0" podUID="b24f838e-1881-4921-8f51-8bb32afc4177" containerName="cinder-scheduler" probeResult="failure" output="Get \"http://10.217.0.158:8080/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 22 09:48:01 crc kubenswrapper[4789]: I1122 09:48:01.965238 4789 scope.go:117] "RemoveContainer" containerID="4ce6f99fe0532d0774ed3a238dd72bf64bbc780ed6b623e2fc3116d00f8d5b2f" Nov 22 09:48:01 crc kubenswrapper[4789]: E1122 09:48:01.967505 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:48:16 crc kubenswrapper[4789]: I1122 09:48:16.965492 4789 scope.go:117] "RemoveContainer" containerID="4ce6f99fe0532d0774ed3a238dd72bf64bbc780ed6b623e2fc3116d00f8d5b2f" Nov 22 09:48:17 crc kubenswrapper[4789]: I1122 09:48:17.671978 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" event={"ID":"58aa8071-7ecc-4692-8789-2db04ec70647","Type":"ContainerStarted","Data":"2e98f18b68c5e7c627f36e814e1953271056ec4c2d4a51091ea26f228a4fa550"} Nov 22 09:50:12 crc kubenswrapper[4789]: I1122 09:50:12.828627 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-h76zf"] Nov 22 09:50:12 crc kubenswrapper[4789]: E1122 09:50:12.829650 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1776d674-3207-4d31-934f-ce7f68685d0b" containerName="extract-utilities" Nov 22 09:50:12 crc kubenswrapper[4789]: I1122 09:50:12.829666 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="1776d674-3207-4d31-934f-ce7f68685d0b" containerName="extract-utilities" Nov 22 09:50:12 crc kubenswrapper[4789]: E1122 09:50:12.829688 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1776d674-3207-4d31-934f-ce7f68685d0b" containerName="registry-server" Nov 22 09:50:12 crc kubenswrapper[4789]: I1122 09:50:12.829696 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="1776d674-3207-4d31-934f-ce7f68685d0b" containerName="registry-server" Nov 22 09:50:12 crc kubenswrapper[4789]: E1122 09:50:12.829721 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1776d674-3207-4d31-934f-ce7f68685d0b" containerName="extract-content" Nov 22 09:50:12 crc kubenswrapper[4789]: I1122 09:50:12.829730 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="1776d674-3207-4d31-934f-ce7f68685d0b" containerName="extract-content" Nov 22 09:50:12 crc kubenswrapper[4789]: I1122 09:50:12.830033 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="1776d674-3207-4d31-934f-ce7f68685d0b" containerName="registry-server" Nov 22 09:50:12 crc kubenswrapper[4789]: I1122 09:50:12.831734 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-h76zf" Nov 22 09:50:12 crc kubenswrapper[4789]: I1122 09:50:12.869974 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4b60a15-7288-4119-aaba-82731be0e537-utilities\") pod \"certified-operators-h76zf\" (UID: \"b4b60a15-7288-4119-aaba-82731be0e537\") " pod="openshift-marketplace/certified-operators-h76zf" Nov 22 09:50:12 crc kubenswrapper[4789]: I1122 09:50:12.870079 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4b60a15-7288-4119-aaba-82731be0e537-catalog-content\") pod \"certified-operators-h76zf\" (UID: \"b4b60a15-7288-4119-aaba-82731be0e537\") " pod="openshift-marketplace/certified-operators-h76zf" Nov 22 09:50:12 crc kubenswrapper[4789]: I1122 09:50:12.870182 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gdv6b\" (UniqueName: \"kubernetes.io/projected/b4b60a15-7288-4119-aaba-82731be0e537-kube-api-access-gdv6b\") pod \"certified-operators-h76zf\" (UID: \"b4b60a15-7288-4119-aaba-82731be0e537\") " pod="openshift-marketplace/certified-operators-h76zf" Nov 22 09:50:12 crc kubenswrapper[4789]: I1122 09:50:12.870493 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-h76zf"] Nov 22 09:50:12 crc kubenswrapper[4789]: I1122 09:50:12.972569 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4b60a15-7288-4119-aaba-82731be0e537-utilities\") pod \"certified-operators-h76zf\" (UID: \"b4b60a15-7288-4119-aaba-82731be0e537\") " pod="openshift-marketplace/certified-operators-h76zf" Nov 22 09:50:12 crc kubenswrapper[4789]: I1122 09:50:12.972644 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4b60a15-7288-4119-aaba-82731be0e537-catalog-content\") pod \"certified-operators-h76zf\" (UID: \"b4b60a15-7288-4119-aaba-82731be0e537\") " pod="openshift-marketplace/certified-operators-h76zf" Nov 22 09:50:12 crc kubenswrapper[4789]: I1122 09:50:12.972710 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gdv6b\" (UniqueName: \"kubernetes.io/projected/b4b60a15-7288-4119-aaba-82731be0e537-kube-api-access-gdv6b\") pod \"certified-operators-h76zf\" (UID: \"b4b60a15-7288-4119-aaba-82731be0e537\") " pod="openshift-marketplace/certified-operators-h76zf" Nov 22 09:50:12 crc kubenswrapper[4789]: I1122 09:50:12.973519 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4b60a15-7288-4119-aaba-82731be0e537-utilities\") pod \"certified-operators-h76zf\" (UID: \"b4b60a15-7288-4119-aaba-82731be0e537\") " pod="openshift-marketplace/certified-operators-h76zf" Nov 22 09:50:12 crc kubenswrapper[4789]: I1122 09:50:12.973592 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4b60a15-7288-4119-aaba-82731be0e537-catalog-content\") pod \"certified-operators-h76zf\" (UID: \"b4b60a15-7288-4119-aaba-82731be0e537\") " pod="openshift-marketplace/certified-operators-h76zf" Nov 22 09:50:12 crc kubenswrapper[4789]: I1122 09:50:12.991999 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gdv6b\" (UniqueName: \"kubernetes.io/projected/b4b60a15-7288-4119-aaba-82731be0e537-kube-api-access-gdv6b\") pod \"certified-operators-h76zf\" (UID: \"b4b60a15-7288-4119-aaba-82731be0e537\") " pod="openshift-marketplace/certified-operators-h76zf" Nov 22 09:50:13 crc kubenswrapper[4789]: I1122 09:50:13.162233 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-h76zf" Nov 22 09:50:13 crc kubenswrapper[4789]: I1122 09:50:13.752657 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-h76zf"] Nov 22 09:50:13 crc kubenswrapper[4789]: W1122 09:50:13.757790 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb4b60a15_7288_4119_aaba_82731be0e537.slice/crio-d7105a374257c7cf9fb67ca9a03ce066f6f8c280e7f3496592156e48b92ef844 WatchSource:0}: Error finding container d7105a374257c7cf9fb67ca9a03ce066f6f8c280e7f3496592156e48b92ef844: Status 404 returned error can't find the container with id d7105a374257c7cf9fb67ca9a03ce066f6f8c280e7f3496592156e48b92ef844 Nov 22 09:50:13 crc kubenswrapper[4789]: I1122 09:50:13.821834 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h76zf" event={"ID":"b4b60a15-7288-4119-aaba-82731be0e537","Type":"ContainerStarted","Data":"d7105a374257c7cf9fb67ca9a03ce066f6f8c280e7f3496592156e48b92ef844"} Nov 22 09:50:14 crc kubenswrapper[4789]: I1122 09:50:14.837303 4789 generic.go:334] "Generic (PLEG): container finished" podID="b4b60a15-7288-4119-aaba-82731be0e537" containerID="cad7f14ed5296cb87cb0e62dc76c30d127c1b332a1c3d0a9cfa7030e55cceb1a" exitCode=0 Nov 22 09:50:14 crc kubenswrapper[4789]: I1122 09:50:14.837411 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h76zf" event={"ID":"b4b60a15-7288-4119-aaba-82731be0e537","Type":"ContainerDied","Data":"cad7f14ed5296cb87cb0e62dc76c30d127c1b332a1c3d0a9cfa7030e55cceb1a"} Nov 22 09:50:15 crc kubenswrapper[4789]: I1122 09:50:15.375155 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-5dd6b"] Nov 22 09:50:15 crc kubenswrapper[4789]: I1122 09:50:15.378163 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5dd6b" Nov 22 09:50:15 crc kubenswrapper[4789]: I1122 09:50:15.389366 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5dd6b"] Nov 22 09:50:15 crc kubenswrapper[4789]: I1122 09:50:15.432476 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7514e0b6-fae4-4419-8e5f-7798884fe6d5-catalog-content\") pod \"redhat-operators-5dd6b\" (UID: \"7514e0b6-fae4-4419-8e5f-7798884fe6d5\") " pod="openshift-marketplace/redhat-operators-5dd6b" Nov 22 09:50:15 crc kubenswrapper[4789]: I1122 09:50:15.432577 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7514e0b6-fae4-4419-8e5f-7798884fe6d5-utilities\") pod \"redhat-operators-5dd6b\" (UID: \"7514e0b6-fae4-4419-8e5f-7798884fe6d5\") " pod="openshift-marketplace/redhat-operators-5dd6b" Nov 22 09:50:15 crc kubenswrapper[4789]: I1122 09:50:15.432687 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nbp42\" (UniqueName: \"kubernetes.io/projected/7514e0b6-fae4-4419-8e5f-7798884fe6d5-kube-api-access-nbp42\") pod \"redhat-operators-5dd6b\" (UID: \"7514e0b6-fae4-4419-8e5f-7798884fe6d5\") " pod="openshift-marketplace/redhat-operators-5dd6b" Nov 22 09:50:15 crc kubenswrapper[4789]: I1122 09:50:15.534864 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7514e0b6-fae4-4419-8e5f-7798884fe6d5-catalog-content\") pod \"redhat-operators-5dd6b\" (UID: \"7514e0b6-fae4-4419-8e5f-7798884fe6d5\") " pod="openshift-marketplace/redhat-operators-5dd6b" Nov 22 09:50:15 crc kubenswrapper[4789]: I1122 09:50:15.534923 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7514e0b6-fae4-4419-8e5f-7798884fe6d5-utilities\") pod \"redhat-operators-5dd6b\" (UID: \"7514e0b6-fae4-4419-8e5f-7798884fe6d5\") " pod="openshift-marketplace/redhat-operators-5dd6b" Nov 22 09:50:15 crc kubenswrapper[4789]: I1122 09:50:15.534963 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nbp42\" (UniqueName: \"kubernetes.io/projected/7514e0b6-fae4-4419-8e5f-7798884fe6d5-kube-api-access-nbp42\") pod \"redhat-operators-5dd6b\" (UID: \"7514e0b6-fae4-4419-8e5f-7798884fe6d5\") " pod="openshift-marketplace/redhat-operators-5dd6b" Nov 22 09:50:15 crc kubenswrapper[4789]: I1122 09:50:15.535506 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7514e0b6-fae4-4419-8e5f-7798884fe6d5-catalog-content\") pod \"redhat-operators-5dd6b\" (UID: \"7514e0b6-fae4-4419-8e5f-7798884fe6d5\") " pod="openshift-marketplace/redhat-operators-5dd6b" Nov 22 09:50:15 crc kubenswrapper[4789]: I1122 09:50:15.535578 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7514e0b6-fae4-4419-8e5f-7798884fe6d5-utilities\") pod \"redhat-operators-5dd6b\" (UID: \"7514e0b6-fae4-4419-8e5f-7798884fe6d5\") " pod="openshift-marketplace/redhat-operators-5dd6b" Nov 22 09:50:15 crc kubenswrapper[4789]: I1122 09:50:15.564031 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nbp42\" (UniqueName: \"kubernetes.io/projected/7514e0b6-fae4-4419-8e5f-7798884fe6d5-kube-api-access-nbp42\") pod \"redhat-operators-5dd6b\" (UID: \"7514e0b6-fae4-4419-8e5f-7798884fe6d5\") " pod="openshift-marketplace/redhat-operators-5dd6b" Nov 22 09:50:15 crc kubenswrapper[4789]: I1122 09:50:15.699542 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5dd6b" Nov 22 09:50:16 crc kubenswrapper[4789]: W1122 09:50:16.207345 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7514e0b6_fae4_4419_8e5f_7798884fe6d5.slice/crio-8b855b552af9ea44a75acb028f0799dcce03ae40bfad5a654453b835ff13ce9a WatchSource:0}: Error finding container 8b855b552af9ea44a75acb028f0799dcce03ae40bfad5a654453b835ff13ce9a: Status 404 returned error can't find the container with id 8b855b552af9ea44a75acb028f0799dcce03ae40bfad5a654453b835ff13ce9a Nov 22 09:50:16 crc kubenswrapper[4789]: I1122 09:50:16.218526 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5dd6b"] Nov 22 09:50:16 crc kubenswrapper[4789]: I1122 09:50:16.870117 4789 generic.go:334] "Generic (PLEG): container finished" podID="7514e0b6-fae4-4419-8e5f-7798884fe6d5" containerID="692ddc87bb90852c79f33d81b8a1b71234f38c9c98fcede0d1076bab7c8be7b4" exitCode=0 Nov 22 09:50:16 crc kubenswrapper[4789]: I1122 09:50:16.870647 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5dd6b" event={"ID":"7514e0b6-fae4-4419-8e5f-7798884fe6d5","Type":"ContainerDied","Data":"692ddc87bb90852c79f33d81b8a1b71234f38c9c98fcede0d1076bab7c8be7b4"} Nov 22 09:50:16 crc kubenswrapper[4789]: I1122 09:50:16.870710 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5dd6b" event={"ID":"7514e0b6-fae4-4419-8e5f-7798884fe6d5","Type":"ContainerStarted","Data":"8b855b552af9ea44a75acb028f0799dcce03ae40bfad5a654453b835ff13ce9a"} Nov 22 09:50:16 crc kubenswrapper[4789]: I1122 09:50:16.875160 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h76zf" event={"ID":"b4b60a15-7288-4119-aaba-82731be0e537","Type":"ContainerStarted","Data":"7196c7a7960d8de8eb3521c013841d1a17ec1a381a049bdd41937224c8cdd791"} Nov 22 09:50:17 crc kubenswrapper[4789]: I1122 09:50:17.885580 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5dd6b" event={"ID":"7514e0b6-fae4-4419-8e5f-7798884fe6d5","Type":"ContainerStarted","Data":"c93637949a843fcf9578f5209f87d7e8ebe8c382a467bcf07b45cec69a45cef2"} Nov 22 09:50:17 crc kubenswrapper[4789]: I1122 09:50:17.888545 4789 generic.go:334] "Generic (PLEG): container finished" podID="b4b60a15-7288-4119-aaba-82731be0e537" containerID="7196c7a7960d8de8eb3521c013841d1a17ec1a381a049bdd41937224c8cdd791" exitCode=0 Nov 22 09:50:17 crc kubenswrapper[4789]: I1122 09:50:17.888588 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h76zf" event={"ID":"b4b60a15-7288-4119-aaba-82731be0e537","Type":"ContainerDied","Data":"7196c7a7960d8de8eb3521c013841d1a17ec1a381a049bdd41937224c8cdd791"} Nov 22 09:50:18 crc kubenswrapper[4789]: I1122 09:50:18.901331 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h76zf" event={"ID":"b4b60a15-7288-4119-aaba-82731be0e537","Type":"ContainerStarted","Data":"efffe7bc6ea582cd9da599afe05fbd77c1d4c3357da564e0352d9dfd6d3dde2e"} Nov 22 09:50:18 crc kubenswrapper[4789]: I1122 09:50:18.944001 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-h76zf" podStartSLOduration=3.489698379 podStartE2EDuration="6.943981098s" podCreationTimestamp="2025-11-22 09:50:12 +0000 UTC" firstStartedPulling="2025-11-22 09:50:14.842216088 +0000 UTC m=+6829.076616371" lastFinishedPulling="2025-11-22 09:50:18.296498817 +0000 UTC m=+6832.530899090" observedRunningTime="2025-11-22 09:50:18.932796139 +0000 UTC m=+6833.167196422" watchObservedRunningTime="2025-11-22 09:50:18.943981098 +0000 UTC m=+6833.178381381" Nov 22 09:50:21 crc kubenswrapper[4789]: I1122 09:50:21.932721 4789 generic.go:334] "Generic (PLEG): container finished" podID="7514e0b6-fae4-4419-8e5f-7798884fe6d5" containerID="c93637949a843fcf9578f5209f87d7e8ebe8c382a467bcf07b45cec69a45cef2" exitCode=0 Nov 22 09:50:21 crc kubenswrapper[4789]: I1122 09:50:21.932796 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5dd6b" event={"ID":"7514e0b6-fae4-4419-8e5f-7798884fe6d5","Type":"ContainerDied","Data":"c93637949a843fcf9578f5209f87d7e8ebe8c382a467bcf07b45cec69a45cef2"} Nov 22 09:50:23 crc kubenswrapper[4789]: I1122 09:50:23.162860 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-h76zf" Nov 22 09:50:23 crc kubenswrapper[4789]: I1122 09:50:23.167041 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-h76zf" Nov 22 09:50:23 crc kubenswrapper[4789]: I1122 09:50:23.220708 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-h76zf" Nov 22 09:50:24 crc kubenswrapper[4789]: I1122 09:50:24.029166 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-h76zf" Nov 22 09:50:24 crc kubenswrapper[4789]: I1122 09:50:24.370832 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-h76zf"] Nov 22 09:50:25 crc kubenswrapper[4789]: I1122 09:50:25.981850 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5dd6b" event={"ID":"7514e0b6-fae4-4419-8e5f-7798884fe6d5","Type":"ContainerStarted","Data":"52a2f05ca668d5b2b7064b1fa6591b796addb944751f6d2abb937cc0640e7f42"} Nov 22 09:50:25 crc kubenswrapper[4789]: I1122 09:50:25.981976 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-h76zf" podUID="b4b60a15-7288-4119-aaba-82731be0e537" containerName="registry-server" containerID="cri-o://efffe7bc6ea582cd9da599afe05fbd77c1d4c3357da564e0352d9dfd6d3dde2e" gracePeriod=2 Nov 22 09:50:26 crc kubenswrapper[4789]: I1122 09:50:26.011225 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-5dd6b" podStartSLOduration=2.853442134 podStartE2EDuration="11.011208019s" podCreationTimestamp="2025-11-22 09:50:15 +0000 UTC" firstStartedPulling="2025-11-22 09:50:16.875231775 +0000 UTC m=+6831.109632058" lastFinishedPulling="2025-11-22 09:50:25.03299767 +0000 UTC m=+6839.267397943" observedRunningTime="2025-11-22 09:50:26.007588232 +0000 UTC m=+6840.241988505" watchObservedRunningTime="2025-11-22 09:50:26.011208019 +0000 UTC m=+6840.245608292" Nov 22 09:50:26 crc kubenswrapper[4789]: I1122 09:50:26.693294 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-h76zf" Nov 22 09:50:26 crc kubenswrapper[4789]: I1122 09:50:26.783529 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4b60a15-7288-4119-aaba-82731be0e537-catalog-content\") pod \"b4b60a15-7288-4119-aaba-82731be0e537\" (UID: \"b4b60a15-7288-4119-aaba-82731be0e537\") " Nov 22 09:50:26 crc kubenswrapper[4789]: I1122 09:50:26.783608 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gdv6b\" (UniqueName: \"kubernetes.io/projected/b4b60a15-7288-4119-aaba-82731be0e537-kube-api-access-gdv6b\") pod \"b4b60a15-7288-4119-aaba-82731be0e537\" (UID: \"b4b60a15-7288-4119-aaba-82731be0e537\") " Nov 22 09:50:26 crc kubenswrapper[4789]: I1122 09:50:26.783775 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4b60a15-7288-4119-aaba-82731be0e537-utilities\") pod \"b4b60a15-7288-4119-aaba-82731be0e537\" (UID: \"b4b60a15-7288-4119-aaba-82731be0e537\") " Nov 22 09:50:26 crc kubenswrapper[4789]: I1122 09:50:26.784530 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b4b60a15-7288-4119-aaba-82731be0e537-utilities" (OuterVolumeSpecName: "utilities") pod "b4b60a15-7288-4119-aaba-82731be0e537" (UID: "b4b60a15-7288-4119-aaba-82731be0e537"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 09:50:26 crc kubenswrapper[4789]: I1122 09:50:26.785303 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4b60a15-7288-4119-aaba-82731be0e537-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 09:50:26 crc kubenswrapper[4789]: I1122 09:50:26.792859 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b4b60a15-7288-4119-aaba-82731be0e537-kube-api-access-gdv6b" (OuterVolumeSpecName: "kube-api-access-gdv6b") pod "b4b60a15-7288-4119-aaba-82731be0e537" (UID: "b4b60a15-7288-4119-aaba-82731be0e537"). InnerVolumeSpecName "kube-api-access-gdv6b". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 09:50:26 crc kubenswrapper[4789]: I1122 09:50:26.849659 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b4b60a15-7288-4119-aaba-82731be0e537-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b4b60a15-7288-4119-aaba-82731be0e537" (UID: "b4b60a15-7288-4119-aaba-82731be0e537"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 09:50:26 crc kubenswrapper[4789]: I1122 09:50:26.887401 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4b60a15-7288-4119-aaba-82731be0e537-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 09:50:26 crc kubenswrapper[4789]: I1122 09:50:26.887439 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gdv6b\" (UniqueName: \"kubernetes.io/projected/b4b60a15-7288-4119-aaba-82731be0e537-kube-api-access-gdv6b\") on node \"crc\" DevicePath \"\"" Nov 22 09:50:26 crc kubenswrapper[4789]: I1122 09:50:26.992619 4789 generic.go:334] "Generic (PLEG): container finished" podID="b4b60a15-7288-4119-aaba-82731be0e537" containerID="efffe7bc6ea582cd9da599afe05fbd77c1d4c3357da564e0352d9dfd6d3dde2e" exitCode=0 Nov 22 09:50:26 crc kubenswrapper[4789]: I1122 09:50:26.992657 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h76zf" event={"ID":"b4b60a15-7288-4119-aaba-82731be0e537","Type":"ContainerDied","Data":"efffe7bc6ea582cd9da599afe05fbd77c1d4c3357da564e0352d9dfd6d3dde2e"} Nov 22 09:50:26 crc kubenswrapper[4789]: I1122 09:50:26.992682 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h76zf" event={"ID":"b4b60a15-7288-4119-aaba-82731be0e537","Type":"ContainerDied","Data":"d7105a374257c7cf9fb67ca9a03ce066f6f8c280e7f3496592156e48b92ef844"} Nov 22 09:50:26 crc kubenswrapper[4789]: I1122 09:50:26.992700 4789 scope.go:117] "RemoveContainer" containerID="efffe7bc6ea582cd9da599afe05fbd77c1d4c3357da564e0352d9dfd6d3dde2e" Nov 22 09:50:26 crc kubenswrapper[4789]: I1122 09:50:26.992716 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-h76zf" Nov 22 09:50:27 crc kubenswrapper[4789]: I1122 09:50:27.034923 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-h76zf"] Nov 22 09:50:27 crc kubenswrapper[4789]: I1122 09:50:27.035374 4789 scope.go:117] "RemoveContainer" containerID="7196c7a7960d8de8eb3521c013841d1a17ec1a381a049bdd41937224c8cdd791" Nov 22 09:50:27 crc kubenswrapper[4789]: I1122 09:50:27.043060 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-h76zf"] Nov 22 09:50:27 crc kubenswrapper[4789]: I1122 09:50:27.062967 4789 scope.go:117] "RemoveContainer" containerID="cad7f14ed5296cb87cb0e62dc76c30d127c1b332a1c3d0a9cfa7030e55cceb1a" Nov 22 09:50:27 crc kubenswrapper[4789]: I1122 09:50:27.107840 4789 scope.go:117] "RemoveContainer" containerID="efffe7bc6ea582cd9da599afe05fbd77c1d4c3357da564e0352d9dfd6d3dde2e" Nov 22 09:50:27 crc kubenswrapper[4789]: E1122 09:50:27.108450 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"efffe7bc6ea582cd9da599afe05fbd77c1d4c3357da564e0352d9dfd6d3dde2e\": container with ID starting with efffe7bc6ea582cd9da599afe05fbd77c1d4c3357da564e0352d9dfd6d3dde2e not found: ID does not exist" containerID="efffe7bc6ea582cd9da599afe05fbd77c1d4c3357da564e0352d9dfd6d3dde2e" Nov 22 09:50:27 crc kubenswrapper[4789]: I1122 09:50:27.108491 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"efffe7bc6ea582cd9da599afe05fbd77c1d4c3357da564e0352d9dfd6d3dde2e"} err="failed to get container status \"efffe7bc6ea582cd9da599afe05fbd77c1d4c3357da564e0352d9dfd6d3dde2e\": rpc error: code = NotFound desc = could not find container \"efffe7bc6ea582cd9da599afe05fbd77c1d4c3357da564e0352d9dfd6d3dde2e\": container with ID starting with efffe7bc6ea582cd9da599afe05fbd77c1d4c3357da564e0352d9dfd6d3dde2e not found: ID does not exist" Nov 22 09:50:27 crc kubenswrapper[4789]: I1122 09:50:27.108516 4789 scope.go:117] "RemoveContainer" containerID="7196c7a7960d8de8eb3521c013841d1a17ec1a381a049bdd41937224c8cdd791" Nov 22 09:50:27 crc kubenswrapper[4789]: E1122 09:50:27.109037 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7196c7a7960d8de8eb3521c013841d1a17ec1a381a049bdd41937224c8cdd791\": container with ID starting with 7196c7a7960d8de8eb3521c013841d1a17ec1a381a049bdd41937224c8cdd791 not found: ID does not exist" containerID="7196c7a7960d8de8eb3521c013841d1a17ec1a381a049bdd41937224c8cdd791" Nov 22 09:50:27 crc kubenswrapper[4789]: I1122 09:50:27.109065 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7196c7a7960d8de8eb3521c013841d1a17ec1a381a049bdd41937224c8cdd791"} err="failed to get container status \"7196c7a7960d8de8eb3521c013841d1a17ec1a381a049bdd41937224c8cdd791\": rpc error: code = NotFound desc = could not find container \"7196c7a7960d8de8eb3521c013841d1a17ec1a381a049bdd41937224c8cdd791\": container with ID starting with 7196c7a7960d8de8eb3521c013841d1a17ec1a381a049bdd41937224c8cdd791 not found: ID does not exist" Nov 22 09:50:27 crc kubenswrapper[4789]: I1122 09:50:27.109085 4789 scope.go:117] "RemoveContainer" containerID="cad7f14ed5296cb87cb0e62dc76c30d127c1b332a1c3d0a9cfa7030e55cceb1a" Nov 22 09:50:27 crc kubenswrapper[4789]: E1122 09:50:27.109411 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cad7f14ed5296cb87cb0e62dc76c30d127c1b332a1c3d0a9cfa7030e55cceb1a\": container with ID starting with cad7f14ed5296cb87cb0e62dc76c30d127c1b332a1c3d0a9cfa7030e55cceb1a not found: ID does not exist" containerID="cad7f14ed5296cb87cb0e62dc76c30d127c1b332a1c3d0a9cfa7030e55cceb1a" Nov 22 09:50:27 crc kubenswrapper[4789]: I1122 09:50:27.109459 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cad7f14ed5296cb87cb0e62dc76c30d127c1b332a1c3d0a9cfa7030e55cceb1a"} err="failed to get container status \"cad7f14ed5296cb87cb0e62dc76c30d127c1b332a1c3d0a9cfa7030e55cceb1a\": rpc error: code = NotFound desc = could not find container \"cad7f14ed5296cb87cb0e62dc76c30d127c1b332a1c3d0a9cfa7030e55cceb1a\": container with ID starting with cad7f14ed5296cb87cb0e62dc76c30d127c1b332a1c3d0a9cfa7030e55cceb1a not found: ID does not exist" Nov 22 09:50:27 crc kubenswrapper[4789]: I1122 09:50:27.976400 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b4b60a15-7288-4119-aaba-82731be0e537" path="/var/lib/kubelet/pods/b4b60a15-7288-4119-aaba-82731be0e537/volumes" Nov 22 09:50:35 crc kubenswrapper[4789]: I1122 09:50:35.372899 4789 patch_prober.go:28] interesting pod/machine-config-daemon-pxvrm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 09:50:35 crc kubenswrapper[4789]: I1122 09:50:35.373417 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 09:50:35 crc kubenswrapper[4789]: I1122 09:50:35.699761 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-5dd6b" Nov 22 09:50:35 crc kubenswrapper[4789]: I1122 09:50:35.700156 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-5dd6b" Nov 22 09:50:35 crc kubenswrapper[4789]: I1122 09:50:35.783957 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-5dd6b" Nov 22 09:50:36 crc kubenswrapper[4789]: I1122 09:50:36.125381 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-5dd6b" Nov 22 09:50:36 crc kubenswrapper[4789]: I1122 09:50:36.171647 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5dd6b"] Nov 22 09:50:38 crc kubenswrapper[4789]: I1122 09:50:38.102733 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-5dd6b" podUID="7514e0b6-fae4-4419-8e5f-7798884fe6d5" containerName="registry-server" containerID="cri-o://52a2f05ca668d5b2b7064b1fa6591b796addb944751f6d2abb937cc0640e7f42" gracePeriod=2 Nov 22 09:50:39 crc kubenswrapper[4789]: I1122 09:50:39.144240 4789 generic.go:334] "Generic (PLEG): container finished" podID="7514e0b6-fae4-4419-8e5f-7798884fe6d5" containerID="52a2f05ca668d5b2b7064b1fa6591b796addb944751f6d2abb937cc0640e7f42" exitCode=0 Nov 22 09:50:39 crc kubenswrapper[4789]: I1122 09:50:39.144478 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5dd6b" event={"ID":"7514e0b6-fae4-4419-8e5f-7798884fe6d5","Type":"ContainerDied","Data":"52a2f05ca668d5b2b7064b1fa6591b796addb944751f6d2abb937cc0640e7f42"} Nov 22 09:50:39 crc kubenswrapper[4789]: I1122 09:50:39.204564 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5dd6b" Nov 22 09:50:39 crc kubenswrapper[4789]: I1122 09:50:39.276314 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nbp42\" (UniqueName: \"kubernetes.io/projected/7514e0b6-fae4-4419-8e5f-7798884fe6d5-kube-api-access-nbp42\") pod \"7514e0b6-fae4-4419-8e5f-7798884fe6d5\" (UID: \"7514e0b6-fae4-4419-8e5f-7798884fe6d5\") " Nov 22 09:50:39 crc kubenswrapper[4789]: I1122 09:50:39.276776 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7514e0b6-fae4-4419-8e5f-7798884fe6d5-utilities\") pod \"7514e0b6-fae4-4419-8e5f-7798884fe6d5\" (UID: \"7514e0b6-fae4-4419-8e5f-7798884fe6d5\") " Nov 22 09:50:39 crc kubenswrapper[4789]: I1122 09:50:39.276887 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7514e0b6-fae4-4419-8e5f-7798884fe6d5-catalog-content\") pod \"7514e0b6-fae4-4419-8e5f-7798884fe6d5\" (UID: \"7514e0b6-fae4-4419-8e5f-7798884fe6d5\") " Nov 22 09:50:39 crc kubenswrapper[4789]: I1122 09:50:39.278497 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7514e0b6-fae4-4419-8e5f-7798884fe6d5-utilities" (OuterVolumeSpecName: "utilities") pod "7514e0b6-fae4-4419-8e5f-7798884fe6d5" (UID: "7514e0b6-fae4-4419-8e5f-7798884fe6d5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 09:50:39 crc kubenswrapper[4789]: I1122 09:50:39.291164 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7514e0b6-fae4-4419-8e5f-7798884fe6d5-kube-api-access-nbp42" (OuterVolumeSpecName: "kube-api-access-nbp42") pod "7514e0b6-fae4-4419-8e5f-7798884fe6d5" (UID: "7514e0b6-fae4-4419-8e5f-7798884fe6d5"). InnerVolumeSpecName "kube-api-access-nbp42". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 09:50:39 crc kubenswrapper[4789]: I1122 09:50:39.379867 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nbp42\" (UniqueName: \"kubernetes.io/projected/7514e0b6-fae4-4419-8e5f-7798884fe6d5-kube-api-access-nbp42\") on node \"crc\" DevicePath \"\"" Nov 22 09:50:39 crc kubenswrapper[4789]: I1122 09:50:39.379906 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7514e0b6-fae4-4419-8e5f-7798884fe6d5-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 09:50:39 crc kubenswrapper[4789]: I1122 09:50:39.385268 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7514e0b6-fae4-4419-8e5f-7798884fe6d5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7514e0b6-fae4-4419-8e5f-7798884fe6d5" (UID: "7514e0b6-fae4-4419-8e5f-7798884fe6d5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 09:50:39 crc kubenswrapper[4789]: I1122 09:50:39.482091 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7514e0b6-fae4-4419-8e5f-7798884fe6d5-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 09:50:40 crc kubenswrapper[4789]: I1122 09:50:40.156613 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5dd6b" event={"ID":"7514e0b6-fae4-4419-8e5f-7798884fe6d5","Type":"ContainerDied","Data":"8b855b552af9ea44a75acb028f0799dcce03ae40bfad5a654453b835ff13ce9a"} Nov 22 09:50:40 crc kubenswrapper[4789]: I1122 09:50:40.156876 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5dd6b" Nov 22 09:50:40 crc kubenswrapper[4789]: I1122 09:50:40.157002 4789 scope.go:117] "RemoveContainer" containerID="52a2f05ca668d5b2b7064b1fa6591b796addb944751f6d2abb937cc0640e7f42" Nov 22 09:50:40 crc kubenswrapper[4789]: I1122 09:50:40.187504 4789 scope.go:117] "RemoveContainer" containerID="c93637949a843fcf9578f5209f87d7e8ebe8c382a467bcf07b45cec69a45cef2" Nov 22 09:50:40 crc kubenswrapper[4789]: I1122 09:50:40.195430 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5dd6b"] Nov 22 09:50:40 crc kubenswrapper[4789]: I1122 09:50:40.211483 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-5dd6b"] Nov 22 09:50:40 crc kubenswrapper[4789]: I1122 09:50:40.212700 4789 scope.go:117] "RemoveContainer" containerID="692ddc87bb90852c79f33d81b8a1b71234f38c9c98fcede0d1076bab7c8be7b4" Nov 22 09:50:41 crc kubenswrapper[4789]: I1122 09:50:41.979224 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7514e0b6-fae4-4419-8e5f-7798884fe6d5" path="/var/lib/kubelet/pods/7514e0b6-fae4-4419-8e5f-7798884fe6d5/volumes" Nov 22 09:51:05 crc kubenswrapper[4789]: I1122 09:51:05.371953 4789 patch_prober.go:28] interesting pod/machine-config-daemon-pxvrm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 09:51:05 crc kubenswrapper[4789]: I1122 09:51:05.372526 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 09:51:35 crc kubenswrapper[4789]: I1122 09:51:35.372293 4789 patch_prober.go:28] interesting pod/machine-config-daemon-pxvrm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 09:51:35 crc kubenswrapper[4789]: I1122 09:51:35.373013 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 09:51:35 crc kubenswrapper[4789]: I1122 09:51:35.373088 4789 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" Nov 22 09:51:35 crc kubenswrapper[4789]: I1122 09:51:35.374120 4789 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2e98f18b68c5e7c627f36e814e1953271056ec4c2d4a51091ea26f228a4fa550"} pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 22 09:51:35 crc kubenswrapper[4789]: I1122 09:51:35.374234 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" containerID="cri-o://2e98f18b68c5e7c627f36e814e1953271056ec4c2d4a51091ea26f228a4fa550" gracePeriod=600 Nov 22 09:51:36 crc kubenswrapper[4789]: I1122 09:51:36.171201 4789 generic.go:334] "Generic (PLEG): container finished" podID="58aa8071-7ecc-4692-8789-2db04ec70647" containerID="2e98f18b68c5e7c627f36e814e1953271056ec4c2d4a51091ea26f228a4fa550" exitCode=0 Nov 22 09:51:36 crc kubenswrapper[4789]: I1122 09:51:36.171571 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" event={"ID":"58aa8071-7ecc-4692-8789-2db04ec70647","Type":"ContainerDied","Data":"2e98f18b68c5e7c627f36e814e1953271056ec4c2d4a51091ea26f228a4fa550"} Nov 22 09:51:36 crc kubenswrapper[4789]: I1122 09:51:36.171618 4789 scope.go:117] "RemoveContainer" containerID="4ce6f99fe0532d0774ed3a238dd72bf64bbc780ed6b623e2fc3116d00f8d5b2f" Nov 22 09:51:37 crc kubenswrapper[4789]: I1122 09:51:37.186192 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" event={"ID":"58aa8071-7ecc-4692-8789-2db04ec70647","Type":"ContainerStarted","Data":"26c17a090dfedea5cf0834b08473176e1ddf9f8e1cbf99d58db7701ed655462d"} Nov 22 09:52:33 crc kubenswrapper[4789]: I1122 09:52:33.773990 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-4q89t"] Nov 22 09:52:33 crc kubenswrapper[4789]: E1122 09:52:33.775221 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7514e0b6-fae4-4419-8e5f-7798884fe6d5" containerName="registry-server" Nov 22 09:52:33 crc kubenswrapper[4789]: I1122 09:52:33.775243 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="7514e0b6-fae4-4419-8e5f-7798884fe6d5" containerName="registry-server" Nov 22 09:52:33 crc kubenswrapper[4789]: E1122 09:52:33.775261 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4b60a15-7288-4119-aaba-82731be0e537" containerName="extract-content" Nov 22 09:52:33 crc kubenswrapper[4789]: I1122 09:52:33.775272 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4b60a15-7288-4119-aaba-82731be0e537" containerName="extract-content" Nov 22 09:52:33 crc kubenswrapper[4789]: E1122 09:52:33.775292 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7514e0b6-fae4-4419-8e5f-7798884fe6d5" containerName="extract-content" Nov 22 09:52:33 crc kubenswrapper[4789]: I1122 09:52:33.775305 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="7514e0b6-fae4-4419-8e5f-7798884fe6d5" containerName="extract-content" Nov 22 09:52:33 crc kubenswrapper[4789]: E1122 09:52:33.775329 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4b60a15-7288-4119-aaba-82731be0e537" containerName="extract-utilities" Nov 22 09:52:33 crc kubenswrapper[4789]: I1122 09:52:33.775340 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4b60a15-7288-4119-aaba-82731be0e537" containerName="extract-utilities" Nov 22 09:52:33 crc kubenswrapper[4789]: E1122 09:52:33.775384 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7514e0b6-fae4-4419-8e5f-7798884fe6d5" containerName="extract-utilities" Nov 22 09:52:33 crc kubenswrapper[4789]: I1122 09:52:33.775394 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="7514e0b6-fae4-4419-8e5f-7798884fe6d5" containerName="extract-utilities" Nov 22 09:52:33 crc kubenswrapper[4789]: E1122 09:52:33.775428 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4b60a15-7288-4119-aaba-82731be0e537" containerName="registry-server" Nov 22 09:52:33 crc kubenswrapper[4789]: I1122 09:52:33.775439 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4b60a15-7288-4119-aaba-82731be0e537" containerName="registry-server" Nov 22 09:52:33 crc kubenswrapper[4789]: I1122 09:52:33.775784 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="7514e0b6-fae4-4419-8e5f-7798884fe6d5" containerName="registry-server" Nov 22 09:52:33 crc kubenswrapper[4789]: I1122 09:52:33.775805 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4b60a15-7288-4119-aaba-82731be0e537" containerName="registry-server" Nov 22 09:52:33 crc kubenswrapper[4789]: I1122 09:52:33.778933 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4q89t" Nov 22 09:52:33 crc kubenswrapper[4789]: I1122 09:52:33.793375 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4q89t"] Nov 22 09:52:33 crc kubenswrapper[4789]: I1122 09:52:33.895380 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c6816e11-72c8-489f-8bfa-876158739c75-utilities\") pod \"community-operators-4q89t\" (UID: \"c6816e11-72c8-489f-8bfa-876158739c75\") " pod="openshift-marketplace/community-operators-4q89t" Nov 22 09:52:33 crc kubenswrapper[4789]: I1122 09:52:33.895811 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2mkh7\" (UniqueName: \"kubernetes.io/projected/c6816e11-72c8-489f-8bfa-876158739c75-kube-api-access-2mkh7\") pod \"community-operators-4q89t\" (UID: \"c6816e11-72c8-489f-8bfa-876158739c75\") " pod="openshift-marketplace/community-operators-4q89t" Nov 22 09:52:33 crc kubenswrapper[4789]: I1122 09:52:33.895854 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c6816e11-72c8-489f-8bfa-876158739c75-catalog-content\") pod \"community-operators-4q89t\" (UID: \"c6816e11-72c8-489f-8bfa-876158739c75\") " pod="openshift-marketplace/community-operators-4q89t" Nov 22 09:52:33 crc kubenswrapper[4789]: I1122 09:52:33.997886 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2mkh7\" (UniqueName: \"kubernetes.io/projected/c6816e11-72c8-489f-8bfa-876158739c75-kube-api-access-2mkh7\") pod \"community-operators-4q89t\" (UID: \"c6816e11-72c8-489f-8bfa-876158739c75\") " pod="openshift-marketplace/community-operators-4q89t" Nov 22 09:52:33 crc kubenswrapper[4789]: I1122 09:52:33.997938 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c6816e11-72c8-489f-8bfa-876158739c75-catalog-content\") pod \"community-operators-4q89t\" (UID: \"c6816e11-72c8-489f-8bfa-876158739c75\") " pod="openshift-marketplace/community-operators-4q89t" Nov 22 09:52:33 crc kubenswrapper[4789]: I1122 09:52:33.998094 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c6816e11-72c8-489f-8bfa-876158739c75-utilities\") pod \"community-operators-4q89t\" (UID: \"c6816e11-72c8-489f-8bfa-876158739c75\") " pod="openshift-marketplace/community-operators-4q89t" Nov 22 09:52:33 crc kubenswrapper[4789]: I1122 09:52:33.998700 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c6816e11-72c8-489f-8bfa-876158739c75-catalog-content\") pod \"community-operators-4q89t\" (UID: \"c6816e11-72c8-489f-8bfa-876158739c75\") " pod="openshift-marketplace/community-operators-4q89t" Nov 22 09:52:33 crc kubenswrapper[4789]: I1122 09:52:33.998717 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c6816e11-72c8-489f-8bfa-876158739c75-utilities\") pod \"community-operators-4q89t\" (UID: \"c6816e11-72c8-489f-8bfa-876158739c75\") " pod="openshift-marketplace/community-operators-4q89t" Nov 22 09:52:34 crc kubenswrapper[4789]: I1122 09:52:34.017878 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2mkh7\" (UniqueName: \"kubernetes.io/projected/c6816e11-72c8-489f-8bfa-876158739c75-kube-api-access-2mkh7\") pod \"community-operators-4q89t\" (UID: \"c6816e11-72c8-489f-8bfa-876158739c75\") " pod="openshift-marketplace/community-operators-4q89t" Nov 22 09:52:34 crc kubenswrapper[4789]: I1122 09:52:34.138441 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4q89t" Nov 22 09:52:34 crc kubenswrapper[4789]: I1122 09:52:34.696341 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4q89t"] Nov 22 09:52:34 crc kubenswrapper[4789]: I1122 09:52:34.798038 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4q89t" event={"ID":"c6816e11-72c8-489f-8bfa-876158739c75","Type":"ContainerStarted","Data":"ce94953410ff68f778a388c0443fc5751f5568f9ca6f7d2c26b265a7f7229755"} Nov 22 09:52:35 crc kubenswrapper[4789]: I1122 09:52:35.813007 4789 generic.go:334] "Generic (PLEG): container finished" podID="c6816e11-72c8-489f-8bfa-876158739c75" containerID="8e81d08c8750dcef1c970fe7ea79d591ecc0d4bfe219f844efb644eacca712fb" exitCode=0 Nov 22 09:52:35 crc kubenswrapper[4789]: I1122 09:52:35.813095 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4q89t" event={"ID":"c6816e11-72c8-489f-8bfa-876158739c75","Type":"ContainerDied","Data":"8e81d08c8750dcef1c970fe7ea79d591ecc0d4bfe219f844efb644eacca712fb"} Nov 22 09:52:35 crc kubenswrapper[4789]: I1122 09:52:35.816130 4789 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 22 09:52:36 crc kubenswrapper[4789]: I1122 09:52:36.825112 4789 generic.go:334] "Generic (PLEG): container finished" podID="c6816e11-72c8-489f-8bfa-876158739c75" containerID="4f9906c22abbde4e7319c9ffee45eabbee6ac336564f784b1b24c8e320aa1ba1" exitCode=0 Nov 22 09:52:36 crc kubenswrapper[4789]: I1122 09:52:36.825189 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4q89t" event={"ID":"c6816e11-72c8-489f-8bfa-876158739c75","Type":"ContainerDied","Data":"4f9906c22abbde4e7319c9ffee45eabbee6ac336564f784b1b24c8e320aa1ba1"} Nov 22 09:52:37 crc kubenswrapper[4789]: I1122 09:52:37.842862 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4q89t" event={"ID":"c6816e11-72c8-489f-8bfa-876158739c75","Type":"ContainerStarted","Data":"11c6dcdc0609408df3beaeb2e616c3ecb0b6333a5ebb3c7d928e7ced3de05dd5"} Nov 22 09:52:37 crc kubenswrapper[4789]: I1122 09:52:37.864347 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-4q89t" podStartSLOduration=3.450897419 podStartE2EDuration="4.86433081s" podCreationTimestamp="2025-11-22 09:52:33 +0000 UTC" firstStartedPulling="2025-11-22 09:52:35.815769028 +0000 UTC m=+6970.050169321" lastFinishedPulling="2025-11-22 09:52:37.229202429 +0000 UTC m=+6971.463602712" observedRunningTime="2025-11-22 09:52:37.86172878 +0000 UTC m=+6972.096129073" watchObservedRunningTime="2025-11-22 09:52:37.86433081 +0000 UTC m=+6972.098731083" Nov 22 09:52:44 crc kubenswrapper[4789]: I1122 09:52:44.139109 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-4q89t" Nov 22 09:52:44 crc kubenswrapper[4789]: I1122 09:52:44.139606 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-4q89t" Nov 22 09:52:44 crc kubenswrapper[4789]: I1122 09:52:44.194250 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-4q89t" Nov 22 09:52:44 crc kubenswrapper[4789]: I1122 09:52:44.987817 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-4q89t" Nov 22 09:52:45 crc kubenswrapper[4789]: I1122 09:52:45.055109 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-4q89t"] Nov 22 09:52:46 crc kubenswrapper[4789]: I1122 09:52:46.935904 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-4q89t" podUID="c6816e11-72c8-489f-8bfa-876158739c75" containerName="registry-server" containerID="cri-o://11c6dcdc0609408df3beaeb2e616c3ecb0b6333a5ebb3c7d928e7ced3de05dd5" gracePeriod=2 Nov 22 09:52:47 crc kubenswrapper[4789]: I1122 09:52:47.463886 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4q89t" Nov 22 09:52:47 crc kubenswrapper[4789]: I1122 09:52:47.578566 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c6816e11-72c8-489f-8bfa-876158739c75-catalog-content\") pod \"c6816e11-72c8-489f-8bfa-876158739c75\" (UID: \"c6816e11-72c8-489f-8bfa-876158739c75\") " Nov 22 09:52:47 crc kubenswrapper[4789]: I1122 09:52:47.578624 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c6816e11-72c8-489f-8bfa-876158739c75-utilities\") pod \"c6816e11-72c8-489f-8bfa-876158739c75\" (UID: \"c6816e11-72c8-489f-8bfa-876158739c75\") " Nov 22 09:52:47 crc kubenswrapper[4789]: I1122 09:52:47.578666 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2mkh7\" (UniqueName: \"kubernetes.io/projected/c6816e11-72c8-489f-8bfa-876158739c75-kube-api-access-2mkh7\") pod \"c6816e11-72c8-489f-8bfa-876158739c75\" (UID: \"c6816e11-72c8-489f-8bfa-876158739c75\") " Nov 22 09:52:47 crc kubenswrapper[4789]: I1122 09:52:47.580623 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c6816e11-72c8-489f-8bfa-876158739c75-utilities" (OuterVolumeSpecName: "utilities") pod "c6816e11-72c8-489f-8bfa-876158739c75" (UID: "c6816e11-72c8-489f-8bfa-876158739c75"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 09:52:47 crc kubenswrapper[4789]: I1122 09:52:47.593698 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c6816e11-72c8-489f-8bfa-876158739c75-kube-api-access-2mkh7" (OuterVolumeSpecName: "kube-api-access-2mkh7") pod "c6816e11-72c8-489f-8bfa-876158739c75" (UID: "c6816e11-72c8-489f-8bfa-876158739c75"). InnerVolumeSpecName "kube-api-access-2mkh7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 09:52:47 crc kubenswrapper[4789]: I1122 09:52:47.681224 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c6816e11-72c8-489f-8bfa-876158739c75-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 09:52:47 crc kubenswrapper[4789]: I1122 09:52:47.681258 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2mkh7\" (UniqueName: \"kubernetes.io/projected/c6816e11-72c8-489f-8bfa-876158739c75-kube-api-access-2mkh7\") on node \"crc\" DevicePath \"\"" Nov 22 09:52:47 crc kubenswrapper[4789]: I1122 09:52:47.952524 4789 generic.go:334] "Generic (PLEG): container finished" podID="c6816e11-72c8-489f-8bfa-876158739c75" containerID="11c6dcdc0609408df3beaeb2e616c3ecb0b6333a5ebb3c7d928e7ced3de05dd5" exitCode=0 Nov 22 09:52:47 crc kubenswrapper[4789]: I1122 09:52:47.952607 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4q89t" event={"ID":"c6816e11-72c8-489f-8bfa-876158739c75","Type":"ContainerDied","Data":"11c6dcdc0609408df3beaeb2e616c3ecb0b6333a5ebb3c7d928e7ced3de05dd5"} Nov 22 09:52:47 crc kubenswrapper[4789]: I1122 09:52:47.952812 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4q89t" event={"ID":"c6816e11-72c8-489f-8bfa-876158739c75","Type":"ContainerDied","Data":"ce94953410ff68f778a388c0443fc5751f5568f9ca6f7d2c26b265a7f7229755"} Nov 22 09:52:47 crc kubenswrapper[4789]: I1122 09:52:47.952834 4789 scope.go:117] "RemoveContainer" containerID="11c6dcdc0609408df3beaeb2e616c3ecb0b6333a5ebb3c7d928e7ced3de05dd5" Nov 22 09:52:47 crc kubenswrapper[4789]: I1122 09:52:47.952648 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4q89t" Nov 22 09:52:47 crc kubenswrapper[4789]: I1122 09:52:47.985692 4789 scope.go:117] "RemoveContainer" containerID="4f9906c22abbde4e7319c9ffee45eabbee6ac336564f784b1b24c8e320aa1ba1" Nov 22 09:52:48 crc kubenswrapper[4789]: I1122 09:52:48.010222 4789 scope.go:117] "RemoveContainer" containerID="8e81d08c8750dcef1c970fe7ea79d591ecc0d4bfe219f844efb644eacca712fb" Nov 22 09:52:48 crc kubenswrapper[4789]: I1122 09:52:48.068724 4789 scope.go:117] "RemoveContainer" containerID="11c6dcdc0609408df3beaeb2e616c3ecb0b6333a5ebb3c7d928e7ced3de05dd5" Nov 22 09:52:48 crc kubenswrapper[4789]: E1122 09:52:48.069340 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"11c6dcdc0609408df3beaeb2e616c3ecb0b6333a5ebb3c7d928e7ced3de05dd5\": container with ID starting with 11c6dcdc0609408df3beaeb2e616c3ecb0b6333a5ebb3c7d928e7ced3de05dd5 not found: ID does not exist" containerID="11c6dcdc0609408df3beaeb2e616c3ecb0b6333a5ebb3c7d928e7ced3de05dd5" Nov 22 09:52:48 crc kubenswrapper[4789]: I1122 09:52:48.069401 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11c6dcdc0609408df3beaeb2e616c3ecb0b6333a5ebb3c7d928e7ced3de05dd5"} err="failed to get container status \"11c6dcdc0609408df3beaeb2e616c3ecb0b6333a5ebb3c7d928e7ced3de05dd5\": rpc error: code = NotFound desc = could not find container \"11c6dcdc0609408df3beaeb2e616c3ecb0b6333a5ebb3c7d928e7ced3de05dd5\": container with ID starting with 11c6dcdc0609408df3beaeb2e616c3ecb0b6333a5ebb3c7d928e7ced3de05dd5 not found: ID does not exist" Nov 22 09:52:48 crc kubenswrapper[4789]: I1122 09:52:48.069429 4789 scope.go:117] "RemoveContainer" containerID="4f9906c22abbde4e7319c9ffee45eabbee6ac336564f784b1b24c8e320aa1ba1" Nov 22 09:52:48 crc kubenswrapper[4789]: E1122 09:52:48.069872 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4f9906c22abbde4e7319c9ffee45eabbee6ac336564f784b1b24c8e320aa1ba1\": container with ID starting with 4f9906c22abbde4e7319c9ffee45eabbee6ac336564f784b1b24c8e320aa1ba1 not found: ID does not exist" containerID="4f9906c22abbde4e7319c9ffee45eabbee6ac336564f784b1b24c8e320aa1ba1" Nov 22 09:52:48 crc kubenswrapper[4789]: I1122 09:52:48.069902 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4f9906c22abbde4e7319c9ffee45eabbee6ac336564f784b1b24c8e320aa1ba1"} err="failed to get container status \"4f9906c22abbde4e7319c9ffee45eabbee6ac336564f784b1b24c8e320aa1ba1\": rpc error: code = NotFound desc = could not find container \"4f9906c22abbde4e7319c9ffee45eabbee6ac336564f784b1b24c8e320aa1ba1\": container with ID starting with 4f9906c22abbde4e7319c9ffee45eabbee6ac336564f784b1b24c8e320aa1ba1 not found: ID does not exist" Nov 22 09:52:48 crc kubenswrapper[4789]: I1122 09:52:48.069921 4789 scope.go:117] "RemoveContainer" containerID="8e81d08c8750dcef1c970fe7ea79d591ecc0d4bfe219f844efb644eacca712fb" Nov 22 09:52:48 crc kubenswrapper[4789]: E1122 09:52:48.070365 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8e81d08c8750dcef1c970fe7ea79d591ecc0d4bfe219f844efb644eacca712fb\": container with ID starting with 8e81d08c8750dcef1c970fe7ea79d591ecc0d4bfe219f844efb644eacca712fb not found: ID does not exist" containerID="8e81d08c8750dcef1c970fe7ea79d591ecc0d4bfe219f844efb644eacca712fb" Nov 22 09:52:48 crc kubenswrapper[4789]: I1122 09:52:48.070426 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8e81d08c8750dcef1c970fe7ea79d591ecc0d4bfe219f844efb644eacca712fb"} err="failed to get container status \"8e81d08c8750dcef1c970fe7ea79d591ecc0d4bfe219f844efb644eacca712fb\": rpc error: code = NotFound desc = could not find container \"8e81d08c8750dcef1c970fe7ea79d591ecc0d4bfe219f844efb644eacca712fb\": container with ID starting with 8e81d08c8750dcef1c970fe7ea79d591ecc0d4bfe219f844efb644eacca712fb not found: ID does not exist" Nov 22 09:52:49 crc kubenswrapper[4789]: I1122 09:52:49.025451 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c6816e11-72c8-489f-8bfa-876158739c75-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c6816e11-72c8-489f-8bfa-876158739c75" (UID: "c6816e11-72c8-489f-8bfa-876158739c75"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 09:52:49 crc kubenswrapper[4789]: I1122 09:52:49.111028 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c6816e11-72c8-489f-8bfa-876158739c75-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 09:52:49 crc kubenswrapper[4789]: I1122 09:52:49.190962 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-4q89t"] Nov 22 09:52:49 crc kubenswrapper[4789]: I1122 09:52:49.203066 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-4q89t"] Nov 22 09:52:49 crc kubenswrapper[4789]: I1122 09:52:49.975282 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c6816e11-72c8-489f-8bfa-876158739c75" path="/var/lib/kubelet/pods/c6816e11-72c8-489f-8bfa-876158739c75/volumes" Nov 22 09:54:05 crc kubenswrapper[4789]: I1122 09:54:05.372559 4789 patch_prober.go:28] interesting pod/machine-config-daemon-pxvrm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 09:54:05 crc kubenswrapper[4789]: I1122 09:54:05.373151 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 09:54:35 crc kubenswrapper[4789]: I1122 09:54:35.372159 4789 patch_prober.go:28] interesting pod/machine-config-daemon-pxvrm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 09:54:35 crc kubenswrapper[4789]: I1122 09:54:35.373113 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 09:55:05 crc kubenswrapper[4789]: I1122 09:55:05.372149 4789 patch_prober.go:28] interesting pod/machine-config-daemon-pxvrm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 09:55:05 crc kubenswrapper[4789]: I1122 09:55:05.373029 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 09:55:05 crc kubenswrapper[4789]: I1122 09:55:05.373099 4789 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" Nov 22 09:55:05 crc kubenswrapper[4789]: I1122 09:55:05.373998 4789 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"26c17a090dfedea5cf0834b08473176e1ddf9f8e1cbf99d58db7701ed655462d"} pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 22 09:55:05 crc kubenswrapper[4789]: I1122 09:55:05.374094 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" containerID="cri-o://26c17a090dfedea5cf0834b08473176e1ddf9f8e1cbf99d58db7701ed655462d" gracePeriod=600 Nov 22 09:55:06 crc kubenswrapper[4789]: E1122 09:55:06.080510 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:55:06 crc kubenswrapper[4789]: I1122 09:55:06.268008 4789 generic.go:334] "Generic (PLEG): container finished" podID="58aa8071-7ecc-4692-8789-2db04ec70647" containerID="26c17a090dfedea5cf0834b08473176e1ddf9f8e1cbf99d58db7701ed655462d" exitCode=0 Nov 22 09:55:06 crc kubenswrapper[4789]: I1122 09:55:06.268037 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" event={"ID":"58aa8071-7ecc-4692-8789-2db04ec70647","Type":"ContainerDied","Data":"26c17a090dfedea5cf0834b08473176e1ddf9f8e1cbf99d58db7701ed655462d"} Nov 22 09:55:06 crc kubenswrapper[4789]: I1122 09:55:06.268140 4789 scope.go:117] "RemoveContainer" containerID="2e98f18b68c5e7c627f36e814e1953271056ec4c2d4a51091ea26f228a4fa550" Nov 22 09:55:06 crc kubenswrapper[4789]: I1122 09:55:06.269077 4789 scope.go:117] "RemoveContainer" containerID="26c17a090dfedea5cf0834b08473176e1ddf9f8e1cbf99d58db7701ed655462d" Nov 22 09:55:06 crc kubenswrapper[4789]: E1122 09:55:06.269563 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:55:17 crc kubenswrapper[4789]: I1122 09:55:17.973461 4789 scope.go:117] "RemoveContainer" containerID="26c17a090dfedea5cf0834b08473176e1ddf9f8e1cbf99d58db7701ed655462d" Nov 22 09:55:17 crc kubenswrapper[4789]: E1122 09:55:17.975114 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:55:29 crc kubenswrapper[4789]: I1122 09:55:29.965573 4789 scope.go:117] "RemoveContainer" containerID="26c17a090dfedea5cf0834b08473176e1ddf9f8e1cbf99d58db7701ed655462d" Nov 22 09:55:29 crc kubenswrapper[4789]: E1122 09:55:29.967300 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:55:44 crc kubenswrapper[4789]: I1122 09:55:44.965016 4789 scope.go:117] "RemoveContainer" containerID="26c17a090dfedea5cf0834b08473176e1ddf9f8e1cbf99d58db7701ed655462d" Nov 22 09:55:44 crc kubenswrapper[4789]: E1122 09:55:44.965952 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:55:59 crc kubenswrapper[4789]: I1122 09:55:59.966720 4789 scope.go:117] "RemoveContainer" containerID="26c17a090dfedea5cf0834b08473176e1ddf9f8e1cbf99d58db7701ed655462d" Nov 22 09:55:59 crc kubenswrapper[4789]: E1122 09:55:59.967843 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:56:12 crc kubenswrapper[4789]: I1122 09:56:12.966322 4789 scope.go:117] "RemoveContainer" containerID="26c17a090dfedea5cf0834b08473176e1ddf9f8e1cbf99d58db7701ed655462d" Nov 22 09:56:12 crc kubenswrapper[4789]: E1122 09:56:12.967666 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:56:27 crc kubenswrapper[4789]: I1122 09:56:27.978634 4789 scope.go:117] "RemoveContainer" containerID="26c17a090dfedea5cf0834b08473176e1ddf9f8e1cbf99d58db7701ed655462d" Nov 22 09:56:27 crc kubenswrapper[4789]: E1122 09:56:27.979734 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:56:39 crc kubenswrapper[4789]: I1122 09:56:39.966633 4789 scope.go:117] "RemoveContainer" containerID="26c17a090dfedea5cf0834b08473176e1ddf9f8e1cbf99d58db7701ed655462d" Nov 22 09:56:39 crc kubenswrapper[4789]: E1122 09:56:39.967603 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:56:52 crc kubenswrapper[4789]: I1122 09:56:52.965379 4789 scope.go:117] "RemoveContainer" containerID="26c17a090dfedea5cf0834b08473176e1ddf9f8e1cbf99d58db7701ed655462d" Nov 22 09:56:52 crc kubenswrapper[4789]: E1122 09:56:52.966396 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:57:06 crc kubenswrapper[4789]: I1122 09:57:06.966652 4789 scope.go:117] "RemoveContainer" containerID="26c17a090dfedea5cf0834b08473176e1ddf9f8e1cbf99d58db7701ed655462d" Nov 22 09:57:06 crc kubenswrapper[4789]: E1122 09:57:06.968016 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:57:17 crc kubenswrapper[4789]: I1122 09:57:17.971614 4789 scope.go:117] "RemoveContainer" containerID="26c17a090dfedea5cf0834b08473176e1ddf9f8e1cbf99d58db7701ed655462d" Nov 22 09:57:17 crc kubenswrapper[4789]: E1122 09:57:17.972408 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:57:29 crc kubenswrapper[4789]: I1122 09:57:29.128745 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-46d6n"] Nov 22 09:57:29 crc kubenswrapper[4789]: E1122 09:57:29.131640 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6816e11-72c8-489f-8bfa-876158739c75" containerName="registry-server" Nov 22 09:57:29 crc kubenswrapper[4789]: I1122 09:57:29.131662 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6816e11-72c8-489f-8bfa-876158739c75" containerName="registry-server" Nov 22 09:57:29 crc kubenswrapper[4789]: E1122 09:57:29.131691 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6816e11-72c8-489f-8bfa-876158739c75" containerName="extract-content" Nov 22 09:57:29 crc kubenswrapper[4789]: I1122 09:57:29.131700 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6816e11-72c8-489f-8bfa-876158739c75" containerName="extract-content" Nov 22 09:57:29 crc kubenswrapper[4789]: E1122 09:57:29.131713 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6816e11-72c8-489f-8bfa-876158739c75" containerName="extract-utilities" Nov 22 09:57:29 crc kubenswrapper[4789]: I1122 09:57:29.131723 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6816e11-72c8-489f-8bfa-876158739c75" containerName="extract-utilities" Nov 22 09:57:29 crc kubenswrapper[4789]: I1122 09:57:29.132003 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6816e11-72c8-489f-8bfa-876158739c75" containerName="registry-server" Nov 22 09:57:29 crc kubenswrapper[4789]: I1122 09:57:29.133648 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-46d6n" Nov 22 09:57:29 crc kubenswrapper[4789]: I1122 09:57:29.143475 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-46d6n"] Nov 22 09:57:29 crc kubenswrapper[4789]: I1122 09:57:29.207380 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e4e7e61c-b105-4bef-9ee3-bd75eb38f1f8-utilities\") pod \"redhat-marketplace-46d6n\" (UID: \"e4e7e61c-b105-4bef-9ee3-bd75eb38f1f8\") " pod="openshift-marketplace/redhat-marketplace-46d6n" Nov 22 09:57:29 crc kubenswrapper[4789]: I1122 09:57:29.207519 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e4e7e61c-b105-4bef-9ee3-bd75eb38f1f8-catalog-content\") pod \"redhat-marketplace-46d6n\" (UID: \"e4e7e61c-b105-4bef-9ee3-bd75eb38f1f8\") " pod="openshift-marketplace/redhat-marketplace-46d6n" Nov 22 09:57:29 crc kubenswrapper[4789]: I1122 09:57:29.207739 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-66p7z\" (UniqueName: \"kubernetes.io/projected/e4e7e61c-b105-4bef-9ee3-bd75eb38f1f8-kube-api-access-66p7z\") pod \"redhat-marketplace-46d6n\" (UID: \"e4e7e61c-b105-4bef-9ee3-bd75eb38f1f8\") " pod="openshift-marketplace/redhat-marketplace-46d6n" Nov 22 09:57:29 crc kubenswrapper[4789]: I1122 09:57:29.310529 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e4e7e61c-b105-4bef-9ee3-bd75eb38f1f8-utilities\") pod \"redhat-marketplace-46d6n\" (UID: \"e4e7e61c-b105-4bef-9ee3-bd75eb38f1f8\") " pod="openshift-marketplace/redhat-marketplace-46d6n" Nov 22 09:57:29 crc kubenswrapper[4789]: I1122 09:57:29.310574 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e4e7e61c-b105-4bef-9ee3-bd75eb38f1f8-catalog-content\") pod \"redhat-marketplace-46d6n\" (UID: \"e4e7e61c-b105-4bef-9ee3-bd75eb38f1f8\") " pod="openshift-marketplace/redhat-marketplace-46d6n" Nov 22 09:57:29 crc kubenswrapper[4789]: I1122 09:57:29.310623 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-66p7z\" (UniqueName: \"kubernetes.io/projected/e4e7e61c-b105-4bef-9ee3-bd75eb38f1f8-kube-api-access-66p7z\") pod \"redhat-marketplace-46d6n\" (UID: \"e4e7e61c-b105-4bef-9ee3-bd75eb38f1f8\") " pod="openshift-marketplace/redhat-marketplace-46d6n" Nov 22 09:57:29 crc kubenswrapper[4789]: I1122 09:57:29.311061 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e4e7e61c-b105-4bef-9ee3-bd75eb38f1f8-utilities\") pod \"redhat-marketplace-46d6n\" (UID: \"e4e7e61c-b105-4bef-9ee3-bd75eb38f1f8\") " pod="openshift-marketplace/redhat-marketplace-46d6n" Nov 22 09:57:29 crc kubenswrapper[4789]: I1122 09:57:29.311336 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e4e7e61c-b105-4bef-9ee3-bd75eb38f1f8-catalog-content\") pod \"redhat-marketplace-46d6n\" (UID: \"e4e7e61c-b105-4bef-9ee3-bd75eb38f1f8\") " pod="openshift-marketplace/redhat-marketplace-46d6n" Nov 22 09:57:29 crc kubenswrapper[4789]: I1122 09:57:29.332246 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-66p7z\" (UniqueName: \"kubernetes.io/projected/e4e7e61c-b105-4bef-9ee3-bd75eb38f1f8-kube-api-access-66p7z\") pod \"redhat-marketplace-46d6n\" (UID: \"e4e7e61c-b105-4bef-9ee3-bd75eb38f1f8\") " pod="openshift-marketplace/redhat-marketplace-46d6n" Nov 22 09:57:29 crc kubenswrapper[4789]: I1122 09:57:29.458464 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-46d6n" Nov 22 09:57:29 crc kubenswrapper[4789]: I1122 09:57:29.940055 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-46d6n"] Nov 22 09:57:30 crc kubenswrapper[4789]: I1122 09:57:30.900104 4789 generic.go:334] "Generic (PLEG): container finished" podID="e4e7e61c-b105-4bef-9ee3-bd75eb38f1f8" containerID="0510eb12299f6480d3465f08cdcb106c1b25bf0f526e3d65ed56917f029c5fec" exitCode=0 Nov 22 09:57:30 crc kubenswrapper[4789]: I1122 09:57:30.900168 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-46d6n" event={"ID":"e4e7e61c-b105-4bef-9ee3-bd75eb38f1f8","Type":"ContainerDied","Data":"0510eb12299f6480d3465f08cdcb106c1b25bf0f526e3d65ed56917f029c5fec"} Nov 22 09:57:30 crc kubenswrapper[4789]: I1122 09:57:30.900402 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-46d6n" event={"ID":"e4e7e61c-b105-4bef-9ee3-bd75eb38f1f8","Type":"ContainerStarted","Data":"66aecf5ef001080a1ca379472e786888c5c7ae517e3f87f59b16a0bdd4df41e6"} Nov 22 09:57:30 crc kubenswrapper[4789]: I1122 09:57:30.965726 4789 scope.go:117] "RemoveContainer" containerID="26c17a090dfedea5cf0834b08473176e1ddf9f8e1cbf99d58db7701ed655462d" Nov 22 09:57:30 crc kubenswrapper[4789]: E1122 09:57:30.966339 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:57:31 crc kubenswrapper[4789]: I1122 09:57:31.910956 4789 generic.go:334] "Generic (PLEG): container finished" podID="e4e7e61c-b105-4bef-9ee3-bd75eb38f1f8" containerID="39bc3bee5ad8bb12b54852957fd7916987b8041227eb33456001e920e220e0ac" exitCode=0 Nov 22 09:57:31 crc kubenswrapper[4789]: I1122 09:57:31.911100 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-46d6n" event={"ID":"e4e7e61c-b105-4bef-9ee3-bd75eb38f1f8","Type":"ContainerDied","Data":"39bc3bee5ad8bb12b54852957fd7916987b8041227eb33456001e920e220e0ac"} Nov 22 09:57:32 crc kubenswrapper[4789]: I1122 09:57:32.921293 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-46d6n" event={"ID":"e4e7e61c-b105-4bef-9ee3-bd75eb38f1f8","Type":"ContainerStarted","Data":"065529c5427743a34dfe480820e770fe1f4ca50c05aef5a33cd4e92a0007919e"} Nov 22 09:57:32 crc kubenswrapper[4789]: I1122 09:57:32.955703 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-46d6n" podStartSLOduration=2.31044437 podStartE2EDuration="3.955687649s" podCreationTimestamp="2025-11-22 09:57:29 +0000 UTC" firstStartedPulling="2025-11-22 09:57:30.902666879 +0000 UTC m=+7265.137067192" lastFinishedPulling="2025-11-22 09:57:32.547910198 +0000 UTC m=+7266.782310471" observedRunningTime="2025-11-22 09:57:32.947503059 +0000 UTC m=+7267.181903332" watchObservedRunningTime="2025-11-22 09:57:32.955687649 +0000 UTC m=+7267.190087922" Nov 22 09:57:39 crc kubenswrapper[4789]: I1122 09:57:39.459293 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-46d6n" Nov 22 09:57:39 crc kubenswrapper[4789]: I1122 09:57:39.459976 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-46d6n" Nov 22 09:57:39 crc kubenswrapper[4789]: I1122 09:57:39.512595 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-46d6n" Nov 22 09:57:40 crc kubenswrapper[4789]: I1122 09:57:40.053244 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-46d6n" Nov 22 09:57:40 crc kubenswrapper[4789]: I1122 09:57:40.103274 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-46d6n"] Nov 22 09:57:42 crc kubenswrapper[4789]: I1122 09:57:42.016488 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-46d6n" podUID="e4e7e61c-b105-4bef-9ee3-bd75eb38f1f8" containerName="registry-server" containerID="cri-o://065529c5427743a34dfe480820e770fe1f4ca50c05aef5a33cd4e92a0007919e" gracePeriod=2 Nov 22 09:57:43 crc kubenswrapper[4789]: I1122 09:57:43.029217 4789 generic.go:334] "Generic (PLEG): container finished" podID="e4e7e61c-b105-4bef-9ee3-bd75eb38f1f8" containerID="065529c5427743a34dfe480820e770fe1f4ca50c05aef5a33cd4e92a0007919e" exitCode=0 Nov 22 09:57:43 crc kubenswrapper[4789]: I1122 09:57:43.029275 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-46d6n" event={"ID":"e4e7e61c-b105-4bef-9ee3-bd75eb38f1f8","Type":"ContainerDied","Data":"065529c5427743a34dfe480820e770fe1f4ca50c05aef5a33cd4e92a0007919e"} Nov 22 09:57:43 crc kubenswrapper[4789]: I1122 09:57:43.965396 4789 scope.go:117] "RemoveContainer" containerID="26c17a090dfedea5cf0834b08473176e1ddf9f8e1cbf99d58db7701ed655462d" Nov 22 09:57:43 crc kubenswrapper[4789]: E1122 09:57:43.965780 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:57:44 crc kubenswrapper[4789]: I1122 09:57:44.457998 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-46d6n" Nov 22 09:57:44 crc kubenswrapper[4789]: I1122 09:57:44.628152 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e4e7e61c-b105-4bef-9ee3-bd75eb38f1f8-utilities\") pod \"e4e7e61c-b105-4bef-9ee3-bd75eb38f1f8\" (UID: \"e4e7e61c-b105-4bef-9ee3-bd75eb38f1f8\") " Nov 22 09:57:44 crc kubenswrapper[4789]: I1122 09:57:44.628892 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e4e7e61c-b105-4bef-9ee3-bd75eb38f1f8-catalog-content\") pod \"e4e7e61c-b105-4bef-9ee3-bd75eb38f1f8\" (UID: \"e4e7e61c-b105-4bef-9ee3-bd75eb38f1f8\") " Nov 22 09:57:44 crc kubenswrapper[4789]: I1122 09:57:44.629008 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-66p7z\" (UniqueName: \"kubernetes.io/projected/e4e7e61c-b105-4bef-9ee3-bd75eb38f1f8-kube-api-access-66p7z\") pod \"e4e7e61c-b105-4bef-9ee3-bd75eb38f1f8\" (UID: \"e4e7e61c-b105-4bef-9ee3-bd75eb38f1f8\") " Nov 22 09:57:44 crc kubenswrapper[4789]: I1122 09:57:44.630164 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e4e7e61c-b105-4bef-9ee3-bd75eb38f1f8-utilities" (OuterVolumeSpecName: "utilities") pod "e4e7e61c-b105-4bef-9ee3-bd75eb38f1f8" (UID: "e4e7e61c-b105-4bef-9ee3-bd75eb38f1f8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 09:57:44 crc kubenswrapper[4789]: I1122 09:57:44.646095 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e4e7e61c-b105-4bef-9ee3-bd75eb38f1f8-kube-api-access-66p7z" (OuterVolumeSpecName: "kube-api-access-66p7z") pod "e4e7e61c-b105-4bef-9ee3-bd75eb38f1f8" (UID: "e4e7e61c-b105-4bef-9ee3-bd75eb38f1f8"). InnerVolumeSpecName "kube-api-access-66p7z". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 09:57:44 crc kubenswrapper[4789]: I1122 09:57:44.729675 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e4e7e61c-b105-4bef-9ee3-bd75eb38f1f8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e4e7e61c-b105-4bef-9ee3-bd75eb38f1f8" (UID: "e4e7e61c-b105-4bef-9ee3-bd75eb38f1f8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 09:57:44 crc kubenswrapper[4789]: I1122 09:57:44.731657 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e4e7e61c-b105-4bef-9ee3-bd75eb38f1f8-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 09:57:44 crc kubenswrapper[4789]: I1122 09:57:44.731725 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-66p7z\" (UniqueName: \"kubernetes.io/projected/e4e7e61c-b105-4bef-9ee3-bd75eb38f1f8-kube-api-access-66p7z\") on node \"crc\" DevicePath \"\"" Nov 22 09:57:44 crc kubenswrapper[4789]: I1122 09:57:44.731742 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e4e7e61c-b105-4bef-9ee3-bd75eb38f1f8-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 09:57:45 crc kubenswrapper[4789]: I1122 09:57:45.056330 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-46d6n" event={"ID":"e4e7e61c-b105-4bef-9ee3-bd75eb38f1f8","Type":"ContainerDied","Data":"66aecf5ef001080a1ca379472e786888c5c7ae517e3f87f59b16a0bdd4df41e6"} Nov 22 09:57:45 crc kubenswrapper[4789]: I1122 09:57:45.056401 4789 scope.go:117] "RemoveContainer" containerID="065529c5427743a34dfe480820e770fe1f4ca50c05aef5a33cd4e92a0007919e" Nov 22 09:57:45 crc kubenswrapper[4789]: I1122 09:57:45.056433 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-46d6n" Nov 22 09:57:45 crc kubenswrapper[4789]: I1122 09:57:45.079590 4789 scope.go:117] "RemoveContainer" containerID="39bc3bee5ad8bb12b54852957fd7916987b8041227eb33456001e920e220e0ac" Nov 22 09:57:45 crc kubenswrapper[4789]: I1122 09:57:45.101290 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-46d6n"] Nov 22 09:57:45 crc kubenswrapper[4789]: I1122 09:57:45.113067 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-46d6n"] Nov 22 09:57:45 crc kubenswrapper[4789]: I1122 09:57:45.120649 4789 scope.go:117] "RemoveContainer" containerID="0510eb12299f6480d3465f08cdcb106c1b25bf0f526e3d65ed56917f029c5fec" Nov 22 09:57:45 crc kubenswrapper[4789]: I1122 09:57:45.974956 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e4e7e61c-b105-4bef-9ee3-bd75eb38f1f8" path="/var/lib/kubelet/pods/e4e7e61c-b105-4bef-9ee3-bd75eb38f1f8/volumes" Nov 22 09:57:58 crc kubenswrapper[4789]: I1122 09:57:58.966471 4789 scope.go:117] "RemoveContainer" containerID="26c17a090dfedea5cf0834b08473176e1ddf9f8e1cbf99d58db7701ed655462d" Nov 22 09:57:58 crc kubenswrapper[4789]: E1122 09:57:58.967404 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:58:09 crc kubenswrapper[4789]: I1122 09:58:09.965564 4789 scope.go:117] "RemoveContainer" containerID="26c17a090dfedea5cf0834b08473176e1ddf9f8e1cbf99d58db7701ed655462d" Nov 22 09:58:09 crc kubenswrapper[4789]: E1122 09:58:09.966290 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:58:22 crc kubenswrapper[4789]: I1122 09:58:22.965599 4789 scope.go:117] "RemoveContainer" containerID="26c17a090dfedea5cf0834b08473176e1ddf9f8e1cbf99d58db7701ed655462d" Nov 22 09:58:22 crc kubenswrapper[4789]: E1122 09:58:22.967023 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:58:34 crc kubenswrapper[4789]: I1122 09:58:34.965601 4789 scope.go:117] "RemoveContainer" containerID="26c17a090dfedea5cf0834b08473176e1ddf9f8e1cbf99d58db7701ed655462d" Nov 22 09:58:34 crc kubenswrapper[4789]: E1122 09:58:34.966496 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:58:45 crc kubenswrapper[4789]: I1122 09:58:45.965497 4789 scope.go:117] "RemoveContainer" containerID="26c17a090dfedea5cf0834b08473176e1ddf9f8e1cbf99d58db7701ed655462d" Nov 22 09:58:45 crc kubenswrapper[4789]: E1122 09:58:45.966469 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:58:56 crc kubenswrapper[4789]: I1122 09:58:56.966191 4789 scope.go:117] "RemoveContainer" containerID="26c17a090dfedea5cf0834b08473176e1ddf9f8e1cbf99d58db7701ed655462d" Nov 22 09:58:56 crc kubenswrapper[4789]: E1122 09:58:56.967462 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:59:10 crc kubenswrapper[4789]: I1122 09:59:10.965289 4789 scope.go:117] "RemoveContainer" containerID="26c17a090dfedea5cf0834b08473176e1ddf9f8e1cbf99d58db7701ed655462d" Nov 22 09:59:10 crc kubenswrapper[4789]: E1122 09:59:10.966247 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:59:19 crc kubenswrapper[4789]: I1122 09:59:19.010316 4789 generic.go:334] "Generic (PLEG): container finished" podID="18bbf966-a509-4faa-a8de-28045e5c01b4" containerID="629153cbbe0e8ed9142a669788309296f3b1bcc479a907dbaa59d0624a764760" exitCode=1 Nov 22 09:59:19 crc kubenswrapper[4789]: I1122 09:59:19.010398 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest-s00-full" event={"ID":"18bbf966-a509-4faa-a8de-28045e5c01b4","Type":"ContainerDied","Data":"629153cbbe0e8ed9142a669788309296f3b1bcc479a907dbaa59d0624a764760"} Nov 22 09:59:20 crc kubenswrapper[4789]: I1122 09:59:20.778634 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest-s00-full" Nov 22 09:59:20 crc kubenswrapper[4789]: I1122 09:59:20.872543 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/18bbf966-a509-4faa-a8de-28045e5c01b4-ceph\") pod \"18bbf966-a509-4faa-a8de-28045e5c01b4\" (UID: \"18bbf966-a509-4faa-a8de-28045e5c01b4\") " Nov 22 09:59:20 crc kubenswrapper[4789]: I1122 09:59:20.872589 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"18bbf966-a509-4faa-a8de-28045e5c01b4\" (UID: \"18bbf966-a509-4faa-a8de-28045e5c01b4\") " Nov 22 09:59:20 crc kubenswrapper[4789]: I1122 09:59:20.872612 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/18bbf966-a509-4faa-a8de-28045e5c01b4-test-operator-ephemeral-workdir\") pod \"18bbf966-a509-4faa-a8de-28045e5c01b4\" (UID: \"18bbf966-a509-4faa-a8de-28045e5c01b4\") " Nov 22 09:59:20 crc kubenswrapper[4789]: I1122 09:59:20.872658 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/18bbf966-a509-4faa-a8de-28045e5c01b4-ssh-key\") pod \"18bbf966-a509-4faa-a8de-28045e5c01b4\" (UID: \"18bbf966-a509-4faa-a8de-28045e5c01b4\") " Nov 22 09:59:20 crc kubenswrapper[4789]: I1122 09:59:20.872727 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/18bbf966-a509-4faa-a8de-28045e5c01b4-test-operator-ephemeral-temporary\") pod \"18bbf966-a509-4faa-a8de-28045e5c01b4\" (UID: \"18bbf966-a509-4faa-a8de-28045e5c01b4\") " Nov 22 09:59:20 crc kubenswrapper[4789]: I1122 09:59:20.872814 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/18bbf966-a509-4faa-a8de-28045e5c01b4-ca-certs\") pod \"18bbf966-a509-4faa-a8de-28045e5c01b4\" (UID: \"18bbf966-a509-4faa-a8de-28045e5c01b4\") " Nov 22 09:59:20 crc kubenswrapper[4789]: I1122 09:59:20.872886 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/18bbf966-a509-4faa-a8de-28045e5c01b4-openstack-config\") pod \"18bbf966-a509-4faa-a8de-28045e5c01b4\" (UID: \"18bbf966-a509-4faa-a8de-28045e5c01b4\") " Nov 22 09:59:20 crc kubenswrapper[4789]: I1122 09:59:20.872931 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/18bbf966-a509-4faa-a8de-28045e5c01b4-openstack-config-secret\") pod \"18bbf966-a509-4faa-a8de-28045e5c01b4\" (UID: \"18bbf966-a509-4faa-a8de-28045e5c01b4\") " Nov 22 09:59:20 crc kubenswrapper[4789]: I1122 09:59:20.872961 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/18bbf966-a509-4faa-a8de-28045e5c01b4-config-data\") pod \"18bbf966-a509-4faa-a8de-28045e5c01b4\" (UID: \"18bbf966-a509-4faa-a8de-28045e5c01b4\") " Nov 22 09:59:20 crc kubenswrapper[4789]: I1122 09:59:20.872988 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-77qg6\" (UniqueName: \"kubernetes.io/projected/18bbf966-a509-4faa-a8de-28045e5c01b4-kube-api-access-77qg6\") pod \"18bbf966-a509-4faa-a8de-28045e5c01b4\" (UID: \"18bbf966-a509-4faa-a8de-28045e5c01b4\") " Nov 22 09:59:20 crc kubenswrapper[4789]: I1122 09:59:20.879493 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/18bbf966-a509-4faa-a8de-28045e5c01b4-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "18bbf966-a509-4faa-a8de-28045e5c01b4" (UID: "18bbf966-a509-4faa-a8de-28045e5c01b4"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 09:59:20 crc kubenswrapper[4789]: I1122 09:59:20.886687 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/18bbf966-a509-4faa-a8de-28045e5c01b4-config-data" (OuterVolumeSpecName: "config-data") pod "18bbf966-a509-4faa-a8de-28045e5c01b4" (UID: "18bbf966-a509-4faa-a8de-28045e5c01b4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 09:59:20 crc kubenswrapper[4789]: I1122 09:59:20.886915 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18bbf966-a509-4faa-a8de-28045e5c01b4-kube-api-access-77qg6" (OuterVolumeSpecName: "kube-api-access-77qg6") pod "18bbf966-a509-4faa-a8de-28045e5c01b4" (UID: "18bbf966-a509-4faa-a8de-28045e5c01b4"). InnerVolumeSpecName "kube-api-access-77qg6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 09:59:20 crc kubenswrapper[4789]: I1122 09:59:20.891258 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/18bbf966-a509-4faa-a8de-28045e5c01b4-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "18bbf966-a509-4faa-a8de-28045e5c01b4" (UID: "18bbf966-a509-4faa-a8de-28045e5c01b4"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 09:59:20 crc kubenswrapper[4789]: I1122 09:59:20.897957 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "test-operator-logs") pod "18bbf966-a509-4faa-a8de-28045e5c01b4" (UID: "18bbf966-a509-4faa-a8de-28045e5c01b4"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 22 09:59:20 crc kubenswrapper[4789]: I1122 09:59:20.915854 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18bbf966-a509-4faa-a8de-28045e5c01b4-ceph" (OuterVolumeSpecName: "ceph") pod "18bbf966-a509-4faa-a8de-28045e5c01b4" (UID: "18bbf966-a509-4faa-a8de-28045e5c01b4"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 09:59:20 crc kubenswrapper[4789]: I1122 09:59:20.929300 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest-s01-single-test"] Nov 22 09:59:20 crc kubenswrapper[4789]: E1122 09:59:20.929796 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4e7e61c-b105-4bef-9ee3-bd75eb38f1f8" containerName="extract-content" Nov 22 09:59:20 crc kubenswrapper[4789]: I1122 09:59:20.929818 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4e7e61c-b105-4bef-9ee3-bd75eb38f1f8" containerName="extract-content" Nov 22 09:59:20 crc kubenswrapper[4789]: E1122 09:59:20.929838 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18bbf966-a509-4faa-a8de-28045e5c01b4" containerName="tempest-tests-tempest-tests-runner" Nov 22 09:59:20 crc kubenswrapper[4789]: I1122 09:59:20.929846 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="18bbf966-a509-4faa-a8de-28045e5c01b4" containerName="tempest-tests-tempest-tests-runner" Nov 22 09:59:20 crc kubenswrapper[4789]: E1122 09:59:20.929857 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4e7e61c-b105-4bef-9ee3-bd75eb38f1f8" containerName="registry-server" Nov 22 09:59:20 crc kubenswrapper[4789]: I1122 09:59:20.929865 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4e7e61c-b105-4bef-9ee3-bd75eb38f1f8" containerName="registry-server" Nov 22 09:59:20 crc kubenswrapper[4789]: E1122 09:59:20.929905 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4e7e61c-b105-4bef-9ee3-bd75eb38f1f8" containerName="extract-utilities" Nov 22 09:59:20 crc kubenswrapper[4789]: I1122 09:59:20.929913 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4e7e61c-b105-4bef-9ee3-bd75eb38f1f8" containerName="extract-utilities" Nov 22 09:59:20 crc kubenswrapper[4789]: I1122 09:59:20.930149 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4e7e61c-b105-4bef-9ee3-bd75eb38f1f8" containerName="registry-server" Nov 22 09:59:20 crc kubenswrapper[4789]: I1122 09:59:20.930169 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="18bbf966-a509-4faa-a8de-28045e5c01b4" containerName="tempest-tests-tempest-tests-runner" Nov 22 09:59:20 crc kubenswrapper[4789]: I1122 09:59:20.930907 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest-s01-single-test" Nov 22 09:59:20 crc kubenswrapper[4789]: I1122 09:59:20.943074 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s1" Nov 22 09:59:20 crc kubenswrapper[4789]: I1122 09:59:20.956052 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s1" Nov 22 09:59:20 crc kubenswrapper[4789]: I1122 09:59:20.977264 4789 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/18bbf966-a509-4faa-a8de-28045e5c01b4-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 09:59:20 crc kubenswrapper[4789]: I1122 09:59:20.977485 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-77qg6\" (UniqueName: \"kubernetes.io/projected/18bbf966-a509-4faa-a8de-28045e5c01b4-kube-api-access-77qg6\") on node \"crc\" DevicePath \"\"" Nov 22 09:59:20 crc kubenswrapper[4789]: I1122 09:59:20.977509 4789 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/18bbf966-a509-4faa-a8de-28045e5c01b4-ceph\") on node \"crc\" DevicePath \"\"" Nov 22 09:59:20 crc kubenswrapper[4789]: I1122 09:59:20.977536 4789 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Nov 22 09:59:20 crc kubenswrapper[4789]: I1122 09:59:20.977562 4789 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/18bbf966-a509-4faa-a8de-28045e5c01b4-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Nov 22 09:59:20 crc kubenswrapper[4789]: I1122 09:59:20.977578 4789 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/18bbf966-a509-4faa-a8de-28045e5c01b4-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Nov 22 09:59:20 crc kubenswrapper[4789]: I1122 09:59:20.989928 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest-s01-single-test"] Nov 22 09:59:21 crc kubenswrapper[4789]: I1122 09:59:21.007987 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18bbf966-a509-4faa-a8de-28045e5c01b4-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "18bbf966-a509-4faa-a8de-28045e5c01b4" (UID: "18bbf966-a509-4faa-a8de-28045e5c01b4"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 09:59:21 crc kubenswrapper[4789]: I1122 09:59:21.039194 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18bbf966-a509-4faa-a8de-28045e5c01b4-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "18bbf966-a509-4faa-a8de-28045e5c01b4" (UID: "18bbf966-a509-4faa-a8de-28045e5c01b4"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 09:59:21 crc kubenswrapper[4789]: I1122 09:59:21.052201 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest-s00-full" event={"ID":"18bbf966-a509-4faa-a8de-28045e5c01b4","Type":"ContainerDied","Data":"efa8a498577cb4705e8534f7b995927f554ec21e5098e6a05c0a810eb9d019e7"} Nov 22 09:59:21 crc kubenswrapper[4789]: I1122 09:59:21.052252 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="efa8a498577cb4705e8534f7b995927f554ec21e5098e6a05c0a810eb9d019e7" Nov 22 09:59:21 crc kubenswrapper[4789]: I1122 09:59:21.052345 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest-s00-full" Nov 22 09:59:21 crc kubenswrapper[4789]: I1122 09:59:21.057475 4789 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Nov 22 09:59:21 crc kubenswrapper[4789]: I1122 09:59:21.064030 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18bbf966-a509-4faa-a8de-28045e5c01b4-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "18bbf966-a509-4faa-a8de-28045e5c01b4" (UID: "18bbf966-a509-4faa-a8de-28045e5c01b4"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 09:59:21 crc kubenswrapper[4789]: I1122 09:59:21.079985 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/21bbc026-9e56-43bc-a3bf-0ec807812dba-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"21bbc026-9e56-43bc-a3bf-0ec807812dba\") " pod="openstack/tempest-tests-tempest-s01-single-test" Nov 22 09:59:21 crc kubenswrapper[4789]: I1122 09:59:21.080046 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hvj9q\" (UniqueName: \"kubernetes.io/projected/21bbc026-9e56-43bc-a3bf-0ec807812dba-kube-api-access-hvj9q\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"21bbc026-9e56-43bc-a3bf-0ec807812dba\") " pod="openstack/tempest-tests-tempest-s01-single-test" Nov 22 09:59:21 crc kubenswrapper[4789]: I1122 09:59:21.080065 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/21bbc026-9e56-43bc-a3bf-0ec807812dba-ssh-key\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"21bbc026-9e56-43bc-a3bf-0ec807812dba\") " pod="openstack/tempest-tests-tempest-s01-single-test" Nov 22 09:59:21 crc kubenswrapper[4789]: I1122 09:59:21.080082 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/21bbc026-9e56-43bc-a3bf-0ec807812dba-openstack-config-secret\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"21bbc026-9e56-43bc-a3bf-0ec807812dba\") " pod="openstack/tempest-tests-tempest-s01-single-test" Nov 22 09:59:21 crc kubenswrapper[4789]: I1122 09:59:21.080109 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/21bbc026-9e56-43bc-a3bf-0ec807812dba-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"21bbc026-9e56-43bc-a3bf-0ec807812dba\") " pod="openstack/tempest-tests-tempest-s01-single-test" Nov 22 09:59:21 crc kubenswrapper[4789]: I1122 09:59:21.080152 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/21bbc026-9e56-43bc-a3bf-0ec807812dba-ceph\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"21bbc026-9e56-43bc-a3bf-0ec807812dba\") " pod="openstack/tempest-tests-tempest-s01-single-test" Nov 22 09:59:21 crc kubenswrapper[4789]: I1122 09:59:21.080174 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/21bbc026-9e56-43bc-a3bf-0ec807812dba-openstack-config\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"21bbc026-9e56-43bc-a3bf-0ec807812dba\") " pod="openstack/tempest-tests-tempest-s01-single-test" Nov 22 09:59:21 crc kubenswrapper[4789]: I1122 09:59:21.080194 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"21bbc026-9e56-43bc-a3bf-0ec807812dba\") " pod="openstack/tempest-tests-tempest-s01-single-test" Nov 22 09:59:21 crc kubenswrapper[4789]: I1122 09:59:21.080304 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/21bbc026-9e56-43bc-a3bf-0ec807812dba-config-data\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"21bbc026-9e56-43bc-a3bf-0ec807812dba\") " pod="openstack/tempest-tests-tempest-s01-single-test" Nov 22 09:59:21 crc kubenswrapper[4789]: I1122 09:59:21.080402 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/21bbc026-9e56-43bc-a3bf-0ec807812dba-ca-certs\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"21bbc026-9e56-43bc-a3bf-0ec807812dba\") " pod="openstack/tempest-tests-tempest-s01-single-test" Nov 22 09:59:21 crc kubenswrapper[4789]: I1122 09:59:21.080541 4789 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/18bbf966-a509-4faa-a8de-28045e5c01b4-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Nov 22 09:59:21 crc kubenswrapper[4789]: I1122 09:59:21.080555 4789 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/18bbf966-a509-4faa-a8de-28045e5c01b4-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 22 09:59:21 crc kubenswrapper[4789]: I1122 09:59:21.080565 4789 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/18bbf966-a509-4faa-a8de-28045e5c01b4-ca-certs\") on node \"crc\" DevicePath \"\"" Nov 22 09:59:21 crc kubenswrapper[4789]: I1122 09:59:21.080905 4789 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"21bbc026-9e56-43bc-a3bf-0ec807812dba\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/tempest-tests-tempest-s01-single-test" Nov 22 09:59:21 crc kubenswrapper[4789]: I1122 09:59:21.084267 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/18bbf966-a509-4faa-a8de-28045e5c01b4-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "18bbf966-a509-4faa-a8de-28045e5c01b4" (UID: "18bbf966-a509-4faa-a8de-28045e5c01b4"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 09:59:21 crc kubenswrapper[4789]: I1122 09:59:21.118310 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"21bbc026-9e56-43bc-a3bf-0ec807812dba\") " pod="openstack/tempest-tests-tempest-s01-single-test" Nov 22 09:59:21 crc kubenswrapper[4789]: I1122 09:59:21.182078 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/21bbc026-9e56-43bc-a3bf-0ec807812dba-config-data\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"21bbc026-9e56-43bc-a3bf-0ec807812dba\") " pod="openstack/tempest-tests-tempest-s01-single-test" Nov 22 09:59:21 crc kubenswrapper[4789]: I1122 09:59:21.182158 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/21bbc026-9e56-43bc-a3bf-0ec807812dba-ca-certs\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"21bbc026-9e56-43bc-a3bf-0ec807812dba\") " pod="openstack/tempest-tests-tempest-s01-single-test" Nov 22 09:59:21 crc kubenswrapper[4789]: I1122 09:59:21.182226 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/21bbc026-9e56-43bc-a3bf-0ec807812dba-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"21bbc026-9e56-43bc-a3bf-0ec807812dba\") " pod="openstack/tempest-tests-tempest-s01-single-test" Nov 22 09:59:21 crc kubenswrapper[4789]: I1122 09:59:21.182248 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hvj9q\" (UniqueName: \"kubernetes.io/projected/21bbc026-9e56-43bc-a3bf-0ec807812dba-kube-api-access-hvj9q\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"21bbc026-9e56-43bc-a3bf-0ec807812dba\") " pod="openstack/tempest-tests-tempest-s01-single-test" Nov 22 09:59:21 crc kubenswrapper[4789]: I1122 09:59:21.182265 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/21bbc026-9e56-43bc-a3bf-0ec807812dba-ssh-key\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"21bbc026-9e56-43bc-a3bf-0ec807812dba\") " pod="openstack/tempest-tests-tempest-s01-single-test" Nov 22 09:59:21 crc kubenswrapper[4789]: I1122 09:59:21.182281 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/21bbc026-9e56-43bc-a3bf-0ec807812dba-openstack-config-secret\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"21bbc026-9e56-43bc-a3bf-0ec807812dba\") " pod="openstack/tempest-tests-tempest-s01-single-test" Nov 22 09:59:21 crc kubenswrapper[4789]: I1122 09:59:21.182302 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/21bbc026-9e56-43bc-a3bf-0ec807812dba-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"21bbc026-9e56-43bc-a3bf-0ec807812dba\") " pod="openstack/tempest-tests-tempest-s01-single-test" Nov 22 09:59:21 crc kubenswrapper[4789]: I1122 09:59:21.182329 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/21bbc026-9e56-43bc-a3bf-0ec807812dba-ceph\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"21bbc026-9e56-43bc-a3bf-0ec807812dba\") " pod="openstack/tempest-tests-tempest-s01-single-test" Nov 22 09:59:21 crc kubenswrapper[4789]: I1122 09:59:21.182348 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/21bbc026-9e56-43bc-a3bf-0ec807812dba-openstack-config\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"21bbc026-9e56-43bc-a3bf-0ec807812dba\") " pod="openstack/tempest-tests-tempest-s01-single-test" Nov 22 09:59:21 crc kubenswrapper[4789]: I1122 09:59:21.182422 4789 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/18bbf966-a509-4faa-a8de-28045e5c01b4-openstack-config\") on node \"crc\" DevicePath \"\"" Nov 22 09:59:21 crc kubenswrapper[4789]: I1122 09:59:21.182920 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/21bbc026-9e56-43bc-a3bf-0ec807812dba-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"21bbc026-9e56-43bc-a3bf-0ec807812dba\") " pod="openstack/tempest-tests-tempest-s01-single-test" Nov 22 09:59:21 crc kubenswrapper[4789]: I1122 09:59:21.183003 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/21bbc026-9e56-43bc-a3bf-0ec807812dba-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"21bbc026-9e56-43bc-a3bf-0ec807812dba\") " pod="openstack/tempest-tests-tempest-s01-single-test" Nov 22 09:59:21 crc kubenswrapper[4789]: I1122 09:59:21.183308 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/21bbc026-9e56-43bc-a3bf-0ec807812dba-openstack-config\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"21bbc026-9e56-43bc-a3bf-0ec807812dba\") " pod="openstack/tempest-tests-tempest-s01-single-test" Nov 22 09:59:21 crc kubenswrapper[4789]: I1122 09:59:21.183428 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/21bbc026-9e56-43bc-a3bf-0ec807812dba-config-data\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"21bbc026-9e56-43bc-a3bf-0ec807812dba\") " pod="openstack/tempest-tests-tempest-s01-single-test" Nov 22 09:59:21 crc kubenswrapper[4789]: I1122 09:59:21.185370 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/21bbc026-9e56-43bc-a3bf-0ec807812dba-ca-certs\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"21bbc026-9e56-43bc-a3bf-0ec807812dba\") " pod="openstack/tempest-tests-tempest-s01-single-test" Nov 22 09:59:21 crc kubenswrapper[4789]: I1122 09:59:21.185802 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/21bbc026-9e56-43bc-a3bf-0ec807812dba-ceph\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"21bbc026-9e56-43bc-a3bf-0ec807812dba\") " pod="openstack/tempest-tests-tempest-s01-single-test" Nov 22 09:59:21 crc kubenswrapper[4789]: I1122 09:59:21.186086 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/21bbc026-9e56-43bc-a3bf-0ec807812dba-openstack-config-secret\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"21bbc026-9e56-43bc-a3bf-0ec807812dba\") " pod="openstack/tempest-tests-tempest-s01-single-test" Nov 22 09:59:21 crc kubenswrapper[4789]: I1122 09:59:21.187332 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/21bbc026-9e56-43bc-a3bf-0ec807812dba-ssh-key\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"21bbc026-9e56-43bc-a3bf-0ec807812dba\") " pod="openstack/tempest-tests-tempest-s01-single-test" Nov 22 09:59:21 crc kubenswrapper[4789]: I1122 09:59:21.200472 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hvj9q\" (UniqueName: \"kubernetes.io/projected/21bbc026-9e56-43bc-a3bf-0ec807812dba-kube-api-access-hvj9q\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"21bbc026-9e56-43bc-a3bf-0ec807812dba\") " pod="openstack/tempest-tests-tempest-s01-single-test" Nov 22 09:59:21 crc kubenswrapper[4789]: I1122 09:59:21.417586 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest-s01-single-test" Nov 22 09:59:21 crc kubenswrapper[4789]: W1122 09:59:21.973814 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod21bbc026_9e56_43bc_a3bf_0ec807812dba.slice/crio-4a2602ea296cf37f885cb27557be9e340b331a59541e8753479b8f43bdbb2abc WatchSource:0}: Error finding container 4a2602ea296cf37f885cb27557be9e340b331a59541e8753479b8f43bdbb2abc: Status 404 returned error can't find the container with id 4a2602ea296cf37f885cb27557be9e340b331a59541e8753479b8f43bdbb2abc Nov 22 09:59:21 crc kubenswrapper[4789]: I1122 09:59:21.990975 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest-s01-single-test"] Nov 22 09:59:22 crc kubenswrapper[4789]: I1122 09:59:22.066052 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest-s01-single-test" event={"ID":"21bbc026-9e56-43bc-a3bf-0ec807812dba","Type":"ContainerStarted","Data":"4a2602ea296cf37f885cb27557be9e340b331a59541e8753479b8f43bdbb2abc"} Nov 22 09:59:23 crc kubenswrapper[4789]: I1122 09:59:23.089277 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest-s01-single-test" event={"ID":"21bbc026-9e56-43bc-a3bf-0ec807812dba","Type":"ContainerStarted","Data":"c489e6990bf44b2984c49f082d91206b68ee60223dc7f75ea96ca87bd67e5266"} Nov 22 09:59:23 crc kubenswrapper[4789]: I1122 09:59:23.113049 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest-s01-single-test" podStartSLOduration=3.113025746 podStartE2EDuration="3.113025746s" podCreationTimestamp="2025-11-22 09:59:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 09:59:23.105102984 +0000 UTC m=+7377.339503277" watchObservedRunningTime="2025-11-22 09:59:23.113025746 +0000 UTC m=+7377.347426029" Nov 22 09:59:25 crc kubenswrapper[4789]: I1122 09:59:25.966004 4789 scope.go:117] "RemoveContainer" containerID="26c17a090dfedea5cf0834b08473176e1ddf9f8e1cbf99d58db7701ed655462d" Nov 22 09:59:25 crc kubenswrapper[4789]: E1122 09:59:25.966649 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:59:40 crc kubenswrapper[4789]: I1122 09:59:40.966362 4789 scope.go:117] "RemoveContainer" containerID="26c17a090dfedea5cf0834b08473176e1ddf9f8e1cbf99d58db7701ed655462d" Nov 22 09:59:40 crc kubenswrapper[4789]: E1122 09:59:40.984133 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 09:59:54 crc kubenswrapper[4789]: I1122 09:59:54.966263 4789 scope.go:117] "RemoveContainer" containerID="26c17a090dfedea5cf0834b08473176e1ddf9f8e1cbf99d58db7701ed655462d" Nov 22 09:59:54 crc kubenswrapper[4789]: E1122 09:59:54.967181 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 10:00:00 crc kubenswrapper[4789]: I1122 10:00:00.152464 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396760-9cpdg"] Nov 22 10:00:00 crc kubenswrapper[4789]: I1122 10:00:00.154996 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396760-9cpdg" Nov 22 10:00:00 crc kubenswrapper[4789]: I1122 10:00:00.158178 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 22 10:00:00 crc kubenswrapper[4789]: I1122 10:00:00.158182 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 22 10:00:00 crc kubenswrapper[4789]: I1122 10:00:00.162139 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396760-9cpdg"] Nov 22 10:00:00 crc kubenswrapper[4789]: I1122 10:00:00.319035 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9k9cl\" (UniqueName: \"kubernetes.io/projected/92173eea-98c4-48f2-8235-8ea34563b0a2-kube-api-access-9k9cl\") pod \"collect-profiles-29396760-9cpdg\" (UID: \"92173eea-98c4-48f2-8235-8ea34563b0a2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396760-9cpdg" Nov 22 10:00:00 crc kubenswrapper[4789]: I1122 10:00:00.319485 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/92173eea-98c4-48f2-8235-8ea34563b0a2-secret-volume\") pod \"collect-profiles-29396760-9cpdg\" (UID: \"92173eea-98c4-48f2-8235-8ea34563b0a2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396760-9cpdg" Nov 22 10:00:00 crc kubenswrapper[4789]: I1122 10:00:00.319587 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/92173eea-98c4-48f2-8235-8ea34563b0a2-config-volume\") pod \"collect-profiles-29396760-9cpdg\" (UID: \"92173eea-98c4-48f2-8235-8ea34563b0a2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396760-9cpdg" Nov 22 10:00:00 crc kubenswrapper[4789]: I1122 10:00:00.422711 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9k9cl\" (UniqueName: \"kubernetes.io/projected/92173eea-98c4-48f2-8235-8ea34563b0a2-kube-api-access-9k9cl\") pod \"collect-profiles-29396760-9cpdg\" (UID: \"92173eea-98c4-48f2-8235-8ea34563b0a2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396760-9cpdg" Nov 22 10:00:00 crc kubenswrapper[4789]: I1122 10:00:00.422836 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/92173eea-98c4-48f2-8235-8ea34563b0a2-secret-volume\") pod \"collect-profiles-29396760-9cpdg\" (UID: \"92173eea-98c4-48f2-8235-8ea34563b0a2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396760-9cpdg" Nov 22 10:00:00 crc kubenswrapper[4789]: I1122 10:00:00.422913 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/92173eea-98c4-48f2-8235-8ea34563b0a2-config-volume\") pod \"collect-profiles-29396760-9cpdg\" (UID: \"92173eea-98c4-48f2-8235-8ea34563b0a2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396760-9cpdg" Nov 22 10:00:00 crc kubenswrapper[4789]: I1122 10:00:00.424607 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/92173eea-98c4-48f2-8235-8ea34563b0a2-config-volume\") pod \"collect-profiles-29396760-9cpdg\" (UID: \"92173eea-98c4-48f2-8235-8ea34563b0a2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396760-9cpdg" Nov 22 10:00:00 crc kubenswrapper[4789]: I1122 10:00:00.431602 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/92173eea-98c4-48f2-8235-8ea34563b0a2-secret-volume\") pod \"collect-profiles-29396760-9cpdg\" (UID: \"92173eea-98c4-48f2-8235-8ea34563b0a2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396760-9cpdg" Nov 22 10:00:00 crc kubenswrapper[4789]: I1122 10:00:00.441550 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9k9cl\" (UniqueName: \"kubernetes.io/projected/92173eea-98c4-48f2-8235-8ea34563b0a2-kube-api-access-9k9cl\") pod \"collect-profiles-29396760-9cpdg\" (UID: \"92173eea-98c4-48f2-8235-8ea34563b0a2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396760-9cpdg" Nov 22 10:00:00 crc kubenswrapper[4789]: I1122 10:00:00.488847 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396760-9cpdg" Nov 22 10:00:00 crc kubenswrapper[4789]: I1122 10:00:00.999133 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396760-9cpdg"] Nov 22 10:00:01 crc kubenswrapper[4789]: I1122 10:00:01.534289 4789 generic.go:334] "Generic (PLEG): container finished" podID="92173eea-98c4-48f2-8235-8ea34563b0a2" containerID="8f8ad012324b671cbaa6639ec7f247bcdd31a20d0db8c04fa33459c376bab696" exitCode=0 Nov 22 10:00:01 crc kubenswrapper[4789]: I1122 10:00:01.534343 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29396760-9cpdg" event={"ID":"92173eea-98c4-48f2-8235-8ea34563b0a2","Type":"ContainerDied","Data":"8f8ad012324b671cbaa6639ec7f247bcdd31a20d0db8c04fa33459c376bab696"} Nov 22 10:00:01 crc kubenswrapper[4789]: I1122 10:00:01.535444 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29396760-9cpdg" event={"ID":"92173eea-98c4-48f2-8235-8ea34563b0a2","Type":"ContainerStarted","Data":"75a437f361bcafffdc39b74c2850483f49f6b788805ff3dc270f9418ecf6cd1c"} Nov 22 10:00:02 crc kubenswrapper[4789]: I1122 10:00:02.903434 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396760-9cpdg" Nov 22 10:00:02 crc kubenswrapper[4789]: I1122 10:00:02.981872 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/92173eea-98c4-48f2-8235-8ea34563b0a2-config-volume\") pod \"92173eea-98c4-48f2-8235-8ea34563b0a2\" (UID: \"92173eea-98c4-48f2-8235-8ea34563b0a2\") " Nov 22 10:00:02 crc kubenswrapper[4789]: I1122 10:00:02.982620 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9k9cl\" (UniqueName: \"kubernetes.io/projected/92173eea-98c4-48f2-8235-8ea34563b0a2-kube-api-access-9k9cl\") pod \"92173eea-98c4-48f2-8235-8ea34563b0a2\" (UID: \"92173eea-98c4-48f2-8235-8ea34563b0a2\") " Nov 22 10:00:02 crc kubenswrapper[4789]: I1122 10:00:02.982673 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/92173eea-98c4-48f2-8235-8ea34563b0a2-config-volume" (OuterVolumeSpecName: "config-volume") pod "92173eea-98c4-48f2-8235-8ea34563b0a2" (UID: "92173eea-98c4-48f2-8235-8ea34563b0a2"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 10:00:02 crc kubenswrapper[4789]: I1122 10:00:02.982823 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/92173eea-98c4-48f2-8235-8ea34563b0a2-secret-volume\") pod \"92173eea-98c4-48f2-8235-8ea34563b0a2\" (UID: \"92173eea-98c4-48f2-8235-8ea34563b0a2\") " Nov 22 10:00:02 crc kubenswrapper[4789]: I1122 10:00:02.983377 4789 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/92173eea-98c4-48f2-8235-8ea34563b0a2-config-volume\") on node \"crc\" DevicePath \"\"" Nov 22 10:00:02 crc kubenswrapper[4789]: I1122 10:00:02.988018 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92173eea-98c4-48f2-8235-8ea34563b0a2-kube-api-access-9k9cl" (OuterVolumeSpecName: "kube-api-access-9k9cl") pod "92173eea-98c4-48f2-8235-8ea34563b0a2" (UID: "92173eea-98c4-48f2-8235-8ea34563b0a2"). InnerVolumeSpecName "kube-api-access-9k9cl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 10:00:02 crc kubenswrapper[4789]: I1122 10:00:02.988047 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92173eea-98c4-48f2-8235-8ea34563b0a2-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "92173eea-98c4-48f2-8235-8ea34563b0a2" (UID: "92173eea-98c4-48f2-8235-8ea34563b0a2"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 10:00:03 crc kubenswrapper[4789]: I1122 10:00:03.085689 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9k9cl\" (UniqueName: \"kubernetes.io/projected/92173eea-98c4-48f2-8235-8ea34563b0a2-kube-api-access-9k9cl\") on node \"crc\" DevicePath \"\"" Nov 22 10:00:03 crc kubenswrapper[4789]: I1122 10:00:03.085739 4789 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/92173eea-98c4-48f2-8235-8ea34563b0a2-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 22 10:00:03 crc kubenswrapper[4789]: I1122 10:00:03.565447 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29396760-9cpdg" event={"ID":"92173eea-98c4-48f2-8235-8ea34563b0a2","Type":"ContainerDied","Data":"75a437f361bcafffdc39b74c2850483f49f6b788805ff3dc270f9418ecf6cd1c"} Nov 22 10:00:03 crc kubenswrapper[4789]: I1122 10:00:03.565489 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="75a437f361bcafffdc39b74c2850483f49f6b788805ff3dc270f9418ecf6cd1c" Nov 22 10:00:03 crc kubenswrapper[4789]: I1122 10:00:03.565985 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396760-9cpdg" Nov 22 10:00:03 crc kubenswrapper[4789]: I1122 10:00:03.990659 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396715-qcwt2"] Nov 22 10:00:04 crc kubenswrapper[4789]: I1122 10:00:04.006157 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396715-qcwt2"] Nov 22 10:00:05 crc kubenswrapper[4789]: I1122 10:00:05.976302 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="55bbc3c2-372c-4ce9-bbae-4b111aa57ca7" path="/var/lib/kubelet/pods/55bbc3c2-372c-4ce9-bbae-4b111aa57ca7/volumes" Nov 22 10:00:09 crc kubenswrapper[4789]: I1122 10:00:09.965737 4789 scope.go:117] "RemoveContainer" containerID="26c17a090dfedea5cf0834b08473176e1ddf9f8e1cbf99d58db7701ed655462d" Nov 22 10:00:10 crc kubenswrapper[4789]: I1122 10:00:10.636717 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" event={"ID":"58aa8071-7ecc-4692-8789-2db04ec70647","Type":"ContainerStarted","Data":"7b5ed03a8e4ebb046ae3b2e18435bfca04a6eb59a1c3958782493a816f47f210"} Nov 22 10:00:30 crc kubenswrapper[4789]: I1122 10:00:30.222837 4789 scope.go:117] "RemoveContainer" containerID="abdc68875fb3d8f78fcd55346727187c234742fc0f7da6c9ef937ec8ca749d8d" Nov 22 10:00:47 crc kubenswrapper[4789]: I1122 10:00:47.452141 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-w8786"] Nov 22 10:00:47 crc kubenswrapper[4789]: E1122 10:00:47.453183 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92173eea-98c4-48f2-8235-8ea34563b0a2" containerName="collect-profiles" Nov 22 10:00:47 crc kubenswrapper[4789]: I1122 10:00:47.453201 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="92173eea-98c4-48f2-8235-8ea34563b0a2" containerName="collect-profiles" Nov 22 10:00:47 crc kubenswrapper[4789]: I1122 10:00:47.453446 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="92173eea-98c4-48f2-8235-8ea34563b0a2" containerName="collect-profiles" Nov 22 10:00:47 crc kubenswrapper[4789]: I1122 10:00:47.455255 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w8786" Nov 22 10:00:47 crc kubenswrapper[4789]: I1122 10:00:47.477842 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-w8786"] Nov 22 10:00:47 crc kubenswrapper[4789]: I1122 10:00:47.603651 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1bde6612-2541-46cd-a1e1-20d0736213ee-utilities\") pod \"redhat-operators-w8786\" (UID: \"1bde6612-2541-46cd-a1e1-20d0736213ee\") " pod="openshift-marketplace/redhat-operators-w8786" Nov 22 10:00:47 crc kubenswrapper[4789]: I1122 10:00:47.604015 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1bde6612-2541-46cd-a1e1-20d0736213ee-catalog-content\") pod \"redhat-operators-w8786\" (UID: \"1bde6612-2541-46cd-a1e1-20d0736213ee\") " pod="openshift-marketplace/redhat-operators-w8786" Nov 22 10:00:47 crc kubenswrapper[4789]: I1122 10:00:47.604423 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r4vmm\" (UniqueName: \"kubernetes.io/projected/1bde6612-2541-46cd-a1e1-20d0736213ee-kube-api-access-r4vmm\") pod \"redhat-operators-w8786\" (UID: \"1bde6612-2541-46cd-a1e1-20d0736213ee\") " pod="openshift-marketplace/redhat-operators-w8786" Nov 22 10:00:47 crc kubenswrapper[4789]: I1122 10:00:47.705785 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r4vmm\" (UniqueName: \"kubernetes.io/projected/1bde6612-2541-46cd-a1e1-20d0736213ee-kube-api-access-r4vmm\") pod \"redhat-operators-w8786\" (UID: \"1bde6612-2541-46cd-a1e1-20d0736213ee\") " pod="openshift-marketplace/redhat-operators-w8786" Nov 22 10:00:47 crc kubenswrapper[4789]: I1122 10:00:47.705846 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1bde6612-2541-46cd-a1e1-20d0736213ee-utilities\") pod \"redhat-operators-w8786\" (UID: \"1bde6612-2541-46cd-a1e1-20d0736213ee\") " pod="openshift-marketplace/redhat-operators-w8786" Nov 22 10:00:47 crc kubenswrapper[4789]: I1122 10:00:47.705958 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1bde6612-2541-46cd-a1e1-20d0736213ee-catalog-content\") pod \"redhat-operators-w8786\" (UID: \"1bde6612-2541-46cd-a1e1-20d0736213ee\") " pod="openshift-marketplace/redhat-operators-w8786" Nov 22 10:00:47 crc kubenswrapper[4789]: I1122 10:00:47.706374 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1bde6612-2541-46cd-a1e1-20d0736213ee-utilities\") pod \"redhat-operators-w8786\" (UID: \"1bde6612-2541-46cd-a1e1-20d0736213ee\") " pod="openshift-marketplace/redhat-operators-w8786" Nov 22 10:00:47 crc kubenswrapper[4789]: I1122 10:00:47.706430 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1bde6612-2541-46cd-a1e1-20d0736213ee-catalog-content\") pod \"redhat-operators-w8786\" (UID: \"1bde6612-2541-46cd-a1e1-20d0736213ee\") " pod="openshift-marketplace/redhat-operators-w8786" Nov 22 10:00:47 crc kubenswrapper[4789]: I1122 10:00:47.735336 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r4vmm\" (UniqueName: \"kubernetes.io/projected/1bde6612-2541-46cd-a1e1-20d0736213ee-kube-api-access-r4vmm\") pod \"redhat-operators-w8786\" (UID: \"1bde6612-2541-46cd-a1e1-20d0736213ee\") " pod="openshift-marketplace/redhat-operators-w8786" Nov 22 10:00:47 crc kubenswrapper[4789]: I1122 10:00:47.800493 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w8786" Nov 22 10:00:48 crc kubenswrapper[4789]: I1122 10:00:48.299156 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-w8786"] Nov 22 10:00:49 crc kubenswrapper[4789]: I1122 10:00:49.065078 4789 generic.go:334] "Generic (PLEG): container finished" podID="1bde6612-2541-46cd-a1e1-20d0736213ee" containerID="df1edd7ac419c3d7d6ae52f37130f9b4b381db03cdef5a166aa8d9eb5c9fb69b" exitCode=0 Nov 22 10:00:49 crc kubenswrapper[4789]: I1122 10:00:49.065130 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w8786" event={"ID":"1bde6612-2541-46cd-a1e1-20d0736213ee","Type":"ContainerDied","Data":"df1edd7ac419c3d7d6ae52f37130f9b4b381db03cdef5a166aa8d9eb5c9fb69b"} Nov 22 10:00:49 crc kubenswrapper[4789]: I1122 10:00:49.065193 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w8786" event={"ID":"1bde6612-2541-46cd-a1e1-20d0736213ee","Type":"ContainerStarted","Data":"2252453e8555f0e6340aefd9d1d2f127f816f52f7d28a9df2edc4465d339eab6"} Nov 22 10:00:49 crc kubenswrapper[4789]: I1122 10:00:49.067559 4789 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 22 10:00:50 crc kubenswrapper[4789]: I1122 10:00:50.078337 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w8786" event={"ID":"1bde6612-2541-46cd-a1e1-20d0736213ee","Type":"ContainerStarted","Data":"707b09a27593500d23e60d1e5ac6c850f0ffb2faad87cf8a57a00b66f48a2909"} Nov 22 10:00:51 crc kubenswrapper[4789]: I1122 10:00:51.101866 4789 generic.go:334] "Generic (PLEG): container finished" podID="1bde6612-2541-46cd-a1e1-20d0736213ee" containerID="707b09a27593500d23e60d1e5ac6c850f0ffb2faad87cf8a57a00b66f48a2909" exitCode=0 Nov 22 10:00:51 crc kubenswrapper[4789]: I1122 10:00:51.101945 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w8786" event={"ID":"1bde6612-2541-46cd-a1e1-20d0736213ee","Type":"ContainerDied","Data":"707b09a27593500d23e60d1e5ac6c850f0ffb2faad87cf8a57a00b66f48a2909"} Nov 22 10:00:52 crc kubenswrapper[4789]: I1122 10:00:52.113729 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w8786" event={"ID":"1bde6612-2541-46cd-a1e1-20d0736213ee","Type":"ContainerStarted","Data":"284b98bfe6e47281f5523a41aa2a127c4f9df8046af4dc79a2ea1131f9c1067a"} Nov 22 10:00:52 crc kubenswrapper[4789]: I1122 10:00:52.138190 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-w8786" podStartSLOduration=2.686200444 podStartE2EDuration="5.138170796s" podCreationTimestamp="2025-11-22 10:00:47 +0000 UTC" firstStartedPulling="2025-11-22 10:00:49.067268419 +0000 UTC m=+7463.301668692" lastFinishedPulling="2025-11-22 10:00:51.519238741 +0000 UTC m=+7465.753639044" observedRunningTime="2025-11-22 10:00:52.128439085 +0000 UTC m=+7466.362839358" watchObservedRunningTime="2025-11-22 10:00:52.138170796 +0000 UTC m=+7466.372571069" Nov 22 10:00:57 crc kubenswrapper[4789]: I1122 10:00:57.801270 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-w8786" Nov 22 10:00:57 crc kubenswrapper[4789]: I1122 10:00:57.801859 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-w8786" Nov 22 10:00:57 crc kubenswrapper[4789]: I1122 10:00:57.847516 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-w8786" Nov 22 10:00:58 crc kubenswrapper[4789]: I1122 10:00:58.221378 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-w8786" Nov 22 10:00:58 crc kubenswrapper[4789]: I1122 10:00:58.284440 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-w8786"] Nov 22 10:01:00 crc kubenswrapper[4789]: I1122 10:01:00.148051 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29396761-w9qjq"] Nov 22 10:01:00 crc kubenswrapper[4789]: I1122 10:01:00.150155 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29396761-w9qjq" Nov 22 10:01:00 crc kubenswrapper[4789]: I1122 10:01:00.165694 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29396761-w9qjq"] Nov 22 10:01:00 crc kubenswrapper[4789]: I1122 10:01:00.199037 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-w8786" podUID="1bde6612-2541-46cd-a1e1-20d0736213ee" containerName="registry-server" containerID="cri-o://284b98bfe6e47281f5523a41aa2a127c4f9df8046af4dc79a2ea1131f9c1067a" gracePeriod=2 Nov 22 10:01:00 crc kubenswrapper[4789]: I1122 10:01:00.266230 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21453b21-488b-4068-8ba4-1fa71816009d-combined-ca-bundle\") pod \"keystone-cron-29396761-w9qjq\" (UID: \"21453b21-488b-4068-8ba4-1fa71816009d\") " pod="openstack/keystone-cron-29396761-w9qjq" Nov 22 10:01:00 crc kubenswrapper[4789]: I1122 10:01:00.266314 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9hjrg\" (UniqueName: \"kubernetes.io/projected/21453b21-488b-4068-8ba4-1fa71816009d-kube-api-access-9hjrg\") pod \"keystone-cron-29396761-w9qjq\" (UID: \"21453b21-488b-4068-8ba4-1fa71816009d\") " pod="openstack/keystone-cron-29396761-w9qjq" Nov 22 10:01:00 crc kubenswrapper[4789]: I1122 10:01:00.266334 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21453b21-488b-4068-8ba4-1fa71816009d-config-data\") pod \"keystone-cron-29396761-w9qjq\" (UID: \"21453b21-488b-4068-8ba4-1fa71816009d\") " pod="openstack/keystone-cron-29396761-w9qjq" Nov 22 10:01:00 crc kubenswrapper[4789]: I1122 10:01:00.266380 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/21453b21-488b-4068-8ba4-1fa71816009d-fernet-keys\") pod \"keystone-cron-29396761-w9qjq\" (UID: \"21453b21-488b-4068-8ba4-1fa71816009d\") " pod="openstack/keystone-cron-29396761-w9qjq" Nov 22 10:01:00 crc kubenswrapper[4789]: I1122 10:01:00.367810 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21453b21-488b-4068-8ba4-1fa71816009d-config-data\") pod \"keystone-cron-29396761-w9qjq\" (UID: \"21453b21-488b-4068-8ba4-1fa71816009d\") " pod="openstack/keystone-cron-29396761-w9qjq" Nov 22 10:01:00 crc kubenswrapper[4789]: I1122 10:01:00.367881 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/21453b21-488b-4068-8ba4-1fa71816009d-fernet-keys\") pod \"keystone-cron-29396761-w9qjq\" (UID: \"21453b21-488b-4068-8ba4-1fa71816009d\") " pod="openstack/keystone-cron-29396761-w9qjq" Nov 22 10:01:00 crc kubenswrapper[4789]: I1122 10:01:00.368084 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21453b21-488b-4068-8ba4-1fa71816009d-combined-ca-bundle\") pod \"keystone-cron-29396761-w9qjq\" (UID: \"21453b21-488b-4068-8ba4-1fa71816009d\") " pod="openstack/keystone-cron-29396761-w9qjq" Nov 22 10:01:00 crc kubenswrapper[4789]: I1122 10:01:00.368158 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9hjrg\" (UniqueName: \"kubernetes.io/projected/21453b21-488b-4068-8ba4-1fa71816009d-kube-api-access-9hjrg\") pod \"keystone-cron-29396761-w9qjq\" (UID: \"21453b21-488b-4068-8ba4-1fa71816009d\") " pod="openstack/keystone-cron-29396761-w9qjq" Nov 22 10:01:00 crc kubenswrapper[4789]: I1122 10:01:00.378900 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21453b21-488b-4068-8ba4-1fa71816009d-combined-ca-bundle\") pod \"keystone-cron-29396761-w9qjq\" (UID: \"21453b21-488b-4068-8ba4-1fa71816009d\") " pod="openstack/keystone-cron-29396761-w9qjq" Nov 22 10:01:00 crc kubenswrapper[4789]: I1122 10:01:00.378940 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/21453b21-488b-4068-8ba4-1fa71816009d-fernet-keys\") pod \"keystone-cron-29396761-w9qjq\" (UID: \"21453b21-488b-4068-8ba4-1fa71816009d\") " pod="openstack/keystone-cron-29396761-w9qjq" Nov 22 10:01:00 crc kubenswrapper[4789]: I1122 10:01:00.379935 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21453b21-488b-4068-8ba4-1fa71816009d-config-data\") pod \"keystone-cron-29396761-w9qjq\" (UID: \"21453b21-488b-4068-8ba4-1fa71816009d\") " pod="openstack/keystone-cron-29396761-w9qjq" Nov 22 10:01:00 crc kubenswrapper[4789]: I1122 10:01:00.385305 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9hjrg\" (UniqueName: \"kubernetes.io/projected/21453b21-488b-4068-8ba4-1fa71816009d-kube-api-access-9hjrg\") pod \"keystone-cron-29396761-w9qjq\" (UID: \"21453b21-488b-4068-8ba4-1fa71816009d\") " pod="openstack/keystone-cron-29396761-w9qjq" Nov 22 10:01:00 crc kubenswrapper[4789]: I1122 10:01:00.473564 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29396761-w9qjq" Nov 22 10:01:00 crc kubenswrapper[4789]: I1122 10:01:00.894485 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w8786" Nov 22 10:01:00 crc kubenswrapper[4789]: I1122 10:01:00.973815 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29396761-w9qjq"] Nov 22 10:01:00 crc kubenswrapper[4789]: W1122 10:01:00.976665 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod21453b21_488b_4068_8ba4_1fa71816009d.slice/crio-008d98b91fb9b84012f793453b13fc26f8327bf0d4e7ebf7c102dc907e984987 WatchSource:0}: Error finding container 008d98b91fb9b84012f793453b13fc26f8327bf0d4e7ebf7c102dc907e984987: Status 404 returned error can't find the container with id 008d98b91fb9b84012f793453b13fc26f8327bf0d4e7ebf7c102dc907e984987 Nov 22 10:01:00 crc kubenswrapper[4789]: I1122 10:01:00.979870 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r4vmm\" (UniqueName: \"kubernetes.io/projected/1bde6612-2541-46cd-a1e1-20d0736213ee-kube-api-access-r4vmm\") pod \"1bde6612-2541-46cd-a1e1-20d0736213ee\" (UID: \"1bde6612-2541-46cd-a1e1-20d0736213ee\") " Nov 22 10:01:00 crc kubenswrapper[4789]: I1122 10:01:00.980176 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1bde6612-2541-46cd-a1e1-20d0736213ee-catalog-content\") pod \"1bde6612-2541-46cd-a1e1-20d0736213ee\" (UID: \"1bde6612-2541-46cd-a1e1-20d0736213ee\") " Nov 22 10:01:00 crc kubenswrapper[4789]: I1122 10:01:00.980202 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1bde6612-2541-46cd-a1e1-20d0736213ee-utilities\") pod \"1bde6612-2541-46cd-a1e1-20d0736213ee\" (UID: \"1bde6612-2541-46cd-a1e1-20d0736213ee\") " Nov 22 10:01:00 crc kubenswrapper[4789]: I1122 10:01:00.981025 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1bde6612-2541-46cd-a1e1-20d0736213ee-utilities" (OuterVolumeSpecName: "utilities") pod "1bde6612-2541-46cd-a1e1-20d0736213ee" (UID: "1bde6612-2541-46cd-a1e1-20d0736213ee"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 10:01:00 crc kubenswrapper[4789]: I1122 10:01:00.981431 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1bde6612-2541-46cd-a1e1-20d0736213ee-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 10:01:00 crc kubenswrapper[4789]: I1122 10:01:00.986488 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bde6612-2541-46cd-a1e1-20d0736213ee-kube-api-access-r4vmm" (OuterVolumeSpecName: "kube-api-access-r4vmm") pod "1bde6612-2541-46cd-a1e1-20d0736213ee" (UID: "1bde6612-2541-46cd-a1e1-20d0736213ee"). InnerVolumeSpecName "kube-api-access-r4vmm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 10:01:01 crc kubenswrapper[4789]: I1122 10:01:01.083472 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r4vmm\" (UniqueName: \"kubernetes.io/projected/1bde6612-2541-46cd-a1e1-20d0736213ee-kube-api-access-r4vmm\") on node \"crc\" DevicePath \"\"" Nov 22 10:01:01 crc kubenswrapper[4789]: I1122 10:01:01.210708 4789 generic.go:334] "Generic (PLEG): container finished" podID="1bde6612-2541-46cd-a1e1-20d0736213ee" containerID="284b98bfe6e47281f5523a41aa2a127c4f9df8046af4dc79a2ea1131f9c1067a" exitCode=0 Nov 22 10:01:01 crc kubenswrapper[4789]: I1122 10:01:01.210819 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w8786" event={"ID":"1bde6612-2541-46cd-a1e1-20d0736213ee","Type":"ContainerDied","Data":"284b98bfe6e47281f5523a41aa2a127c4f9df8046af4dc79a2ea1131f9c1067a"} Nov 22 10:01:01 crc kubenswrapper[4789]: I1122 10:01:01.212235 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w8786" event={"ID":"1bde6612-2541-46cd-a1e1-20d0736213ee","Type":"ContainerDied","Data":"2252453e8555f0e6340aefd9d1d2f127f816f52f7d28a9df2edc4465d339eab6"} Nov 22 10:01:01 crc kubenswrapper[4789]: I1122 10:01:01.210868 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w8786" Nov 22 10:01:01 crc kubenswrapper[4789]: I1122 10:01:01.212293 4789 scope.go:117] "RemoveContainer" containerID="284b98bfe6e47281f5523a41aa2a127c4f9df8046af4dc79a2ea1131f9c1067a" Nov 22 10:01:01 crc kubenswrapper[4789]: I1122 10:01:01.213983 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29396761-w9qjq" event={"ID":"21453b21-488b-4068-8ba4-1fa71816009d","Type":"ContainerStarted","Data":"008d98b91fb9b84012f793453b13fc26f8327bf0d4e7ebf7c102dc907e984987"} Nov 22 10:01:01 crc kubenswrapper[4789]: I1122 10:01:01.366576 4789 scope.go:117] "RemoveContainer" containerID="707b09a27593500d23e60d1e5ac6c850f0ffb2faad87cf8a57a00b66f48a2909" Nov 22 10:01:01 crc kubenswrapper[4789]: I1122 10:01:01.422013 4789 scope.go:117] "RemoveContainer" containerID="df1edd7ac419c3d7d6ae52f37130f9b4b381db03cdef5a166aa8d9eb5c9fb69b" Nov 22 10:01:01 crc kubenswrapper[4789]: I1122 10:01:01.499697 4789 scope.go:117] "RemoveContainer" containerID="284b98bfe6e47281f5523a41aa2a127c4f9df8046af4dc79a2ea1131f9c1067a" Nov 22 10:01:01 crc kubenswrapper[4789]: E1122 10:01:01.500341 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"284b98bfe6e47281f5523a41aa2a127c4f9df8046af4dc79a2ea1131f9c1067a\": container with ID starting with 284b98bfe6e47281f5523a41aa2a127c4f9df8046af4dc79a2ea1131f9c1067a not found: ID does not exist" containerID="284b98bfe6e47281f5523a41aa2a127c4f9df8046af4dc79a2ea1131f9c1067a" Nov 22 10:01:01 crc kubenswrapper[4789]: I1122 10:01:01.500395 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"284b98bfe6e47281f5523a41aa2a127c4f9df8046af4dc79a2ea1131f9c1067a"} err="failed to get container status \"284b98bfe6e47281f5523a41aa2a127c4f9df8046af4dc79a2ea1131f9c1067a\": rpc error: code = NotFound desc = could not find container \"284b98bfe6e47281f5523a41aa2a127c4f9df8046af4dc79a2ea1131f9c1067a\": container with ID starting with 284b98bfe6e47281f5523a41aa2a127c4f9df8046af4dc79a2ea1131f9c1067a not found: ID does not exist" Nov 22 10:01:01 crc kubenswrapper[4789]: I1122 10:01:01.500430 4789 scope.go:117] "RemoveContainer" containerID="707b09a27593500d23e60d1e5ac6c850f0ffb2faad87cf8a57a00b66f48a2909" Nov 22 10:01:01 crc kubenswrapper[4789]: E1122 10:01:01.500786 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"707b09a27593500d23e60d1e5ac6c850f0ffb2faad87cf8a57a00b66f48a2909\": container with ID starting with 707b09a27593500d23e60d1e5ac6c850f0ffb2faad87cf8a57a00b66f48a2909 not found: ID does not exist" containerID="707b09a27593500d23e60d1e5ac6c850f0ffb2faad87cf8a57a00b66f48a2909" Nov 22 10:01:01 crc kubenswrapper[4789]: I1122 10:01:01.500823 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"707b09a27593500d23e60d1e5ac6c850f0ffb2faad87cf8a57a00b66f48a2909"} err="failed to get container status \"707b09a27593500d23e60d1e5ac6c850f0ffb2faad87cf8a57a00b66f48a2909\": rpc error: code = NotFound desc = could not find container \"707b09a27593500d23e60d1e5ac6c850f0ffb2faad87cf8a57a00b66f48a2909\": container with ID starting with 707b09a27593500d23e60d1e5ac6c850f0ffb2faad87cf8a57a00b66f48a2909 not found: ID does not exist" Nov 22 10:01:01 crc kubenswrapper[4789]: I1122 10:01:01.500847 4789 scope.go:117] "RemoveContainer" containerID="df1edd7ac419c3d7d6ae52f37130f9b4b381db03cdef5a166aa8d9eb5c9fb69b" Nov 22 10:01:01 crc kubenswrapper[4789]: E1122 10:01:01.501362 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"df1edd7ac419c3d7d6ae52f37130f9b4b381db03cdef5a166aa8d9eb5c9fb69b\": container with ID starting with df1edd7ac419c3d7d6ae52f37130f9b4b381db03cdef5a166aa8d9eb5c9fb69b not found: ID does not exist" containerID="df1edd7ac419c3d7d6ae52f37130f9b4b381db03cdef5a166aa8d9eb5c9fb69b" Nov 22 10:01:01 crc kubenswrapper[4789]: I1122 10:01:01.501409 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"df1edd7ac419c3d7d6ae52f37130f9b4b381db03cdef5a166aa8d9eb5c9fb69b"} err="failed to get container status \"df1edd7ac419c3d7d6ae52f37130f9b4b381db03cdef5a166aa8d9eb5c9fb69b\": rpc error: code = NotFound desc = could not find container \"df1edd7ac419c3d7d6ae52f37130f9b4b381db03cdef5a166aa8d9eb5c9fb69b\": container with ID starting with df1edd7ac419c3d7d6ae52f37130f9b4b381db03cdef5a166aa8d9eb5c9fb69b not found: ID does not exist" Nov 22 10:01:04 crc kubenswrapper[4789]: I1122 10:01:04.100259 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1bde6612-2541-46cd-a1e1-20d0736213ee-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1bde6612-2541-46cd-a1e1-20d0736213ee" (UID: "1bde6612-2541-46cd-a1e1-20d0736213ee"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 10:01:04 crc kubenswrapper[4789]: I1122 10:01:04.167919 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1bde6612-2541-46cd-a1e1-20d0736213ee-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 10:01:04 crc kubenswrapper[4789]: I1122 10:01:04.243457 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-w8786"] Nov 22 10:01:04 crc kubenswrapper[4789]: I1122 10:01:04.248573 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29396761-w9qjq" event={"ID":"21453b21-488b-4068-8ba4-1fa71816009d","Type":"ContainerStarted","Data":"ea1f5f394962fef551350002ee2a76e2b34d171b655daefce9615b816b4f1a8e"} Nov 22 10:01:04 crc kubenswrapper[4789]: I1122 10:01:04.253386 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-w8786"] Nov 22 10:01:04 crc kubenswrapper[4789]: I1122 10:01:04.279543 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29396761-w9qjq" podStartSLOduration=4.279522717 podStartE2EDuration="4.279522717s" podCreationTimestamp="2025-11-22 10:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 10:01:04.262925282 +0000 UTC m=+7478.497325575" watchObservedRunningTime="2025-11-22 10:01:04.279522717 +0000 UTC m=+7478.513923000" Nov 22 10:01:05 crc kubenswrapper[4789]: I1122 10:01:05.987851 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bde6612-2541-46cd-a1e1-20d0736213ee" path="/var/lib/kubelet/pods/1bde6612-2541-46cd-a1e1-20d0736213ee/volumes" Nov 22 10:01:07 crc kubenswrapper[4789]: I1122 10:01:07.282614 4789 generic.go:334] "Generic (PLEG): container finished" podID="21453b21-488b-4068-8ba4-1fa71816009d" containerID="ea1f5f394962fef551350002ee2a76e2b34d171b655daefce9615b816b4f1a8e" exitCode=0 Nov 22 10:01:07 crc kubenswrapper[4789]: I1122 10:01:07.282700 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29396761-w9qjq" event={"ID":"21453b21-488b-4068-8ba4-1fa71816009d","Type":"ContainerDied","Data":"ea1f5f394962fef551350002ee2a76e2b34d171b655daefce9615b816b4f1a8e"} Nov 22 10:01:08 crc kubenswrapper[4789]: I1122 10:01:08.693452 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29396761-w9qjq" Nov 22 10:01:08 crc kubenswrapper[4789]: I1122 10:01:08.761651 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21453b21-488b-4068-8ba4-1fa71816009d-combined-ca-bundle\") pod \"21453b21-488b-4068-8ba4-1fa71816009d\" (UID: \"21453b21-488b-4068-8ba4-1fa71816009d\") " Nov 22 10:01:08 crc kubenswrapper[4789]: I1122 10:01:08.761821 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/21453b21-488b-4068-8ba4-1fa71816009d-fernet-keys\") pod \"21453b21-488b-4068-8ba4-1fa71816009d\" (UID: \"21453b21-488b-4068-8ba4-1fa71816009d\") " Nov 22 10:01:08 crc kubenswrapper[4789]: I1122 10:01:08.762878 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9hjrg\" (UniqueName: \"kubernetes.io/projected/21453b21-488b-4068-8ba4-1fa71816009d-kube-api-access-9hjrg\") pod \"21453b21-488b-4068-8ba4-1fa71816009d\" (UID: \"21453b21-488b-4068-8ba4-1fa71816009d\") " Nov 22 10:01:08 crc kubenswrapper[4789]: I1122 10:01:08.762947 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21453b21-488b-4068-8ba4-1fa71816009d-config-data\") pod \"21453b21-488b-4068-8ba4-1fa71816009d\" (UID: \"21453b21-488b-4068-8ba4-1fa71816009d\") " Nov 22 10:01:08 crc kubenswrapper[4789]: I1122 10:01:08.768392 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/21453b21-488b-4068-8ba4-1fa71816009d-kube-api-access-9hjrg" (OuterVolumeSpecName: "kube-api-access-9hjrg") pod "21453b21-488b-4068-8ba4-1fa71816009d" (UID: "21453b21-488b-4068-8ba4-1fa71816009d"). InnerVolumeSpecName "kube-api-access-9hjrg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 10:01:08 crc kubenswrapper[4789]: I1122 10:01:08.769015 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21453b21-488b-4068-8ba4-1fa71816009d-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "21453b21-488b-4068-8ba4-1fa71816009d" (UID: "21453b21-488b-4068-8ba4-1fa71816009d"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 10:01:08 crc kubenswrapper[4789]: I1122 10:01:08.805670 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21453b21-488b-4068-8ba4-1fa71816009d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "21453b21-488b-4068-8ba4-1fa71816009d" (UID: "21453b21-488b-4068-8ba4-1fa71816009d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 10:01:08 crc kubenswrapper[4789]: I1122 10:01:08.831982 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21453b21-488b-4068-8ba4-1fa71816009d-config-data" (OuterVolumeSpecName: "config-data") pod "21453b21-488b-4068-8ba4-1fa71816009d" (UID: "21453b21-488b-4068-8ba4-1fa71816009d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 10:01:08 crc kubenswrapper[4789]: I1122 10:01:08.865397 4789 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21453b21-488b-4068-8ba4-1fa71816009d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 10:01:08 crc kubenswrapper[4789]: I1122 10:01:08.865444 4789 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/21453b21-488b-4068-8ba4-1fa71816009d-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 22 10:01:08 crc kubenswrapper[4789]: I1122 10:01:08.865459 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9hjrg\" (UniqueName: \"kubernetes.io/projected/21453b21-488b-4068-8ba4-1fa71816009d-kube-api-access-9hjrg\") on node \"crc\" DevicePath \"\"" Nov 22 10:01:08 crc kubenswrapper[4789]: I1122 10:01:08.865471 4789 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21453b21-488b-4068-8ba4-1fa71816009d-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 10:01:09 crc kubenswrapper[4789]: I1122 10:01:09.304008 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29396761-w9qjq" event={"ID":"21453b21-488b-4068-8ba4-1fa71816009d","Type":"ContainerDied","Data":"008d98b91fb9b84012f793453b13fc26f8327bf0d4e7ebf7c102dc907e984987"} Nov 22 10:01:09 crc kubenswrapper[4789]: I1122 10:01:09.304475 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="008d98b91fb9b84012f793453b13fc26f8327bf0d4e7ebf7c102dc907e984987" Nov 22 10:01:09 crc kubenswrapper[4789]: I1122 10:01:09.304095 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29396761-w9qjq" Nov 22 10:01:26 crc kubenswrapper[4789]: I1122 10:01:26.796980 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-5xn8v"] Nov 22 10:01:26 crc kubenswrapper[4789]: E1122 10:01:26.797948 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bde6612-2541-46cd-a1e1-20d0736213ee" containerName="extract-content" Nov 22 10:01:26 crc kubenswrapper[4789]: I1122 10:01:26.797961 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bde6612-2541-46cd-a1e1-20d0736213ee" containerName="extract-content" Nov 22 10:01:26 crc kubenswrapper[4789]: E1122 10:01:26.797995 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bde6612-2541-46cd-a1e1-20d0736213ee" containerName="extract-utilities" Nov 22 10:01:26 crc kubenswrapper[4789]: I1122 10:01:26.798002 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bde6612-2541-46cd-a1e1-20d0736213ee" containerName="extract-utilities" Nov 22 10:01:26 crc kubenswrapper[4789]: E1122 10:01:26.798014 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bde6612-2541-46cd-a1e1-20d0736213ee" containerName="registry-server" Nov 22 10:01:26 crc kubenswrapper[4789]: I1122 10:01:26.798020 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bde6612-2541-46cd-a1e1-20d0736213ee" containerName="registry-server" Nov 22 10:01:26 crc kubenswrapper[4789]: E1122 10:01:26.798032 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21453b21-488b-4068-8ba4-1fa71816009d" containerName="keystone-cron" Nov 22 10:01:26 crc kubenswrapper[4789]: I1122 10:01:26.798038 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="21453b21-488b-4068-8ba4-1fa71816009d" containerName="keystone-cron" Nov 22 10:01:26 crc kubenswrapper[4789]: I1122 10:01:26.798190 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bde6612-2541-46cd-a1e1-20d0736213ee" containerName="registry-server" Nov 22 10:01:26 crc kubenswrapper[4789]: I1122 10:01:26.798217 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="21453b21-488b-4068-8ba4-1fa71816009d" containerName="keystone-cron" Nov 22 10:01:26 crc kubenswrapper[4789]: I1122 10:01:26.801897 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5xn8v" Nov 22 10:01:26 crc kubenswrapper[4789]: I1122 10:01:26.821622 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5xn8v"] Nov 22 10:01:26 crc kubenswrapper[4789]: I1122 10:01:26.836360 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d8c25bc7-f00d-4fa0-9ac7-4ec21c03a05d-catalog-content\") pod \"certified-operators-5xn8v\" (UID: \"d8c25bc7-f00d-4fa0-9ac7-4ec21c03a05d\") " pod="openshift-marketplace/certified-operators-5xn8v" Nov 22 10:01:26 crc kubenswrapper[4789]: I1122 10:01:26.836685 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xntbb\" (UniqueName: \"kubernetes.io/projected/d8c25bc7-f00d-4fa0-9ac7-4ec21c03a05d-kube-api-access-xntbb\") pod \"certified-operators-5xn8v\" (UID: \"d8c25bc7-f00d-4fa0-9ac7-4ec21c03a05d\") " pod="openshift-marketplace/certified-operators-5xn8v" Nov 22 10:01:26 crc kubenswrapper[4789]: I1122 10:01:26.836816 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d8c25bc7-f00d-4fa0-9ac7-4ec21c03a05d-utilities\") pod \"certified-operators-5xn8v\" (UID: \"d8c25bc7-f00d-4fa0-9ac7-4ec21c03a05d\") " pod="openshift-marketplace/certified-operators-5xn8v" Nov 22 10:01:26 crc kubenswrapper[4789]: I1122 10:01:26.938329 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d8c25bc7-f00d-4fa0-9ac7-4ec21c03a05d-catalog-content\") pod \"certified-operators-5xn8v\" (UID: \"d8c25bc7-f00d-4fa0-9ac7-4ec21c03a05d\") " pod="openshift-marketplace/certified-operators-5xn8v" Nov 22 10:01:26 crc kubenswrapper[4789]: I1122 10:01:26.938417 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xntbb\" (UniqueName: \"kubernetes.io/projected/d8c25bc7-f00d-4fa0-9ac7-4ec21c03a05d-kube-api-access-xntbb\") pod \"certified-operators-5xn8v\" (UID: \"d8c25bc7-f00d-4fa0-9ac7-4ec21c03a05d\") " pod="openshift-marketplace/certified-operators-5xn8v" Nov 22 10:01:26 crc kubenswrapper[4789]: I1122 10:01:26.938444 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d8c25bc7-f00d-4fa0-9ac7-4ec21c03a05d-utilities\") pod \"certified-operators-5xn8v\" (UID: \"d8c25bc7-f00d-4fa0-9ac7-4ec21c03a05d\") " pod="openshift-marketplace/certified-operators-5xn8v" Nov 22 10:01:26 crc kubenswrapper[4789]: I1122 10:01:26.939025 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d8c25bc7-f00d-4fa0-9ac7-4ec21c03a05d-utilities\") pod \"certified-operators-5xn8v\" (UID: \"d8c25bc7-f00d-4fa0-9ac7-4ec21c03a05d\") " pod="openshift-marketplace/certified-operators-5xn8v" Nov 22 10:01:26 crc kubenswrapper[4789]: I1122 10:01:26.939186 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d8c25bc7-f00d-4fa0-9ac7-4ec21c03a05d-catalog-content\") pod \"certified-operators-5xn8v\" (UID: \"d8c25bc7-f00d-4fa0-9ac7-4ec21c03a05d\") " pod="openshift-marketplace/certified-operators-5xn8v" Nov 22 10:01:26 crc kubenswrapper[4789]: I1122 10:01:26.958574 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xntbb\" (UniqueName: \"kubernetes.io/projected/d8c25bc7-f00d-4fa0-9ac7-4ec21c03a05d-kube-api-access-xntbb\") pod \"certified-operators-5xn8v\" (UID: \"d8c25bc7-f00d-4fa0-9ac7-4ec21c03a05d\") " pod="openshift-marketplace/certified-operators-5xn8v" Nov 22 10:01:27 crc kubenswrapper[4789]: I1122 10:01:27.133978 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5xn8v" Nov 22 10:01:27 crc kubenswrapper[4789]: I1122 10:01:27.734604 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5xn8v"] Nov 22 10:01:28 crc kubenswrapper[4789]: I1122 10:01:28.510426 4789 generic.go:334] "Generic (PLEG): container finished" podID="d8c25bc7-f00d-4fa0-9ac7-4ec21c03a05d" containerID="0d17024fc5ea4590a23935bc41095692afa75a987ed1c37e67425003df728c04" exitCode=0 Nov 22 10:01:28 crc kubenswrapper[4789]: I1122 10:01:28.510474 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5xn8v" event={"ID":"d8c25bc7-f00d-4fa0-9ac7-4ec21c03a05d","Type":"ContainerDied","Data":"0d17024fc5ea4590a23935bc41095692afa75a987ed1c37e67425003df728c04"} Nov 22 10:01:28 crc kubenswrapper[4789]: I1122 10:01:28.510725 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5xn8v" event={"ID":"d8c25bc7-f00d-4fa0-9ac7-4ec21c03a05d","Type":"ContainerStarted","Data":"54b2be938868afda2699551d74860eafde63ec8372cdac25587b01ada3ac81ad"} Nov 22 10:01:29 crc kubenswrapper[4789]: I1122 10:01:29.525066 4789 generic.go:334] "Generic (PLEG): container finished" podID="d8c25bc7-f00d-4fa0-9ac7-4ec21c03a05d" containerID="4dbaeb3823ccfc75bc940a2f76c8d8d5d00d38abbd3bf4e2f3c968e7525c226b" exitCode=0 Nov 22 10:01:29 crc kubenswrapper[4789]: I1122 10:01:29.525152 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5xn8v" event={"ID":"d8c25bc7-f00d-4fa0-9ac7-4ec21c03a05d","Type":"ContainerDied","Data":"4dbaeb3823ccfc75bc940a2f76c8d8d5d00d38abbd3bf4e2f3c968e7525c226b"} Nov 22 10:01:30 crc kubenswrapper[4789]: I1122 10:01:30.538176 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5xn8v" event={"ID":"d8c25bc7-f00d-4fa0-9ac7-4ec21c03a05d","Type":"ContainerStarted","Data":"8458ab54541f6e4a35662f1df3393d6574e634a5ab9758bb63c83d8f5a2ed1f2"} Nov 22 10:01:30 crc kubenswrapper[4789]: I1122 10:01:30.564840 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-5xn8v" podStartSLOduration=3.080195046 podStartE2EDuration="4.564810246s" podCreationTimestamp="2025-11-22 10:01:26 +0000 UTC" firstStartedPulling="2025-11-22 10:01:28.512696021 +0000 UTC m=+7502.747096314" lastFinishedPulling="2025-11-22 10:01:29.997311231 +0000 UTC m=+7504.231711514" observedRunningTime="2025-11-22 10:01:30.552400803 +0000 UTC m=+7504.786801066" watchObservedRunningTime="2025-11-22 10:01:30.564810246 +0000 UTC m=+7504.799210519" Nov 22 10:01:37 crc kubenswrapper[4789]: I1122 10:01:37.134627 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-5xn8v" Nov 22 10:01:37 crc kubenswrapper[4789]: I1122 10:01:37.135180 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-5xn8v" Nov 22 10:01:37 crc kubenswrapper[4789]: I1122 10:01:37.206562 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-5xn8v" Nov 22 10:01:37 crc kubenswrapper[4789]: I1122 10:01:37.658853 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-5xn8v" Nov 22 10:01:37 crc kubenswrapper[4789]: I1122 10:01:37.725560 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5xn8v"] Nov 22 10:01:39 crc kubenswrapper[4789]: I1122 10:01:39.632791 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-5xn8v" podUID="d8c25bc7-f00d-4fa0-9ac7-4ec21c03a05d" containerName="registry-server" containerID="cri-o://8458ab54541f6e4a35662f1df3393d6574e634a5ab9758bb63c83d8f5a2ed1f2" gracePeriod=2 Nov 22 10:01:40 crc kubenswrapper[4789]: I1122 10:01:40.438262 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5xn8v" Nov 22 10:01:40 crc kubenswrapper[4789]: I1122 10:01:40.535891 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d8c25bc7-f00d-4fa0-9ac7-4ec21c03a05d-catalog-content\") pod \"d8c25bc7-f00d-4fa0-9ac7-4ec21c03a05d\" (UID: \"d8c25bc7-f00d-4fa0-9ac7-4ec21c03a05d\") " Nov 22 10:01:40 crc kubenswrapper[4789]: I1122 10:01:40.536021 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d8c25bc7-f00d-4fa0-9ac7-4ec21c03a05d-utilities\") pod \"d8c25bc7-f00d-4fa0-9ac7-4ec21c03a05d\" (UID: \"d8c25bc7-f00d-4fa0-9ac7-4ec21c03a05d\") " Nov 22 10:01:40 crc kubenswrapper[4789]: I1122 10:01:40.536150 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xntbb\" (UniqueName: \"kubernetes.io/projected/d8c25bc7-f00d-4fa0-9ac7-4ec21c03a05d-kube-api-access-xntbb\") pod \"d8c25bc7-f00d-4fa0-9ac7-4ec21c03a05d\" (UID: \"d8c25bc7-f00d-4fa0-9ac7-4ec21c03a05d\") " Nov 22 10:01:40 crc kubenswrapper[4789]: I1122 10:01:40.538500 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d8c25bc7-f00d-4fa0-9ac7-4ec21c03a05d-utilities" (OuterVolumeSpecName: "utilities") pod "d8c25bc7-f00d-4fa0-9ac7-4ec21c03a05d" (UID: "d8c25bc7-f00d-4fa0-9ac7-4ec21c03a05d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 10:01:40 crc kubenswrapper[4789]: I1122 10:01:40.543419 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d8c25bc7-f00d-4fa0-9ac7-4ec21c03a05d-kube-api-access-xntbb" (OuterVolumeSpecName: "kube-api-access-xntbb") pod "d8c25bc7-f00d-4fa0-9ac7-4ec21c03a05d" (UID: "d8c25bc7-f00d-4fa0-9ac7-4ec21c03a05d"). InnerVolumeSpecName "kube-api-access-xntbb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 10:01:40 crc kubenswrapper[4789]: I1122 10:01:40.638434 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d8c25bc7-f00d-4fa0-9ac7-4ec21c03a05d-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 10:01:40 crc kubenswrapper[4789]: I1122 10:01:40.638469 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xntbb\" (UniqueName: \"kubernetes.io/projected/d8c25bc7-f00d-4fa0-9ac7-4ec21c03a05d-kube-api-access-xntbb\") on node \"crc\" DevicePath \"\"" Nov 22 10:01:40 crc kubenswrapper[4789]: I1122 10:01:40.643442 4789 generic.go:334] "Generic (PLEG): container finished" podID="d8c25bc7-f00d-4fa0-9ac7-4ec21c03a05d" containerID="8458ab54541f6e4a35662f1df3393d6574e634a5ab9758bb63c83d8f5a2ed1f2" exitCode=0 Nov 22 10:01:40 crc kubenswrapper[4789]: I1122 10:01:40.643490 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5xn8v" event={"ID":"d8c25bc7-f00d-4fa0-9ac7-4ec21c03a05d","Type":"ContainerDied","Data":"8458ab54541f6e4a35662f1df3393d6574e634a5ab9758bb63c83d8f5a2ed1f2"} Nov 22 10:01:40 crc kubenswrapper[4789]: I1122 10:01:40.643520 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5xn8v" Nov 22 10:01:40 crc kubenswrapper[4789]: I1122 10:01:40.643541 4789 scope.go:117] "RemoveContainer" containerID="8458ab54541f6e4a35662f1df3393d6574e634a5ab9758bb63c83d8f5a2ed1f2" Nov 22 10:01:40 crc kubenswrapper[4789]: I1122 10:01:40.643527 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5xn8v" event={"ID":"d8c25bc7-f00d-4fa0-9ac7-4ec21c03a05d","Type":"ContainerDied","Data":"54b2be938868afda2699551d74860eafde63ec8372cdac25587b01ada3ac81ad"} Nov 22 10:01:40 crc kubenswrapper[4789]: I1122 10:01:40.673988 4789 scope.go:117] "RemoveContainer" containerID="4dbaeb3823ccfc75bc940a2f76c8d8d5d00d38abbd3bf4e2f3c968e7525c226b" Nov 22 10:01:40 crc kubenswrapper[4789]: I1122 10:01:40.701526 4789 scope.go:117] "RemoveContainer" containerID="0d17024fc5ea4590a23935bc41095692afa75a987ed1c37e67425003df728c04" Nov 22 10:01:40 crc kubenswrapper[4789]: I1122 10:01:40.735677 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d8c25bc7-f00d-4fa0-9ac7-4ec21c03a05d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d8c25bc7-f00d-4fa0-9ac7-4ec21c03a05d" (UID: "d8c25bc7-f00d-4fa0-9ac7-4ec21c03a05d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 10:01:40 crc kubenswrapper[4789]: I1122 10:01:40.739665 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d8c25bc7-f00d-4fa0-9ac7-4ec21c03a05d-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 10:01:40 crc kubenswrapper[4789]: I1122 10:01:40.747718 4789 scope.go:117] "RemoveContainer" containerID="8458ab54541f6e4a35662f1df3393d6574e634a5ab9758bb63c83d8f5a2ed1f2" Nov 22 10:01:40 crc kubenswrapper[4789]: E1122 10:01:40.754193 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8458ab54541f6e4a35662f1df3393d6574e634a5ab9758bb63c83d8f5a2ed1f2\": container with ID starting with 8458ab54541f6e4a35662f1df3393d6574e634a5ab9758bb63c83d8f5a2ed1f2 not found: ID does not exist" containerID="8458ab54541f6e4a35662f1df3393d6574e634a5ab9758bb63c83d8f5a2ed1f2" Nov 22 10:01:40 crc kubenswrapper[4789]: I1122 10:01:40.754236 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8458ab54541f6e4a35662f1df3393d6574e634a5ab9758bb63c83d8f5a2ed1f2"} err="failed to get container status \"8458ab54541f6e4a35662f1df3393d6574e634a5ab9758bb63c83d8f5a2ed1f2\": rpc error: code = NotFound desc = could not find container \"8458ab54541f6e4a35662f1df3393d6574e634a5ab9758bb63c83d8f5a2ed1f2\": container with ID starting with 8458ab54541f6e4a35662f1df3393d6574e634a5ab9758bb63c83d8f5a2ed1f2 not found: ID does not exist" Nov 22 10:01:40 crc kubenswrapper[4789]: I1122 10:01:40.754260 4789 scope.go:117] "RemoveContainer" containerID="4dbaeb3823ccfc75bc940a2f76c8d8d5d00d38abbd3bf4e2f3c968e7525c226b" Nov 22 10:01:40 crc kubenswrapper[4789]: E1122 10:01:40.754544 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4dbaeb3823ccfc75bc940a2f76c8d8d5d00d38abbd3bf4e2f3c968e7525c226b\": container with ID starting with 4dbaeb3823ccfc75bc940a2f76c8d8d5d00d38abbd3bf4e2f3c968e7525c226b not found: ID does not exist" containerID="4dbaeb3823ccfc75bc940a2f76c8d8d5d00d38abbd3bf4e2f3c968e7525c226b" Nov 22 10:01:40 crc kubenswrapper[4789]: I1122 10:01:40.754567 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4dbaeb3823ccfc75bc940a2f76c8d8d5d00d38abbd3bf4e2f3c968e7525c226b"} err="failed to get container status \"4dbaeb3823ccfc75bc940a2f76c8d8d5d00d38abbd3bf4e2f3c968e7525c226b\": rpc error: code = NotFound desc = could not find container \"4dbaeb3823ccfc75bc940a2f76c8d8d5d00d38abbd3bf4e2f3c968e7525c226b\": container with ID starting with 4dbaeb3823ccfc75bc940a2f76c8d8d5d00d38abbd3bf4e2f3c968e7525c226b not found: ID does not exist" Nov 22 10:01:40 crc kubenswrapper[4789]: I1122 10:01:40.754583 4789 scope.go:117] "RemoveContainer" containerID="0d17024fc5ea4590a23935bc41095692afa75a987ed1c37e67425003df728c04" Nov 22 10:01:40 crc kubenswrapper[4789]: E1122 10:01:40.754835 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0d17024fc5ea4590a23935bc41095692afa75a987ed1c37e67425003df728c04\": container with ID starting with 0d17024fc5ea4590a23935bc41095692afa75a987ed1c37e67425003df728c04 not found: ID does not exist" containerID="0d17024fc5ea4590a23935bc41095692afa75a987ed1c37e67425003df728c04" Nov 22 10:01:40 crc kubenswrapper[4789]: I1122 10:01:40.754856 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d17024fc5ea4590a23935bc41095692afa75a987ed1c37e67425003df728c04"} err="failed to get container status \"0d17024fc5ea4590a23935bc41095692afa75a987ed1c37e67425003df728c04\": rpc error: code = NotFound desc = could not find container \"0d17024fc5ea4590a23935bc41095692afa75a987ed1c37e67425003df728c04\": container with ID starting with 0d17024fc5ea4590a23935bc41095692afa75a987ed1c37e67425003df728c04 not found: ID does not exist" Nov 22 10:01:40 crc kubenswrapper[4789]: I1122 10:01:40.990361 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5xn8v"] Nov 22 10:01:41 crc kubenswrapper[4789]: I1122 10:01:41.001085 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-5xn8v"] Nov 22 10:01:41 crc kubenswrapper[4789]: I1122 10:01:41.980254 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d8c25bc7-f00d-4fa0-9ac7-4ec21c03a05d" path="/var/lib/kubelet/pods/d8c25bc7-f00d-4fa0-9ac7-4ec21c03a05d/volumes" Nov 22 10:02:35 crc kubenswrapper[4789]: I1122 10:02:35.373291 4789 patch_prober.go:28] interesting pod/machine-config-daemon-pxvrm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 10:02:35 crc kubenswrapper[4789]: I1122 10:02:35.374095 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 10:02:36 crc kubenswrapper[4789]: I1122 10:02:36.083850 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-bdxgm"] Nov 22 10:02:36 crc kubenswrapper[4789]: E1122 10:02:36.085080 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8c25bc7-f00d-4fa0-9ac7-4ec21c03a05d" containerName="registry-server" Nov 22 10:02:36 crc kubenswrapper[4789]: I1122 10:02:36.085110 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8c25bc7-f00d-4fa0-9ac7-4ec21c03a05d" containerName="registry-server" Nov 22 10:02:36 crc kubenswrapper[4789]: E1122 10:02:36.085157 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8c25bc7-f00d-4fa0-9ac7-4ec21c03a05d" containerName="extract-content" Nov 22 10:02:36 crc kubenswrapper[4789]: I1122 10:02:36.085172 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8c25bc7-f00d-4fa0-9ac7-4ec21c03a05d" containerName="extract-content" Nov 22 10:02:36 crc kubenswrapper[4789]: E1122 10:02:36.085186 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8c25bc7-f00d-4fa0-9ac7-4ec21c03a05d" containerName="extract-utilities" Nov 22 10:02:36 crc kubenswrapper[4789]: I1122 10:02:36.085195 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8c25bc7-f00d-4fa0-9ac7-4ec21c03a05d" containerName="extract-utilities" Nov 22 10:02:36 crc kubenswrapper[4789]: I1122 10:02:36.085450 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8c25bc7-f00d-4fa0-9ac7-4ec21c03a05d" containerName="registry-server" Nov 22 10:02:36 crc kubenswrapper[4789]: I1122 10:02:36.087713 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bdxgm" Nov 22 10:02:36 crc kubenswrapper[4789]: I1122 10:02:36.108698 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bdxgm"] Nov 22 10:02:36 crc kubenswrapper[4789]: I1122 10:02:36.140095 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/31270cda-46da-459a-bec4-188f357009de-utilities\") pod \"community-operators-bdxgm\" (UID: \"31270cda-46da-459a-bec4-188f357009de\") " pod="openshift-marketplace/community-operators-bdxgm" Nov 22 10:02:36 crc kubenswrapper[4789]: I1122 10:02:36.140151 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v5b82\" (UniqueName: \"kubernetes.io/projected/31270cda-46da-459a-bec4-188f357009de-kube-api-access-v5b82\") pod \"community-operators-bdxgm\" (UID: \"31270cda-46da-459a-bec4-188f357009de\") " pod="openshift-marketplace/community-operators-bdxgm" Nov 22 10:02:36 crc kubenswrapper[4789]: I1122 10:02:36.140193 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/31270cda-46da-459a-bec4-188f357009de-catalog-content\") pod \"community-operators-bdxgm\" (UID: \"31270cda-46da-459a-bec4-188f357009de\") " pod="openshift-marketplace/community-operators-bdxgm" Nov 22 10:02:36 crc kubenswrapper[4789]: I1122 10:02:36.243167 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/31270cda-46da-459a-bec4-188f357009de-catalog-content\") pod \"community-operators-bdxgm\" (UID: \"31270cda-46da-459a-bec4-188f357009de\") " pod="openshift-marketplace/community-operators-bdxgm" Nov 22 10:02:36 crc kubenswrapper[4789]: I1122 10:02:36.243430 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/31270cda-46da-459a-bec4-188f357009de-utilities\") pod \"community-operators-bdxgm\" (UID: \"31270cda-46da-459a-bec4-188f357009de\") " pod="openshift-marketplace/community-operators-bdxgm" Nov 22 10:02:36 crc kubenswrapper[4789]: I1122 10:02:36.243460 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v5b82\" (UniqueName: \"kubernetes.io/projected/31270cda-46da-459a-bec4-188f357009de-kube-api-access-v5b82\") pod \"community-operators-bdxgm\" (UID: \"31270cda-46da-459a-bec4-188f357009de\") " pod="openshift-marketplace/community-operators-bdxgm" Nov 22 10:02:36 crc kubenswrapper[4789]: I1122 10:02:36.243984 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/31270cda-46da-459a-bec4-188f357009de-catalog-content\") pod \"community-operators-bdxgm\" (UID: \"31270cda-46da-459a-bec4-188f357009de\") " pod="openshift-marketplace/community-operators-bdxgm" Nov 22 10:02:36 crc kubenswrapper[4789]: I1122 10:02:36.244418 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/31270cda-46da-459a-bec4-188f357009de-utilities\") pod \"community-operators-bdxgm\" (UID: \"31270cda-46da-459a-bec4-188f357009de\") " pod="openshift-marketplace/community-operators-bdxgm" Nov 22 10:02:36 crc kubenswrapper[4789]: I1122 10:02:36.263012 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v5b82\" (UniqueName: \"kubernetes.io/projected/31270cda-46da-459a-bec4-188f357009de-kube-api-access-v5b82\") pod \"community-operators-bdxgm\" (UID: \"31270cda-46da-459a-bec4-188f357009de\") " pod="openshift-marketplace/community-operators-bdxgm" Nov 22 10:02:36 crc kubenswrapper[4789]: I1122 10:02:36.428103 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bdxgm" Nov 22 10:02:36 crc kubenswrapper[4789]: I1122 10:02:36.872428 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bdxgm"] Nov 22 10:02:37 crc kubenswrapper[4789]: I1122 10:02:37.260593 4789 generic.go:334] "Generic (PLEG): container finished" podID="31270cda-46da-459a-bec4-188f357009de" containerID="2921f478cd997682a2c269700f86921e812ae3a93c2e5ef9fff4fef7ebd5074a" exitCode=0 Nov 22 10:02:37 crc kubenswrapper[4789]: I1122 10:02:37.260652 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bdxgm" event={"ID":"31270cda-46da-459a-bec4-188f357009de","Type":"ContainerDied","Data":"2921f478cd997682a2c269700f86921e812ae3a93c2e5ef9fff4fef7ebd5074a"} Nov 22 10:02:37 crc kubenswrapper[4789]: I1122 10:02:37.261182 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bdxgm" event={"ID":"31270cda-46da-459a-bec4-188f357009de","Type":"ContainerStarted","Data":"e2c3ef711d874b86b72f9a597c3078ed13456b01d5db32d647fde90590e9c2b2"} Nov 22 10:02:38 crc kubenswrapper[4789]: I1122 10:02:38.273976 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bdxgm" event={"ID":"31270cda-46da-459a-bec4-188f357009de","Type":"ContainerStarted","Data":"63fc05743606154cc4e94d8ecb28f12b62dfde03c45b205b88e4f63b0442b174"} Nov 22 10:02:39 crc kubenswrapper[4789]: I1122 10:02:39.288259 4789 generic.go:334] "Generic (PLEG): container finished" podID="31270cda-46da-459a-bec4-188f357009de" containerID="63fc05743606154cc4e94d8ecb28f12b62dfde03c45b205b88e4f63b0442b174" exitCode=0 Nov 22 10:02:39 crc kubenswrapper[4789]: I1122 10:02:39.288338 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bdxgm" event={"ID":"31270cda-46da-459a-bec4-188f357009de","Type":"ContainerDied","Data":"63fc05743606154cc4e94d8ecb28f12b62dfde03c45b205b88e4f63b0442b174"} Nov 22 10:02:40 crc kubenswrapper[4789]: I1122 10:02:40.303238 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bdxgm" event={"ID":"31270cda-46da-459a-bec4-188f357009de","Type":"ContainerStarted","Data":"66448fe04dc78df5fed599a44f732ca0946095d080ec9432faec9bff409521d5"} Nov 22 10:02:40 crc kubenswrapper[4789]: I1122 10:02:40.336085 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-bdxgm" podStartSLOduration=1.70871548 podStartE2EDuration="4.336055478s" podCreationTimestamp="2025-11-22 10:02:36 +0000 UTC" firstStartedPulling="2025-11-22 10:02:37.26214219 +0000 UTC m=+7571.496542483" lastFinishedPulling="2025-11-22 10:02:39.889482208 +0000 UTC m=+7574.123882481" observedRunningTime="2025-11-22 10:02:40.329348479 +0000 UTC m=+7574.563748792" watchObservedRunningTime="2025-11-22 10:02:40.336055478 +0000 UTC m=+7574.570455791" Nov 22 10:02:46 crc kubenswrapper[4789]: I1122 10:02:46.429123 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-bdxgm" Nov 22 10:02:46 crc kubenswrapper[4789]: I1122 10:02:46.429686 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-bdxgm" Nov 22 10:02:46 crc kubenswrapper[4789]: I1122 10:02:46.475847 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-bdxgm" Nov 22 10:02:47 crc kubenswrapper[4789]: I1122 10:02:47.428054 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-bdxgm" Nov 22 10:02:47 crc kubenswrapper[4789]: I1122 10:02:47.478718 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-bdxgm"] Nov 22 10:02:49 crc kubenswrapper[4789]: I1122 10:02:49.385215 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-bdxgm" podUID="31270cda-46da-459a-bec4-188f357009de" containerName="registry-server" containerID="cri-o://66448fe04dc78df5fed599a44f732ca0946095d080ec9432faec9bff409521d5" gracePeriod=2 Nov 22 10:02:49 crc kubenswrapper[4789]: I1122 10:02:49.879057 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bdxgm" Nov 22 10:02:49 crc kubenswrapper[4789]: I1122 10:02:49.933135 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v5b82\" (UniqueName: \"kubernetes.io/projected/31270cda-46da-459a-bec4-188f357009de-kube-api-access-v5b82\") pod \"31270cda-46da-459a-bec4-188f357009de\" (UID: \"31270cda-46da-459a-bec4-188f357009de\") " Nov 22 10:02:49 crc kubenswrapper[4789]: I1122 10:02:49.933224 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/31270cda-46da-459a-bec4-188f357009de-utilities\") pod \"31270cda-46da-459a-bec4-188f357009de\" (UID: \"31270cda-46da-459a-bec4-188f357009de\") " Nov 22 10:02:49 crc kubenswrapper[4789]: I1122 10:02:49.933484 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/31270cda-46da-459a-bec4-188f357009de-catalog-content\") pod \"31270cda-46da-459a-bec4-188f357009de\" (UID: \"31270cda-46da-459a-bec4-188f357009de\") " Nov 22 10:02:49 crc kubenswrapper[4789]: I1122 10:02:49.935241 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/31270cda-46da-459a-bec4-188f357009de-utilities" (OuterVolumeSpecName: "utilities") pod "31270cda-46da-459a-bec4-188f357009de" (UID: "31270cda-46da-459a-bec4-188f357009de"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 10:02:49 crc kubenswrapper[4789]: I1122 10:02:49.947117 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31270cda-46da-459a-bec4-188f357009de-kube-api-access-v5b82" (OuterVolumeSpecName: "kube-api-access-v5b82") pod "31270cda-46da-459a-bec4-188f357009de" (UID: "31270cda-46da-459a-bec4-188f357009de"). InnerVolumeSpecName "kube-api-access-v5b82". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 10:02:50 crc kubenswrapper[4789]: I1122 10:02:50.005119 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/31270cda-46da-459a-bec4-188f357009de-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "31270cda-46da-459a-bec4-188f357009de" (UID: "31270cda-46da-459a-bec4-188f357009de"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 10:02:50 crc kubenswrapper[4789]: I1122 10:02:50.036500 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v5b82\" (UniqueName: \"kubernetes.io/projected/31270cda-46da-459a-bec4-188f357009de-kube-api-access-v5b82\") on node \"crc\" DevicePath \"\"" Nov 22 10:02:50 crc kubenswrapper[4789]: I1122 10:02:50.036544 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/31270cda-46da-459a-bec4-188f357009de-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 10:02:50 crc kubenswrapper[4789]: I1122 10:02:50.036559 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/31270cda-46da-459a-bec4-188f357009de-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 10:02:50 crc kubenswrapper[4789]: I1122 10:02:50.400608 4789 generic.go:334] "Generic (PLEG): container finished" podID="31270cda-46da-459a-bec4-188f357009de" containerID="66448fe04dc78df5fed599a44f732ca0946095d080ec9432faec9bff409521d5" exitCode=0 Nov 22 10:02:50 crc kubenswrapper[4789]: I1122 10:02:50.400685 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bdxgm" Nov 22 10:02:50 crc kubenswrapper[4789]: I1122 10:02:50.400692 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bdxgm" event={"ID":"31270cda-46da-459a-bec4-188f357009de","Type":"ContainerDied","Data":"66448fe04dc78df5fed599a44f732ca0946095d080ec9432faec9bff409521d5"} Nov 22 10:02:50 crc kubenswrapper[4789]: I1122 10:02:50.400807 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bdxgm" event={"ID":"31270cda-46da-459a-bec4-188f357009de","Type":"ContainerDied","Data":"e2c3ef711d874b86b72f9a597c3078ed13456b01d5db32d647fde90590e9c2b2"} Nov 22 10:02:50 crc kubenswrapper[4789]: I1122 10:02:50.400860 4789 scope.go:117] "RemoveContainer" containerID="66448fe04dc78df5fed599a44f732ca0946095d080ec9432faec9bff409521d5" Nov 22 10:02:50 crc kubenswrapper[4789]: I1122 10:02:50.429557 4789 scope.go:117] "RemoveContainer" containerID="63fc05743606154cc4e94d8ecb28f12b62dfde03c45b205b88e4f63b0442b174" Nov 22 10:02:50 crc kubenswrapper[4789]: I1122 10:02:50.448111 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-bdxgm"] Nov 22 10:02:50 crc kubenswrapper[4789]: I1122 10:02:50.467089 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-bdxgm"] Nov 22 10:02:50 crc kubenswrapper[4789]: I1122 10:02:50.469038 4789 scope.go:117] "RemoveContainer" containerID="2921f478cd997682a2c269700f86921e812ae3a93c2e5ef9fff4fef7ebd5074a" Nov 22 10:02:50 crc kubenswrapper[4789]: I1122 10:02:50.516596 4789 scope.go:117] "RemoveContainer" containerID="66448fe04dc78df5fed599a44f732ca0946095d080ec9432faec9bff409521d5" Nov 22 10:02:50 crc kubenswrapper[4789]: E1122 10:02:50.518052 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"66448fe04dc78df5fed599a44f732ca0946095d080ec9432faec9bff409521d5\": container with ID starting with 66448fe04dc78df5fed599a44f732ca0946095d080ec9432faec9bff409521d5 not found: ID does not exist" containerID="66448fe04dc78df5fed599a44f732ca0946095d080ec9432faec9bff409521d5" Nov 22 10:02:50 crc kubenswrapper[4789]: I1122 10:02:50.518142 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"66448fe04dc78df5fed599a44f732ca0946095d080ec9432faec9bff409521d5"} err="failed to get container status \"66448fe04dc78df5fed599a44f732ca0946095d080ec9432faec9bff409521d5\": rpc error: code = NotFound desc = could not find container \"66448fe04dc78df5fed599a44f732ca0946095d080ec9432faec9bff409521d5\": container with ID starting with 66448fe04dc78df5fed599a44f732ca0946095d080ec9432faec9bff409521d5 not found: ID does not exist" Nov 22 10:02:50 crc kubenswrapper[4789]: I1122 10:02:50.518193 4789 scope.go:117] "RemoveContainer" containerID="63fc05743606154cc4e94d8ecb28f12b62dfde03c45b205b88e4f63b0442b174" Nov 22 10:02:50 crc kubenswrapper[4789]: E1122 10:02:50.518731 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"63fc05743606154cc4e94d8ecb28f12b62dfde03c45b205b88e4f63b0442b174\": container with ID starting with 63fc05743606154cc4e94d8ecb28f12b62dfde03c45b205b88e4f63b0442b174 not found: ID does not exist" containerID="63fc05743606154cc4e94d8ecb28f12b62dfde03c45b205b88e4f63b0442b174" Nov 22 10:02:50 crc kubenswrapper[4789]: I1122 10:02:50.518833 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"63fc05743606154cc4e94d8ecb28f12b62dfde03c45b205b88e4f63b0442b174"} err="failed to get container status \"63fc05743606154cc4e94d8ecb28f12b62dfde03c45b205b88e4f63b0442b174\": rpc error: code = NotFound desc = could not find container \"63fc05743606154cc4e94d8ecb28f12b62dfde03c45b205b88e4f63b0442b174\": container with ID starting with 63fc05743606154cc4e94d8ecb28f12b62dfde03c45b205b88e4f63b0442b174 not found: ID does not exist" Nov 22 10:02:50 crc kubenswrapper[4789]: I1122 10:02:50.518860 4789 scope.go:117] "RemoveContainer" containerID="2921f478cd997682a2c269700f86921e812ae3a93c2e5ef9fff4fef7ebd5074a" Nov 22 10:02:50 crc kubenswrapper[4789]: E1122 10:02:50.519284 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2921f478cd997682a2c269700f86921e812ae3a93c2e5ef9fff4fef7ebd5074a\": container with ID starting with 2921f478cd997682a2c269700f86921e812ae3a93c2e5ef9fff4fef7ebd5074a not found: ID does not exist" containerID="2921f478cd997682a2c269700f86921e812ae3a93c2e5ef9fff4fef7ebd5074a" Nov 22 10:02:50 crc kubenswrapper[4789]: I1122 10:02:50.519327 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2921f478cd997682a2c269700f86921e812ae3a93c2e5ef9fff4fef7ebd5074a"} err="failed to get container status \"2921f478cd997682a2c269700f86921e812ae3a93c2e5ef9fff4fef7ebd5074a\": rpc error: code = NotFound desc = could not find container \"2921f478cd997682a2c269700f86921e812ae3a93c2e5ef9fff4fef7ebd5074a\": container with ID starting with 2921f478cd997682a2c269700f86921e812ae3a93c2e5ef9fff4fef7ebd5074a not found: ID does not exist" Nov 22 10:02:51 crc kubenswrapper[4789]: I1122 10:02:51.977976 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31270cda-46da-459a-bec4-188f357009de" path="/var/lib/kubelet/pods/31270cda-46da-459a-bec4-188f357009de/volumes" Nov 22 10:03:05 crc kubenswrapper[4789]: I1122 10:03:05.372707 4789 patch_prober.go:28] interesting pod/machine-config-daemon-pxvrm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 10:03:05 crc kubenswrapper[4789]: I1122 10:03:05.373330 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 10:03:35 crc kubenswrapper[4789]: I1122 10:03:35.372324 4789 patch_prober.go:28] interesting pod/machine-config-daemon-pxvrm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 10:03:35 crc kubenswrapper[4789]: I1122 10:03:35.373131 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 10:03:35 crc kubenswrapper[4789]: I1122 10:03:35.373189 4789 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" Nov 22 10:03:35 crc kubenswrapper[4789]: I1122 10:03:35.374181 4789 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7b5ed03a8e4ebb046ae3b2e18435bfca04a6eb59a1c3958782493a816f47f210"} pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 22 10:03:35 crc kubenswrapper[4789]: I1122 10:03:35.374259 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" containerID="cri-o://7b5ed03a8e4ebb046ae3b2e18435bfca04a6eb59a1c3958782493a816f47f210" gracePeriod=600 Nov 22 10:03:35 crc kubenswrapper[4789]: I1122 10:03:35.878906 4789 generic.go:334] "Generic (PLEG): container finished" podID="58aa8071-7ecc-4692-8789-2db04ec70647" containerID="7b5ed03a8e4ebb046ae3b2e18435bfca04a6eb59a1c3958782493a816f47f210" exitCode=0 Nov 22 10:03:35 crc kubenswrapper[4789]: I1122 10:03:35.879009 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" event={"ID":"58aa8071-7ecc-4692-8789-2db04ec70647","Type":"ContainerDied","Data":"7b5ed03a8e4ebb046ae3b2e18435bfca04a6eb59a1c3958782493a816f47f210"} Nov 22 10:03:35 crc kubenswrapper[4789]: I1122 10:03:35.879243 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" event={"ID":"58aa8071-7ecc-4692-8789-2db04ec70647","Type":"ContainerStarted","Data":"6e22a5a8fc3638ca6aab7b67ea550d9b289803b56baeb39c04dfbc919861ab85"} Nov 22 10:03:35 crc kubenswrapper[4789]: I1122 10:03:35.879272 4789 scope.go:117] "RemoveContainer" containerID="26c17a090dfedea5cf0834b08473176e1ddf9f8e1cbf99d58db7701ed655462d" Nov 22 10:05:35 crc kubenswrapper[4789]: I1122 10:05:35.372375 4789 patch_prober.go:28] interesting pod/machine-config-daemon-pxvrm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 10:05:35 crc kubenswrapper[4789]: I1122 10:05:35.373263 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 10:06:05 crc kubenswrapper[4789]: I1122 10:06:05.372462 4789 patch_prober.go:28] interesting pod/machine-config-daemon-pxvrm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 10:06:05 crc kubenswrapper[4789]: I1122 10:06:05.373215 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 10:06:35 crc kubenswrapper[4789]: I1122 10:06:35.372894 4789 patch_prober.go:28] interesting pod/machine-config-daemon-pxvrm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 10:06:35 crc kubenswrapper[4789]: I1122 10:06:35.373798 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 10:06:35 crc kubenswrapper[4789]: I1122 10:06:35.374010 4789 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" Nov 22 10:06:35 crc kubenswrapper[4789]: I1122 10:06:35.375546 4789 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6e22a5a8fc3638ca6aab7b67ea550d9b289803b56baeb39c04dfbc919861ab85"} pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 22 10:06:35 crc kubenswrapper[4789]: I1122 10:06:35.375662 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" containerID="cri-o://6e22a5a8fc3638ca6aab7b67ea550d9b289803b56baeb39c04dfbc919861ab85" gracePeriod=600 Nov 22 10:06:35 crc kubenswrapper[4789]: E1122 10:06:35.502407 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 10:06:36 crc kubenswrapper[4789]: I1122 10:06:36.478297 4789 generic.go:334] "Generic (PLEG): container finished" podID="58aa8071-7ecc-4692-8789-2db04ec70647" containerID="6e22a5a8fc3638ca6aab7b67ea550d9b289803b56baeb39c04dfbc919861ab85" exitCode=0 Nov 22 10:06:36 crc kubenswrapper[4789]: I1122 10:06:36.478336 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" event={"ID":"58aa8071-7ecc-4692-8789-2db04ec70647","Type":"ContainerDied","Data":"6e22a5a8fc3638ca6aab7b67ea550d9b289803b56baeb39c04dfbc919861ab85"} Nov 22 10:06:36 crc kubenswrapper[4789]: I1122 10:06:36.478565 4789 scope.go:117] "RemoveContainer" containerID="7b5ed03a8e4ebb046ae3b2e18435bfca04a6eb59a1c3958782493a816f47f210" Nov 22 10:06:36 crc kubenswrapper[4789]: I1122 10:06:36.479341 4789 scope.go:117] "RemoveContainer" containerID="6e22a5a8fc3638ca6aab7b67ea550d9b289803b56baeb39c04dfbc919861ab85" Nov 22 10:06:36 crc kubenswrapper[4789]: E1122 10:06:36.479677 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 10:06:47 crc kubenswrapper[4789]: I1122 10:06:47.971010 4789 scope.go:117] "RemoveContainer" containerID="6e22a5a8fc3638ca6aab7b67ea550d9b289803b56baeb39c04dfbc919861ab85" Nov 22 10:06:47 crc kubenswrapper[4789]: E1122 10:06:47.971607 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 10:06:59 crc kubenswrapper[4789]: I1122 10:06:59.967568 4789 scope.go:117] "RemoveContainer" containerID="6e22a5a8fc3638ca6aab7b67ea550d9b289803b56baeb39c04dfbc919861ab85" Nov 22 10:06:59 crc kubenswrapper[4789]: E1122 10:06:59.968552 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 10:07:10 crc kubenswrapper[4789]: I1122 10:07:10.965779 4789 scope.go:117] "RemoveContainer" containerID="6e22a5a8fc3638ca6aab7b67ea550d9b289803b56baeb39c04dfbc919861ab85" Nov 22 10:07:10 crc kubenswrapper[4789]: E1122 10:07:10.966576 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 10:07:22 crc kubenswrapper[4789]: I1122 10:07:22.965471 4789 scope.go:117] "RemoveContainer" containerID="6e22a5a8fc3638ca6aab7b67ea550d9b289803b56baeb39c04dfbc919861ab85" Nov 22 10:07:22 crc kubenswrapper[4789]: E1122 10:07:22.966384 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 10:07:36 crc kubenswrapper[4789]: I1122 10:07:36.139553 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-95m7j"] Nov 22 10:07:36 crc kubenswrapper[4789]: E1122 10:07:36.140952 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31270cda-46da-459a-bec4-188f357009de" containerName="extract-content" Nov 22 10:07:36 crc kubenswrapper[4789]: I1122 10:07:36.140970 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="31270cda-46da-459a-bec4-188f357009de" containerName="extract-content" Nov 22 10:07:36 crc kubenswrapper[4789]: E1122 10:07:36.141007 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31270cda-46da-459a-bec4-188f357009de" containerName="extract-utilities" Nov 22 10:07:36 crc kubenswrapper[4789]: I1122 10:07:36.141017 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="31270cda-46da-459a-bec4-188f357009de" containerName="extract-utilities" Nov 22 10:07:36 crc kubenswrapper[4789]: E1122 10:07:36.141039 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31270cda-46da-459a-bec4-188f357009de" containerName="registry-server" Nov 22 10:07:36 crc kubenswrapper[4789]: I1122 10:07:36.141049 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="31270cda-46da-459a-bec4-188f357009de" containerName="registry-server" Nov 22 10:07:36 crc kubenswrapper[4789]: I1122 10:07:36.141290 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="31270cda-46da-459a-bec4-188f357009de" containerName="registry-server" Nov 22 10:07:36 crc kubenswrapper[4789]: I1122 10:07:36.143985 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-95m7j" Nov 22 10:07:36 crc kubenswrapper[4789]: I1122 10:07:36.159704 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-95m7j"] Nov 22 10:07:36 crc kubenswrapper[4789]: I1122 10:07:36.324502 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/52e85bac-f240-4bde-947f-0165427690a5-catalog-content\") pod \"redhat-marketplace-95m7j\" (UID: \"52e85bac-f240-4bde-947f-0165427690a5\") " pod="openshift-marketplace/redhat-marketplace-95m7j" Nov 22 10:07:36 crc kubenswrapper[4789]: I1122 10:07:36.324713 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-smjrq\" (UniqueName: \"kubernetes.io/projected/52e85bac-f240-4bde-947f-0165427690a5-kube-api-access-smjrq\") pod \"redhat-marketplace-95m7j\" (UID: \"52e85bac-f240-4bde-947f-0165427690a5\") " pod="openshift-marketplace/redhat-marketplace-95m7j" Nov 22 10:07:36 crc kubenswrapper[4789]: I1122 10:07:36.324839 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/52e85bac-f240-4bde-947f-0165427690a5-utilities\") pod \"redhat-marketplace-95m7j\" (UID: \"52e85bac-f240-4bde-947f-0165427690a5\") " pod="openshift-marketplace/redhat-marketplace-95m7j" Nov 22 10:07:36 crc kubenswrapper[4789]: I1122 10:07:36.427174 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/52e85bac-f240-4bde-947f-0165427690a5-catalog-content\") pod \"redhat-marketplace-95m7j\" (UID: \"52e85bac-f240-4bde-947f-0165427690a5\") " pod="openshift-marketplace/redhat-marketplace-95m7j" Nov 22 10:07:36 crc kubenswrapper[4789]: I1122 10:07:36.427274 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-smjrq\" (UniqueName: \"kubernetes.io/projected/52e85bac-f240-4bde-947f-0165427690a5-kube-api-access-smjrq\") pod \"redhat-marketplace-95m7j\" (UID: \"52e85bac-f240-4bde-947f-0165427690a5\") " pod="openshift-marketplace/redhat-marketplace-95m7j" Nov 22 10:07:36 crc kubenswrapper[4789]: I1122 10:07:36.427352 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/52e85bac-f240-4bde-947f-0165427690a5-utilities\") pod \"redhat-marketplace-95m7j\" (UID: \"52e85bac-f240-4bde-947f-0165427690a5\") " pod="openshift-marketplace/redhat-marketplace-95m7j" Nov 22 10:07:36 crc kubenswrapper[4789]: I1122 10:07:36.427700 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/52e85bac-f240-4bde-947f-0165427690a5-catalog-content\") pod \"redhat-marketplace-95m7j\" (UID: \"52e85bac-f240-4bde-947f-0165427690a5\") " pod="openshift-marketplace/redhat-marketplace-95m7j" Nov 22 10:07:36 crc kubenswrapper[4789]: I1122 10:07:36.427792 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/52e85bac-f240-4bde-947f-0165427690a5-utilities\") pod \"redhat-marketplace-95m7j\" (UID: \"52e85bac-f240-4bde-947f-0165427690a5\") " pod="openshift-marketplace/redhat-marketplace-95m7j" Nov 22 10:07:36 crc kubenswrapper[4789]: I1122 10:07:36.451987 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-smjrq\" (UniqueName: \"kubernetes.io/projected/52e85bac-f240-4bde-947f-0165427690a5-kube-api-access-smjrq\") pod \"redhat-marketplace-95m7j\" (UID: \"52e85bac-f240-4bde-947f-0165427690a5\") " pod="openshift-marketplace/redhat-marketplace-95m7j" Nov 22 10:07:36 crc kubenswrapper[4789]: I1122 10:07:36.475867 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-95m7j" Nov 22 10:07:36 crc kubenswrapper[4789]: I1122 10:07:36.927575 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-95m7j"] Nov 22 10:07:37 crc kubenswrapper[4789]: I1122 10:07:37.066048 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-95m7j" event={"ID":"52e85bac-f240-4bde-947f-0165427690a5","Type":"ContainerStarted","Data":"27e2af6bebb25c78f99a79aa7e08f29baa08b314f4c399924dede9eb88915890"} Nov 22 10:07:37 crc kubenswrapper[4789]: I1122 10:07:37.973419 4789 scope.go:117] "RemoveContainer" containerID="6e22a5a8fc3638ca6aab7b67ea550d9b289803b56baeb39c04dfbc919861ab85" Nov 22 10:07:37 crc kubenswrapper[4789]: E1122 10:07:37.974263 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 10:07:38 crc kubenswrapper[4789]: I1122 10:07:38.076582 4789 generic.go:334] "Generic (PLEG): container finished" podID="52e85bac-f240-4bde-947f-0165427690a5" containerID="5ab3178e3276ff581bb3f66058c7ecc40d24e268f3798e8ebbd965d2b96e6abd" exitCode=0 Nov 22 10:07:38 crc kubenswrapper[4789]: I1122 10:07:38.076645 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-95m7j" event={"ID":"52e85bac-f240-4bde-947f-0165427690a5","Type":"ContainerDied","Data":"5ab3178e3276ff581bb3f66058c7ecc40d24e268f3798e8ebbd965d2b96e6abd"} Nov 22 10:07:38 crc kubenswrapper[4789]: I1122 10:07:38.080283 4789 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 22 10:07:39 crc kubenswrapper[4789]: E1122 10:07:39.969620 4789 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod52e85bac_f240_4bde_947f_0165427690a5.slice/crio-conmon-08c1061d7b04369a5549e6d2c784abe1589f4ae1b9d65c706a3edba7c3492ce1.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod52e85bac_f240_4bde_947f_0165427690a5.slice/crio-08c1061d7b04369a5549e6d2c784abe1589f4ae1b9d65c706a3edba7c3492ce1.scope\": RecentStats: unable to find data in memory cache]" Nov 22 10:07:40 crc kubenswrapper[4789]: I1122 10:07:40.097638 4789 generic.go:334] "Generic (PLEG): container finished" podID="52e85bac-f240-4bde-947f-0165427690a5" containerID="08c1061d7b04369a5549e6d2c784abe1589f4ae1b9d65c706a3edba7c3492ce1" exitCode=0 Nov 22 10:07:40 crc kubenswrapper[4789]: I1122 10:07:40.097693 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-95m7j" event={"ID":"52e85bac-f240-4bde-947f-0165427690a5","Type":"ContainerDied","Data":"08c1061d7b04369a5549e6d2c784abe1589f4ae1b9d65c706a3edba7c3492ce1"} Nov 22 10:07:42 crc kubenswrapper[4789]: I1122 10:07:42.121437 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-95m7j" event={"ID":"52e85bac-f240-4bde-947f-0165427690a5","Type":"ContainerStarted","Data":"346e1660d8b265d88b0e2b3a49912c06bb08d047304fbd87cbeb296f762faa1d"} Nov 22 10:07:46 crc kubenswrapper[4789]: I1122 10:07:46.475944 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-95m7j" Nov 22 10:07:46 crc kubenswrapper[4789]: I1122 10:07:46.476780 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-95m7j" Nov 22 10:07:46 crc kubenswrapper[4789]: I1122 10:07:46.523336 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-95m7j" Nov 22 10:07:46 crc kubenswrapper[4789]: I1122 10:07:46.544310 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-95m7j" podStartSLOduration=7.454136485 podStartE2EDuration="10.544293199s" podCreationTimestamp="2025-11-22 10:07:36 +0000 UTC" firstStartedPulling="2025-11-22 10:07:38.079944945 +0000 UTC m=+7872.314345228" lastFinishedPulling="2025-11-22 10:07:41.170101669 +0000 UTC m=+7875.404501942" observedRunningTime="2025-11-22 10:07:42.154530789 +0000 UTC m=+7876.388931062" watchObservedRunningTime="2025-11-22 10:07:46.544293199 +0000 UTC m=+7880.778693472" Nov 22 10:07:47 crc kubenswrapper[4789]: I1122 10:07:47.225327 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-95m7j" Nov 22 10:07:47 crc kubenswrapper[4789]: I1122 10:07:47.271208 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-95m7j"] Nov 22 10:07:49 crc kubenswrapper[4789]: I1122 10:07:49.212873 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-95m7j" podUID="52e85bac-f240-4bde-947f-0165427690a5" containerName="registry-server" containerID="cri-o://346e1660d8b265d88b0e2b3a49912c06bb08d047304fbd87cbeb296f762faa1d" gracePeriod=2 Nov 22 10:07:50 crc kubenswrapper[4789]: I1122 10:07:50.225550 4789 generic.go:334] "Generic (PLEG): container finished" podID="52e85bac-f240-4bde-947f-0165427690a5" containerID="346e1660d8b265d88b0e2b3a49912c06bb08d047304fbd87cbeb296f762faa1d" exitCode=0 Nov 22 10:07:50 crc kubenswrapper[4789]: I1122 10:07:50.225924 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-95m7j" event={"ID":"52e85bac-f240-4bde-947f-0165427690a5","Type":"ContainerDied","Data":"346e1660d8b265d88b0e2b3a49912c06bb08d047304fbd87cbeb296f762faa1d"} Nov 22 10:07:50 crc kubenswrapper[4789]: I1122 10:07:50.444597 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-95m7j" Nov 22 10:07:50 crc kubenswrapper[4789]: I1122 10:07:50.541854 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/52e85bac-f240-4bde-947f-0165427690a5-catalog-content\") pod \"52e85bac-f240-4bde-947f-0165427690a5\" (UID: \"52e85bac-f240-4bde-947f-0165427690a5\") " Nov 22 10:07:50 crc kubenswrapper[4789]: I1122 10:07:50.542052 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/52e85bac-f240-4bde-947f-0165427690a5-utilities\") pod \"52e85bac-f240-4bde-947f-0165427690a5\" (UID: \"52e85bac-f240-4bde-947f-0165427690a5\") " Nov 22 10:07:50 crc kubenswrapper[4789]: I1122 10:07:50.542210 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-smjrq\" (UniqueName: \"kubernetes.io/projected/52e85bac-f240-4bde-947f-0165427690a5-kube-api-access-smjrq\") pod \"52e85bac-f240-4bde-947f-0165427690a5\" (UID: \"52e85bac-f240-4bde-947f-0165427690a5\") " Nov 22 10:07:50 crc kubenswrapper[4789]: I1122 10:07:50.544258 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/52e85bac-f240-4bde-947f-0165427690a5-utilities" (OuterVolumeSpecName: "utilities") pod "52e85bac-f240-4bde-947f-0165427690a5" (UID: "52e85bac-f240-4bde-947f-0165427690a5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 10:07:50 crc kubenswrapper[4789]: I1122 10:07:50.548032 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/52e85bac-f240-4bde-947f-0165427690a5-kube-api-access-smjrq" (OuterVolumeSpecName: "kube-api-access-smjrq") pod "52e85bac-f240-4bde-947f-0165427690a5" (UID: "52e85bac-f240-4bde-947f-0165427690a5"). InnerVolumeSpecName "kube-api-access-smjrq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 10:07:50 crc kubenswrapper[4789]: I1122 10:07:50.644852 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-smjrq\" (UniqueName: \"kubernetes.io/projected/52e85bac-f240-4bde-947f-0165427690a5-kube-api-access-smjrq\") on node \"crc\" DevicePath \"\"" Nov 22 10:07:50 crc kubenswrapper[4789]: I1122 10:07:50.644891 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/52e85bac-f240-4bde-947f-0165427690a5-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 10:07:51 crc kubenswrapper[4789]: I1122 10:07:51.243124 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-95m7j" event={"ID":"52e85bac-f240-4bde-947f-0165427690a5","Type":"ContainerDied","Data":"27e2af6bebb25c78f99a79aa7e08f29baa08b314f4c399924dede9eb88915890"} Nov 22 10:07:51 crc kubenswrapper[4789]: I1122 10:07:51.243212 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-95m7j" Nov 22 10:07:51 crc kubenswrapper[4789]: I1122 10:07:51.243625 4789 scope.go:117] "RemoveContainer" containerID="346e1660d8b265d88b0e2b3a49912c06bb08d047304fbd87cbeb296f762faa1d" Nov 22 10:07:51 crc kubenswrapper[4789]: I1122 10:07:51.275302 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/52e85bac-f240-4bde-947f-0165427690a5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "52e85bac-f240-4bde-947f-0165427690a5" (UID: "52e85bac-f240-4bde-947f-0165427690a5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 10:07:51 crc kubenswrapper[4789]: I1122 10:07:51.282396 4789 scope.go:117] "RemoveContainer" containerID="08c1061d7b04369a5549e6d2c784abe1589f4ae1b9d65c706a3edba7c3492ce1" Nov 22 10:07:51 crc kubenswrapper[4789]: I1122 10:07:51.327321 4789 scope.go:117] "RemoveContainer" containerID="5ab3178e3276ff581bb3f66058c7ecc40d24e268f3798e8ebbd965d2b96e6abd" Nov 22 10:07:51 crc kubenswrapper[4789]: I1122 10:07:51.359940 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/52e85bac-f240-4bde-947f-0165427690a5-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 10:07:51 crc kubenswrapper[4789]: I1122 10:07:51.596878 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-95m7j"] Nov 22 10:07:51 crc kubenswrapper[4789]: I1122 10:07:51.609063 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-95m7j"] Nov 22 10:07:51 crc kubenswrapper[4789]: I1122 10:07:51.965518 4789 scope.go:117] "RemoveContainer" containerID="6e22a5a8fc3638ca6aab7b67ea550d9b289803b56baeb39c04dfbc919861ab85" Nov 22 10:07:51 crc kubenswrapper[4789]: E1122 10:07:51.965903 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 10:07:51 crc kubenswrapper[4789]: I1122 10:07:51.977888 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="52e85bac-f240-4bde-947f-0165427690a5" path="/var/lib/kubelet/pods/52e85bac-f240-4bde-947f-0165427690a5/volumes" Nov 22 10:08:05 crc kubenswrapper[4789]: I1122 10:08:05.966609 4789 scope.go:117] "RemoveContainer" containerID="6e22a5a8fc3638ca6aab7b67ea550d9b289803b56baeb39c04dfbc919861ab85" Nov 22 10:08:05 crc kubenswrapper[4789]: E1122 10:08:05.968242 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 10:08:09 crc kubenswrapper[4789]: I1122 10:08:09.478332 4789 generic.go:334] "Generic (PLEG): container finished" podID="21bbc026-9e56-43bc-a3bf-0ec807812dba" containerID="c489e6990bf44b2984c49f082d91206b68ee60223dc7f75ea96ca87bd67e5266" exitCode=123 Nov 22 10:08:09 crc kubenswrapper[4789]: I1122 10:08:09.478548 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest-s01-single-test" event={"ID":"21bbc026-9e56-43bc-a3bf-0ec807812dba","Type":"ContainerDied","Data":"c489e6990bf44b2984c49f082d91206b68ee60223dc7f75ea96ca87bd67e5266"} Nov 22 10:08:10 crc kubenswrapper[4789]: I1122 10:08:10.992356 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest-s01-single-test" Nov 22 10:08:11 crc kubenswrapper[4789]: I1122 10:08:11.004422 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/21bbc026-9e56-43bc-a3bf-0ec807812dba-ceph\") pod \"21bbc026-9e56-43bc-a3bf-0ec807812dba\" (UID: \"21bbc026-9e56-43bc-a3bf-0ec807812dba\") " Nov 22 10:08:11 crc kubenswrapper[4789]: I1122 10:08:11.004522 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/21bbc026-9e56-43bc-a3bf-0ec807812dba-test-operator-ephemeral-temporary\") pod \"21bbc026-9e56-43bc-a3bf-0ec807812dba\" (UID: \"21bbc026-9e56-43bc-a3bf-0ec807812dba\") " Nov 22 10:08:11 crc kubenswrapper[4789]: I1122 10:08:11.004564 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/21bbc026-9e56-43bc-a3bf-0ec807812dba-config-data\") pod \"21bbc026-9e56-43bc-a3bf-0ec807812dba\" (UID: \"21bbc026-9e56-43bc-a3bf-0ec807812dba\") " Nov 22 10:08:11 crc kubenswrapper[4789]: I1122 10:08:11.004602 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/21bbc026-9e56-43bc-a3bf-0ec807812dba-openstack-config-secret\") pod \"21bbc026-9e56-43bc-a3bf-0ec807812dba\" (UID: \"21bbc026-9e56-43bc-a3bf-0ec807812dba\") " Nov 22 10:08:11 crc kubenswrapper[4789]: I1122 10:08:11.004692 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/21bbc026-9e56-43bc-a3bf-0ec807812dba-openstack-config\") pod \"21bbc026-9e56-43bc-a3bf-0ec807812dba\" (UID: \"21bbc026-9e56-43bc-a3bf-0ec807812dba\") " Nov 22 10:08:11 crc kubenswrapper[4789]: I1122 10:08:11.004849 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hvj9q\" (UniqueName: \"kubernetes.io/projected/21bbc026-9e56-43bc-a3bf-0ec807812dba-kube-api-access-hvj9q\") pod \"21bbc026-9e56-43bc-a3bf-0ec807812dba\" (UID: \"21bbc026-9e56-43bc-a3bf-0ec807812dba\") " Nov 22 10:08:11 crc kubenswrapper[4789]: I1122 10:08:11.005014 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"21bbc026-9e56-43bc-a3bf-0ec807812dba\" (UID: \"21bbc026-9e56-43bc-a3bf-0ec807812dba\") " Nov 22 10:08:11 crc kubenswrapper[4789]: I1122 10:08:11.005146 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/21bbc026-9e56-43bc-a3bf-0ec807812dba-test-operator-ephemeral-workdir\") pod \"21bbc026-9e56-43bc-a3bf-0ec807812dba\" (UID: \"21bbc026-9e56-43bc-a3bf-0ec807812dba\") " Nov 22 10:08:11 crc kubenswrapper[4789]: I1122 10:08:11.005355 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/21bbc026-9e56-43bc-a3bf-0ec807812dba-ssh-key\") pod \"21bbc026-9e56-43bc-a3bf-0ec807812dba\" (UID: \"21bbc026-9e56-43bc-a3bf-0ec807812dba\") " Nov 22 10:08:11 crc kubenswrapper[4789]: I1122 10:08:11.005455 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/21bbc026-9e56-43bc-a3bf-0ec807812dba-ca-certs\") pod \"21bbc026-9e56-43bc-a3bf-0ec807812dba\" (UID: \"21bbc026-9e56-43bc-a3bf-0ec807812dba\") " Nov 22 10:08:11 crc kubenswrapper[4789]: I1122 10:08:11.005652 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/21bbc026-9e56-43bc-a3bf-0ec807812dba-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "21bbc026-9e56-43bc-a3bf-0ec807812dba" (UID: "21bbc026-9e56-43bc-a3bf-0ec807812dba"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 10:08:11 crc kubenswrapper[4789]: I1122 10:08:11.006499 4789 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/21bbc026-9e56-43bc-a3bf-0ec807812dba-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Nov 22 10:08:11 crc kubenswrapper[4789]: I1122 10:08:11.006560 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/21bbc026-9e56-43bc-a3bf-0ec807812dba-config-data" (OuterVolumeSpecName: "config-data") pod "21bbc026-9e56-43bc-a3bf-0ec807812dba" (UID: "21bbc026-9e56-43bc-a3bf-0ec807812dba"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 10:08:11 crc kubenswrapper[4789]: I1122 10:08:11.007112 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/21bbc026-9e56-43bc-a3bf-0ec807812dba-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "21bbc026-9e56-43bc-a3bf-0ec807812dba" (UID: "21bbc026-9e56-43bc-a3bf-0ec807812dba"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 10:08:11 crc kubenswrapper[4789]: I1122 10:08:11.021446 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21bbc026-9e56-43bc-a3bf-0ec807812dba-ceph" (OuterVolumeSpecName: "ceph") pod "21bbc026-9e56-43bc-a3bf-0ec807812dba" (UID: "21bbc026-9e56-43bc-a3bf-0ec807812dba"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 10:08:11 crc kubenswrapper[4789]: I1122 10:08:11.021748 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/21bbc026-9e56-43bc-a3bf-0ec807812dba-kube-api-access-hvj9q" (OuterVolumeSpecName: "kube-api-access-hvj9q") pod "21bbc026-9e56-43bc-a3bf-0ec807812dba" (UID: "21bbc026-9e56-43bc-a3bf-0ec807812dba"). InnerVolumeSpecName "kube-api-access-hvj9q". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 10:08:11 crc kubenswrapper[4789]: I1122 10:08:11.027291 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "test-operator-logs") pod "21bbc026-9e56-43bc-a3bf-0ec807812dba" (UID: "21bbc026-9e56-43bc-a3bf-0ec807812dba"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 22 10:08:11 crc kubenswrapper[4789]: I1122 10:08:11.057428 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21bbc026-9e56-43bc-a3bf-0ec807812dba-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "21bbc026-9e56-43bc-a3bf-0ec807812dba" (UID: "21bbc026-9e56-43bc-a3bf-0ec807812dba"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 10:08:11 crc kubenswrapper[4789]: I1122 10:08:11.065384 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21bbc026-9e56-43bc-a3bf-0ec807812dba-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "21bbc026-9e56-43bc-a3bf-0ec807812dba" (UID: "21bbc026-9e56-43bc-a3bf-0ec807812dba"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 10:08:11 crc kubenswrapper[4789]: I1122 10:08:11.072154 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21bbc026-9e56-43bc-a3bf-0ec807812dba-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "21bbc026-9e56-43bc-a3bf-0ec807812dba" (UID: "21bbc026-9e56-43bc-a3bf-0ec807812dba"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 10:08:11 crc kubenswrapper[4789]: I1122 10:08:11.093278 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/21bbc026-9e56-43bc-a3bf-0ec807812dba-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "21bbc026-9e56-43bc-a3bf-0ec807812dba" (UID: "21bbc026-9e56-43bc-a3bf-0ec807812dba"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 10:08:11 crc kubenswrapper[4789]: I1122 10:08:11.107930 4789 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/21bbc026-9e56-43bc-a3bf-0ec807812dba-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 10:08:11 crc kubenswrapper[4789]: I1122 10:08:11.107965 4789 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/21bbc026-9e56-43bc-a3bf-0ec807812dba-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Nov 22 10:08:11 crc kubenswrapper[4789]: I1122 10:08:11.107975 4789 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/21bbc026-9e56-43bc-a3bf-0ec807812dba-openstack-config\") on node \"crc\" DevicePath \"\"" Nov 22 10:08:11 crc kubenswrapper[4789]: I1122 10:08:11.107985 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hvj9q\" (UniqueName: \"kubernetes.io/projected/21bbc026-9e56-43bc-a3bf-0ec807812dba-kube-api-access-hvj9q\") on node \"crc\" DevicePath \"\"" Nov 22 10:08:11 crc kubenswrapper[4789]: I1122 10:08:11.108021 4789 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Nov 22 10:08:11 crc kubenswrapper[4789]: I1122 10:08:11.108033 4789 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/21bbc026-9e56-43bc-a3bf-0ec807812dba-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Nov 22 10:08:11 crc kubenswrapper[4789]: I1122 10:08:11.108043 4789 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/21bbc026-9e56-43bc-a3bf-0ec807812dba-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 22 10:08:11 crc kubenswrapper[4789]: I1122 10:08:11.108055 4789 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/21bbc026-9e56-43bc-a3bf-0ec807812dba-ca-certs\") on node \"crc\" DevicePath \"\"" Nov 22 10:08:11 crc kubenswrapper[4789]: I1122 10:08:11.108064 4789 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/21bbc026-9e56-43bc-a3bf-0ec807812dba-ceph\") on node \"crc\" DevicePath \"\"" Nov 22 10:08:11 crc kubenswrapper[4789]: I1122 10:08:11.130846 4789 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Nov 22 10:08:11 crc kubenswrapper[4789]: I1122 10:08:11.209291 4789 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Nov 22 10:08:11 crc kubenswrapper[4789]: I1122 10:08:11.507868 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest-s01-single-test" event={"ID":"21bbc026-9e56-43bc-a3bf-0ec807812dba","Type":"ContainerDied","Data":"4a2602ea296cf37f885cb27557be9e340b331a59541e8753479b8f43bdbb2abc"} Nov 22 10:08:11 crc kubenswrapper[4789]: I1122 10:08:11.508212 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4a2602ea296cf37f885cb27557be9e340b331a59541e8753479b8f43bdbb2abc" Nov 22 10:08:11 crc kubenswrapper[4789]: I1122 10:08:11.508015 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest-s01-single-test" Nov 22 10:08:17 crc kubenswrapper[4789]: I1122 10:08:17.840745 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Nov 22 10:08:17 crc kubenswrapper[4789]: E1122 10:08:17.841803 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52e85bac-f240-4bde-947f-0165427690a5" containerName="extract-content" Nov 22 10:08:17 crc kubenswrapper[4789]: I1122 10:08:17.841818 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="52e85bac-f240-4bde-947f-0165427690a5" containerName="extract-content" Nov 22 10:08:17 crc kubenswrapper[4789]: E1122 10:08:17.841840 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21bbc026-9e56-43bc-a3bf-0ec807812dba" containerName="tempest-tests-tempest-tests-runner" Nov 22 10:08:17 crc kubenswrapper[4789]: I1122 10:08:17.841846 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="21bbc026-9e56-43bc-a3bf-0ec807812dba" containerName="tempest-tests-tempest-tests-runner" Nov 22 10:08:17 crc kubenswrapper[4789]: E1122 10:08:17.841865 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52e85bac-f240-4bde-947f-0165427690a5" containerName="extract-utilities" Nov 22 10:08:17 crc kubenswrapper[4789]: I1122 10:08:17.841871 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="52e85bac-f240-4bde-947f-0165427690a5" containerName="extract-utilities" Nov 22 10:08:17 crc kubenswrapper[4789]: E1122 10:08:17.841886 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52e85bac-f240-4bde-947f-0165427690a5" containerName="registry-server" Nov 22 10:08:17 crc kubenswrapper[4789]: I1122 10:08:17.841891 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="52e85bac-f240-4bde-947f-0165427690a5" containerName="registry-server" Nov 22 10:08:17 crc kubenswrapper[4789]: I1122 10:08:17.842066 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="52e85bac-f240-4bde-947f-0165427690a5" containerName="registry-server" Nov 22 10:08:17 crc kubenswrapper[4789]: I1122 10:08:17.842080 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="21bbc026-9e56-43bc-a3bf-0ec807812dba" containerName="tempest-tests-tempest-tests-runner" Nov 22 10:08:17 crc kubenswrapper[4789]: I1122 10:08:17.842791 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 22 10:08:17 crc kubenswrapper[4789]: I1122 10:08:17.845978 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-49sx9" Nov 22 10:08:17 crc kubenswrapper[4789]: I1122 10:08:17.850950 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Nov 22 10:08:17 crc kubenswrapper[4789]: I1122 10:08:17.951904 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8c265\" (UniqueName: \"kubernetes.io/projected/b7a4abc5-a120-4e83-ae40-efd239ae25fe-kube-api-access-8c265\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"b7a4abc5-a120-4e83-ae40-efd239ae25fe\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 22 10:08:17 crc kubenswrapper[4789]: I1122 10:08:17.952209 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"b7a4abc5-a120-4e83-ae40-efd239ae25fe\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 22 10:08:18 crc kubenswrapper[4789]: I1122 10:08:18.054263 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8c265\" (UniqueName: \"kubernetes.io/projected/b7a4abc5-a120-4e83-ae40-efd239ae25fe-kube-api-access-8c265\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"b7a4abc5-a120-4e83-ae40-efd239ae25fe\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 22 10:08:18 crc kubenswrapper[4789]: I1122 10:08:18.054528 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"b7a4abc5-a120-4e83-ae40-efd239ae25fe\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 22 10:08:18 crc kubenswrapper[4789]: I1122 10:08:18.055070 4789 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"b7a4abc5-a120-4e83-ae40-efd239ae25fe\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 22 10:08:18 crc kubenswrapper[4789]: I1122 10:08:18.093058 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"b7a4abc5-a120-4e83-ae40-efd239ae25fe\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 22 10:08:18 crc kubenswrapper[4789]: I1122 10:08:18.093977 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8c265\" (UniqueName: \"kubernetes.io/projected/b7a4abc5-a120-4e83-ae40-efd239ae25fe-kube-api-access-8c265\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"b7a4abc5-a120-4e83-ae40-efd239ae25fe\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 22 10:08:18 crc kubenswrapper[4789]: I1122 10:08:18.172266 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 22 10:08:18 crc kubenswrapper[4789]: I1122 10:08:18.615837 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Nov 22 10:08:19 crc kubenswrapper[4789]: I1122 10:08:19.590253 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"b7a4abc5-a120-4e83-ae40-efd239ae25fe","Type":"ContainerStarted","Data":"7fdd510991eb5f533eb04613b3ce2d7cc80cba297197481c57373a223e8b1b95"} Nov 22 10:08:19 crc kubenswrapper[4789]: I1122 10:08:19.965795 4789 scope.go:117] "RemoveContainer" containerID="6e22a5a8fc3638ca6aab7b67ea550d9b289803b56baeb39c04dfbc919861ab85" Nov 22 10:08:19 crc kubenswrapper[4789]: E1122 10:08:19.966480 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 10:08:20 crc kubenswrapper[4789]: I1122 10:08:20.599189 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"b7a4abc5-a120-4e83-ae40-efd239ae25fe","Type":"ContainerStarted","Data":"d211d79e7e37a604c5026422d89b1143e72b3dbf1fbc4f4d50fd5def7666c5bf"} Nov 22 10:08:20 crc kubenswrapper[4789]: I1122 10:08:20.614027 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=2.669159361 podStartE2EDuration="3.614008589s" podCreationTimestamp="2025-11-22 10:08:17 +0000 UTC" firstStartedPulling="2025-11-22 10:08:18.620165388 +0000 UTC m=+7912.854565671" lastFinishedPulling="2025-11-22 10:08:19.565014626 +0000 UTC m=+7913.799414899" observedRunningTime="2025-11-22 10:08:20.613032812 +0000 UTC m=+7914.847433085" watchObservedRunningTime="2025-11-22 10:08:20.614008589 +0000 UTC m=+7914.848408862" Nov 22 10:08:30 crc kubenswrapper[4789]: I1122 10:08:30.965493 4789 scope.go:117] "RemoveContainer" containerID="6e22a5a8fc3638ca6aab7b67ea550d9b289803b56baeb39c04dfbc919861ab85" Nov 22 10:08:30 crc kubenswrapper[4789]: E1122 10:08:30.966412 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 10:08:37 crc kubenswrapper[4789]: I1122 10:08:37.028646 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tobiko-tests-tobiko-s00-podified-functional"] Nov 22 10:08:37 crc kubenswrapper[4789]: I1122 10:08:37.031815 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Nov 22 10:08:37 crc kubenswrapper[4789]: I1122 10:08:37.034431 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tobiko-tests-tobikotobiko-config" Nov 22 10:08:37 crc kubenswrapper[4789]: I1122 10:08:37.034977 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"tobiko-secret" Nov 22 10:08:37 crc kubenswrapper[4789]: I1122 10:08:37.037079 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tobiko-tests-tobikotobiko-public-key" Nov 22 10:08:37 crc kubenswrapper[4789]: I1122 10:08:37.037179 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"test-operator-clouds-config" Nov 22 10:08:37 crc kubenswrapper[4789]: I1122 10:08:37.039089 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tobiko-tests-tobikotobiko-private-key" Nov 22 10:08:37 crc kubenswrapper[4789]: I1122 10:08:37.052843 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tobiko-tests-tobiko-s00-podified-functional"] Nov 22 10:08:37 crc kubenswrapper[4789]: I1122 10:08:37.134102 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d6bgv\" (UniqueName: \"kubernetes.io/projected/68695f3e-75c7-4752-a896-dbec939c0dd6-kube-api-access-d6bgv\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"68695f3e-75c7-4752-a896-dbec939c0dd6\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Nov 22 10:08:37 crc kubenswrapper[4789]: I1122 10:08:37.134192 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tobiko-private-key\" (UniqueName: \"kubernetes.io/configmap/68695f3e-75c7-4752-a896-dbec939c0dd6-tobiko-private-key\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"68695f3e-75c7-4752-a896-dbec939c0dd6\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Nov 22 10:08:37 crc kubenswrapper[4789]: I1122 10:08:37.134259 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/secret/68695f3e-75c7-4752-a896-dbec939c0dd6-kubeconfig\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"68695f3e-75c7-4752-a896-dbec939c0dd6\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Nov 22 10:08:37 crc kubenswrapper[4789]: I1122 10:08:37.134296 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/68695f3e-75c7-4752-a896-dbec939c0dd6-test-operator-clouds-config\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"68695f3e-75c7-4752-a896-dbec939c0dd6\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Nov 22 10:08:37 crc kubenswrapper[4789]: I1122 10:08:37.134322 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"68695f3e-75c7-4752-a896-dbec939c0dd6\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Nov 22 10:08:37 crc kubenswrapper[4789]: I1122 10:08:37.134442 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/68695f3e-75c7-4752-a896-dbec939c0dd6-ca-certs\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"68695f3e-75c7-4752-a896-dbec939c0dd6\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Nov 22 10:08:37 crc kubenswrapper[4789]: I1122 10:08:37.134565 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/68695f3e-75c7-4752-a896-dbec939c0dd6-test-operator-ephemeral-temporary\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"68695f3e-75c7-4752-a896-dbec939c0dd6\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Nov 22 10:08:37 crc kubenswrapper[4789]: I1122 10:08:37.134626 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tobiko-config\" (UniqueName: \"kubernetes.io/configmap/68695f3e-75c7-4752-a896-dbec939c0dd6-tobiko-config\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"68695f3e-75c7-4752-a896-dbec939c0dd6\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Nov 22 10:08:37 crc kubenswrapper[4789]: I1122 10:08:37.134729 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tobiko-public-key\" (UniqueName: \"kubernetes.io/configmap/68695f3e-75c7-4752-a896-dbec939c0dd6-tobiko-public-key\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"68695f3e-75c7-4752-a896-dbec939c0dd6\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Nov 22 10:08:37 crc kubenswrapper[4789]: I1122 10:08:37.134796 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/68695f3e-75c7-4752-a896-dbec939c0dd6-ceph\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"68695f3e-75c7-4752-a896-dbec939c0dd6\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Nov 22 10:08:37 crc kubenswrapper[4789]: I1122 10:08:37.134848 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/68695f3e-75c7-4752-a896-dbec939c0dd6-openstack-config-secret\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"68695f3e-75c7-4752-a896-dbec939c0dd6\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Nov 22 10:08:37 crc kubenswrapper[4789]: I1122 10:08:37.134883 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/68695f3e-75c7-4752-a896-dbec939c0dd6-test-operator-ephemeral-workdir\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"68695f3e-75c7-4752-a896-dbec939c0dd6\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Nov 22 10:08:37 crc kubenswrapper[4789]: I1122 10:08:37.236862 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/secret/68695f3e-75c7-4752-a896-dbec939c0dd6-kubeconfig\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"68695f3e-75c7-4752-a896-dbec939c0dd6\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Nov 22 10:08:37 crc kubenswrapper[4789]: I1122 10:08:37.236926 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/68695f3e-75c7-4752-a896-dbec939c0dd6-test-operator-clouds-config\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"68695f3e-75c7-4752-a896-dbec939c0dd6\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Nov 22 10:08:37 crc kubenswrapper[4789]: I1122 10:08:37.236959 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"68695f3e-75c7-4752-a896-dbec939c0dd6\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Nov 22 10:08:37 crc kubenswrapper[4789]: I1122 10:08:37.237007 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/68695f3e-75c7-4752-a896-dbec939c0dd6-ca-certs\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"68695f3e-75c7-4752-a896-dbec939c0dd6\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Nov 22 10:08:37 crc kubenswrapper[4789]: I1122 10:08:37.237050 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/68695f3e-75c7-4752-a896-dbec939c0dd6-test-operator-ephemeral-temporary\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"68695f3e-75c7-4752-a896-dbec939c0dd6\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Nov 22 10:08:37 crc kubenswrapper[4789]: I1122 10:08:37.237084 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tobiko-config\" (UniqueName: \"kubernetes.io/configmap/68695f3e-75c7-4752-a896-dbec939c0dd6-tobiko-config\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"68695f3e-75c7-4752-a896-dbec939c0dd6\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Nov 22 10:08:37 crc kubenswrapper[4789]: I1122 10:08:37.237146 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tobiko-public-key\" (UniqueName: \"kubernetes.io/configmap/68695f3e-75c7-4752-a896-dbec939c0dd6-tobiko-public-key\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"68695f3e-75c7-4752-a896-dbec939c0dd6\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Nov 22 10:08:37 crc kubenswrapper[4789]: I1122 10:08:37.237176 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/68695f3e-75c7-4752-a896-dbec939c0dd6-ceph\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"68695f3e-75c7-4752-a896-dbec939c0dd6\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Nov 22 10:08:37 crc kubenswrapper[4789]: I1122 10:08:37.237225 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/68695f3e-75c7-4752-a896-dbec939c0dd6-openstack-config-secret\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"68695f3e-75c7-4752-a896-dbec939c0dd6\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Nov 22 10:08:37 crc kubenswrapper[4789]: I1122 10:08:37.237268 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/68695f3e-75c7-4752-a896-dbec939c0dd6-test-operator-ephemeral-workdir\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"68695f3e-75c7-4752-a896-dbec939c0dd6\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Nov 22 10:08:37 crc kubenswrapper[4789]: I1122 10:08:37.237316 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d6bgv\" (UniqueName: \"kubernetes.io/projected/68695f3e-75c7-4752-a896-dbec939c0dd6-kube-api-access-d6bgv\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"68695f3e-75c7-4752-a896-dbec939c0dd6\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Nov 22 10:08:37 crc kubenswrapper[4789]: I1122 10:08:37.237371 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tobiko-private-key\" (UniqueName: \"kubernetes.io/configmap/68695f3e-75c7-4752-a896-dbec939c0dd6-tobiko-private-key\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"68695f3e-75c7-4752-a896-dbec939c0dd6\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Nov 22 10:08:37 crc kubenswrapper[4789]: I1122 10:08:37.237989 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/68695f3e-75c7-4752-a896-dbec939c0dd6-test-operator-ephemeral-workdir\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"68695f3e-75c7-4752-a896-dbec939c0dd6\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Nov 22 10:08:37 crc kubenswrapper[4789]: I1122 10:08:37.238226 4789 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"68695f3e-75c7-4752-a896-dbec939c0dd6\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Nov 22 10:08:37 crc kubenswrapper[4789]: I1122 10:08:37.238259 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/68695f3e-75c7-4752-a896-dbec939c0dd6-test-operator-ephemeral-temporary\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"68695f3e-75c7-4752-a896-dbec939c0dd6\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Nov 22 10:08:37 crc kubenswrapper[4789]: I1122 10:08:37.238422 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tobiko-config\" (UniqueName: \"kubernetes.io/configmap/68695f3e-75c7-4752-a896-dbec939c0dd6-tobiko-config\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"68695f3e-75c7-4752-a896-dbec939c0dd6\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Nov 22 10:08:37 crc kubenswrapper[4789]: I1122 10:08:37.238529 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tobiko-public-key\" (UniqueName: \"kubernetes.io/configmap/68695f3e-75c7-4752-a896-dbec939c0dd6-tobiko-public-key\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"68695f3e-75c7-4752-a896-dbec939c0dd6\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Nov 22 10:08:37 crc kubenswrapper[4789]: I1122 10:08:37.238552 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/68695f3e-75c7-4752-a896-dbec939c0dd6-test-operator-clouds-config\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"68695f3e-75c7-4752-a896-dbec939c0dd6\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Nov 22 10:08:37 crc kubenswrapper[4789]: I1122 10:08:37.241067 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tobiko-private-key\" (UniqueName: \"kubernetes.io/configmap/68695f3e-75c7-4752-a896-dbec939c0dd6-tobiko-private-key\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"68695f3e-75c7-4752-a896-dbec939c0dd6\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Nov 22 10:08:37 crc kubenswrapper[4789]: I1122 10:08:37.242912 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/secret/68695f3e-75c7-4752-a896-dbec939c0dd6-kubeconfig\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"68695f3e-75c7-4752-a896-dbec939c0dd6\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Nov 22 10:08:37 crc kubenswrapper[4789]: I1122 10:08:37.243144 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/68695f3e-75c7-4752-a896-dbec939c0dd6-ceph\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"68695f3e-75c7-4752-a896-dbec939c0dd6\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Nov 22 10:08:37 crc kubenswrapper[4789]: I1122 10:08:37.245407 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/68695f3e-75c7-4752-a896-dbec939c0dd6-openstack-config-secret\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"68695f3e-75c7-4752-a896-dbec939c0dd6\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Nov 22 10:08:37 crc kubenswrapper[4789]: I1122 10:08:37.252795 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/68695f3e-75c7-4752-a896-dbec939c0dd6-ca-certs\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"68695f3e-75c7-4752-a896-dbec939c0dd6\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Nov 22 10:08:37 crc kubenswrapper[4789]: I1122 10:08:37.256492 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d6bgv\" (UniqueName: \"kubernetes.io/projected/68695f3e-75c7-4752-a896-dbec939c0dd6-kube-api-access-d6bgv\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"68695f3e-75c7-4752-a896-dbec939c0dd6\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Nov 22 10:08:37 crc kubenswrapper[4789]: I1122 10:08:37.270380 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"68695f3e-75c7-4752-a896-dbec939c0dd6\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Nov 22 10:08:37 crc kubenswrapper[4789]: I1122 10:08:37.368320 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Nov 22 10:08:37 crc kubenswrapper[4789]: I1122 10:08:37.931141 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tobiko-tests-tobiko-s00-podified-functional"] Nov 22 10:08:37 crc kubenswrapper[4789]: W1122 10:08:37.935078 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod68695f3e_75c7_4752_a896_dbec939c0dd6.slice/crio-05a9180e9cacf224877958b0949b33aadd330faf8745f7d1ec035c5d77b5d9db WatchSource:0}: Error finding container 05a9180e9cacf224877958b0949b33aadd330faf8745f7d1ec035c5d77b5d9db: Status 404 returned error can't find the container with id 05a9180e9cacf224877958b0949b33aadd330faf8745f7d1ec035c5d77b5d9db Nov 22 10:08:38 crc kubenswrapper[4789]: I1122 10:08:38.777462 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tobiko-tests-tobiko-s00-podified-functional" event={"ID":"68695f3e-75c7-4752-a896-dbec939c0dd6","Type":"ContainerStarted","Data":"05a9180e9cacf224877958b0949b33aadd330faf8745f7d1ec035c5d77b5d9db"} Nov 22 10:08:44 crc kubenswrapper[4789]: I1122 10:08:44.964813 4789 scope.go:117] "RemoveContainer" containerID="6e22a5a8fc3638ca6aab7b67ea550d9b289803b56baeb39c04dfbc919861ab85" Nov 22 10:08:44 crc kubenswrapper[4789]: E1122 10:08:44.965584 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 10:08:54 crc kubenswrapper[4789]: E1122 10:08:54.495994 4789 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tobiko:current-podified" Nov 22 10:08:54 crc kubenswrapper[4789]: E1122 10:08:54.496858 4789 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tobiko-tests-tobiko,Image:quay.io/podified-antelope-centos9/openstack-tobiko:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:TOBIKO_DEBUG_MODE,Value:false,ValueFrom:nil,},EnvVar{Name:TOBIKO_KEYS_FOLDER,Value:/etc/test_operator,ValueFrom:nil,},EnvVar{Name:TOBIKO_LOGS_DIR_NAME,Value:tobiko-tests-tobiko-s00-podified-functional,ValueFrom:nil,},EnvVar{Name:TOBIKO_PYTEST_ADDOPTS,Value:,ValueFrom:nil,},EnvVar{Name:TOBIKO_TESTENV,Value:functional -- tobiko/tests/functional/podified/test_topology.py,ValueFrom:nil,},EnvVar{Name:TOBIKO_VERSION,Value:master,ValueFrom:nil,},EnvVar{Name:TOX_NUM_PROCESSES,Value:2,ValueFrom:nil,},EnvVar{Name:USE_EXTERNAL_FILES,Value:True,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{8 0} {} 8 DecimalSI},memory: {{8589934592 0} {} BinarySI},},Requests:ResourceList{cpu: {{4 0} {} 4 DecimalSI},memory: {{4294967296 0} {} 4Gi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tobiko,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tobiko/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-clouds-config,ReadOnly:true,MountPath:/var/lib/tobiko/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-clouds-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:tobiko-config,ReadOnly:false,MountPath:/etc/tobiko/tobiko.conf,SubPath:tobiko.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ca-bundle.trust.crt,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:tobiko-private-key,ReadOnly:true,MountPath:/etc/test_operator/id_ecdsa,SubPath:id_ecdsa,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:tobiko-public-key,ReadOnly:true,MountPath:/etc/test_operator/id_ecdsa.pub,SubPath:id_ecdsa.pub,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kubeconfig,ReadOnly:true,MountPath:/var/lib/tobiko/.kube/config,SubPath:config,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ceph,ReadOnly:true,MountPath:/etc/ceph,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-d6bgv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[NET_ADMIN NET_RAW],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42495,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42495,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tobiko-tests-tobiko-s00-podified-functional_openstack(68695f3e-75c7-4752-a896-dbec939c0dd6): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 22 10:08:54 crc kubenswrapper[4789]: E1122 10:08:54.498067 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tobiko-tests-tobiko\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tobiko-tests-tobiko-s00-podified-functional" podUID="68695f3e-75c7-4752-a896-dbec939c0dd6" Nov 22 10:08:54 crc kubenswrapper[4789]: E1122 10:08:54.971049 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tobiko-tests-tobiko\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tobiko:current-podified\\\"\"" pod="openstack/tobiko-tests-tobiko-s00-podified-functional" podUID="68695f3e-75c7-4752-a896-dbec939c0dd6" Nov 22 10:08:55 crc kubenswrapper[4789]: I1122 10:08:55.965633 4789 scope.go:117] "RemoveContainer" containerID="6e22a5a8fc3638ca6aab7b67ea550d9b289803b56baeb39c04dfbc919861ab85" Nov 22 10:08:55 crc kubenswrapper[4789]: E1122 10:08:55.966233 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 10:09:06 crc kubenswrapper[4789]: I1122 10:09:06.966031 4789 scope.go:117] "RemoveContainer" containerID="6e22a5a8fc3638ca6aab7b67ea550d9b289803b56baeb39c04dfbc919861ab85" Nov 22 10:09:06 crc kubenswrapper[4789]: E1122 10:09:06.967638 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 10:09:11 crc kubenswrapper[4789]: I1122 10:09:11.665884 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tobiko-tests-tobiko-s00-podified-functional" event={"ID":"68695f3e-75c7-4752-a896-dbec939c0dd6","Type":"ContainerStarted","Data":"1373fe435707924df3a278570d8d4efe3a42a58b949b44a3b4e462ab3708ba23"} Nov 22 10:09:11 crc kubenswrapper[4789]: I1122 10:09:11.692415 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tobiko-tests-tobiko-s00-podified-functional" podStartSLOduration=3.748216667 podStartE2EDuration="36.692396921s" podCreationTimestamp="2025-11-22 10:08:35 +0000 UTC" firstStartedPulling="2025-11-22 10:08:37.937283122 +0000 UTC m=+7932.171683395" lastFinishedPulling="2025-11-22 10:09:10.881463366 +0000 UTC m=+7965.115863649" observedRunningTime="2025-11-22 10:09:11.685232379 +0000 UTC m=+7965.919632662" watchObservedRunningTime="2025-11-22 10:09:11.692396921 +0000 UTC m=+7965.926797194" Nov 22 10:09:18 crc kubenswrapper[4789]: I1122 10:09:18.965690 4789 scope.go:117] "RemoveContainer" containerID="6e22a5a8fc3638ca6aab7b67ea550d9b289803b56baeb39c04dfbc919861ab85" Nov 22 10:09:18 crc kubenswrapper[4789]: E1122 10:09:18.967159 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 10:09:30 crc kubenswrapper[4789]: I1122 10:09:30.964540 4789 scope.go:117] "RemoveContainer" containerID="6e22a5a8fc3638ca6aab7b67ea550d9b289803b56baeb39c04dfbc919861ab85" Nov 22 10:09:30 crc kubenswrapper[4789]: E1122 10:09:30.965279 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 10:09:44 crc kubenswrapper[4789]: I1122 10:09:44.966467 4789 scope.go:117] "RemoveContainer" containerID="6e22a5a8fc3638ca6aab7b67ea550d9b289803b56baeb39c04dfbc919861ab85" Nov 22 10:09:44 crc kubenswrapper[4789]: E1122 10:09:44.967537 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 10:09:56 crc kubenswrapper[4789]: I1122 10:09:56.965706 4789 scope.go:117] "RemoveContainer" containerID="6e22a5a8fc3638ca6aab7b67ea550d9b289803b56baeb39c04dfbc919861ab85" Nov 22 10:09:56 crc kubenswrapper[4789]: E1122 10:09:56.966633 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 10:10:11 crc kubenswrapper[4789]: I1122 10:10:11.965788 4789 scope.go:117] "RemoveContainer" containerID="6e22a5a8fc3638ca6aab7b67ea550d9b289803b56baeb39c04dfbc919861ab85" Nov 22 10:10:11 crc kubenswrapper[4789]: E1122 10:10:11.966772 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 10:10:24 crc kubenswrapper[4789]: I1122 10:10:24.965866 4789 scope.go:117] "RemoveContainer" containerID="6e22a5a8fc3638ca6aab7b67ea550d9b289803b56baeb39c04dfbc919861ab85" Nov 22 10:10:24 crc kubenswrapper[4789]: E1122 10:10:24.966720 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 10:10:26 crc kubenswrapper[4789]: I1122 10:10:26.409257 4789 generic.go:334] "Generic (PLEG): container finished" podID="68695f3e-75c7-4752-a896-dbec939c0dd6" containerID="1373fe435707924df3a278570d8d4efe3a42a58b949b44a3b4e462ab3708ba23" exitCode=0 Nov 22 10:10:26 crc kubenswrapper[4789]: I1122 10:10:26.409354 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tobiko-tests-tobiko-s00-podified-functional" event={"ID":"68695f3e-75c7-4752-a896-dbec939c0dd6","Type":"ContainerDied","Data":"1373fe435707924df3a278570d8d4efe3a42a58b949b44a3b4e462ab3708ba23"} Nov 22 10:10:27 crc kubenswrapper[4789]: I1122 10:10:27.908844 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Nov 22 10:10:27 crc kubenswrapper[4789]: I1122 10:10:27.987877 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tobiko-tests-tobiko-s01-sanity"] Nov 22 10:10:27 crc kubenswrapper[4789]: E1122 10:10:27.988271 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68695f3e-75c7-4752-a896-dbec939c0dd6" containerName="tobiko-tests-tobiko" Nov 22 10:10:27 crc kubenswrapper[4789]: I1122 10:10:27.988284 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="68695f3e-75c7-4752-a896-dbec939c0dd6" containerName="tobiko-tests-tobiko" Nov 22 10:10:27 crc kubenswrapper[4789]: I1122 10:10:27.988501 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="68695f3e-75c7-4752-a896-dbec939c0dd6" containerName="tobiko-tests-tobiko" Nov 22 10:10:27 crc kubenswrapper[4789]: I1122 10:10:27.989355 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tobiko-tests-tobiko-s01-sanity" Nov 22 10:10:27 crc kubenswrapper[4789]: I1122 10:10:27.991666 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tobiko-tests-tobiko-s01-sanity"] Nov 22 10:10:28 crc kubenswrapper[4789]: I1122 10:10:28.006065 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tobiko-private-key\" (UniqueName: \"kubernetes.io/configmap/68695f3e-75c7-4752-a896-dbec939c0dd6-tobiko-private-key\") pod \"68695f3e-75c7-4752-a896-dbec939c0dd6\" (UID: \"68695f3e-75c7-4752-a896-dbec939c0dd6\") " Nov 22 10:10:28 crc kubenswrapper[4789]: I1122 10:10:28.006152 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/68695f3e-75c7-4752-a896-dbec939c0dd6-ca-certs\") pod \"68695f3e-75c7-4752-a896-dbec939c0dd6\" (UID: \"68695f3e-75c7-4752-a896-dbec939c0dd6\") " Nov 22 10:10:28 crc kubenswrapper[4789]: I1122 10:10:28.006185 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/68695f3e-75c7-4752-a896-dbec939c0dd6-test-operator-clouds-config\") pod \"68695f3e-75c7-4752-a896-dbec939c0dd6\" (UID: \"68695f3e-75c7-4752-a896-dbec939c0dd6\") " Nov 22 10:10:28 crc kubenswrapper[4789]: I1122 10:10:28.006221 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/68695f3e-75c7-4752-a896-dbec939c0dd6-ceph\") pod \"68695f3e-75c7-4752-a896-dbec939c0dd6\" (UID: \"68695f3e-75c7-4752-a896-dbec939c0dd6\") " Nov 22 10:10:28 crc kubenswrapper[4789]: I1122 10:10:28.006304 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tobiko-config\" (UniqueName: \"kubernetes.io/configmap/68695f3e-75c7-4752-a896-dbec939c0dd6-tobiko-config\") pod \"68695f3e-75c7-4752-a896-dbec939c0dd6\" (UID: \"68695f3e-75c7-4752-a896-dbec939c0dd6\") " Nov 22 10:10:28 crc kubenswrapper[4789]: I1122 10:10:28.006336 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"68695f3e-75c7-4752-a896-dbec939c0dd6\" (UID: \"68695f3e-75c7-4752-a896-dbec939c0dd6\") " Nov 22 10:10:28 crc kubenswrapper[4789]: I1122 10:10:28.006416 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/68695f3e-75c7-4752-a896-dbec939c0dd6-test-operator-ephemeral-temporary\") pod \"68695f3e-75c7-4752-a896-dbec939c0dd6\" (UID: \"68695f3e-75c7-4752-a896-dbec939c0dd6\") " Nov 22 10:10:28 crc kubenswrapper[4789]: I1122 10:10:28.006452 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/68695f3e-75c7-4752-a896-dbec939c0dd6-test-operator-ephemeral-workdir\") pod \"68695f3e-75c7-4752-a896-dbec939c0dd6\" (UID: \"68695f3e-75c7-4752-a896-dbec939c0dd6\") " Nov 22 10:10:28 crc kubenswrapper[4789]: I1122 10:10:28.006510 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/secret/68695f3e-75c7-4752-a896-dbec939c0dd6-kubeconfig\") pod \"68695f3e-75c7-4752-a896-dbec939c0dd6\" (UID: \"68695f3e-75c7-4752-a896-dbec939c0dd6\") " Nov 22 10:10:28 crc kubenswrapper[4789]: I1122 10:10:28.006580 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/68695f3e-75c7-4752-a896-dbec939c0dd6-openstack-config-secret\") pod \"68695f3e-75c7-4752-a896-dbec939c0dd6\" (UID: \"68695f3e-75c7-4752-a896-dbec939c0dd6\") " Nov 22 10:10:28 crc kubenswrapper[4789]: I1122 10:10:28.006615 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6bgv\" (UniqueName: \"kubernetes.io/projected/68695f3e-75c7-4752-a896-dbec939c0dd6-kube-api-access-d6bgv\") pod \"68695f3e-75c7-4752-a896-dbec939c0dd6\" (UID: \"68695f3e-75c7-4752-a896-dbec939c0dd6\") " Nov 22 10:10:28 crc kubenswrapper[4789]: I1122 10:10:28.006677 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tobiko-public-key\" (UniqueName: \"kubernetes.io/configmap/68695f3e-75c7-4752-a896-dbec939c0dd6-tobiko-public-key\") pod \"68695f3e-75c7-4752-a896-dbec939c0dd6\" (UID: \"68695f3e-75c7-4752-a896-dbec939c0dd6\") " Nov 22 10:10:28 crc kubenswrapper[4789]: I1122 10:10:28.012494 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/68695f3e-75c7-4752-a896-dbec939c0dd6-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "68695f3e-75c7-4752-a896-dbec939c0dd6" (UID: "68695f3e-75c7-4752-a896-dbec939c0dd6"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 10:10:28 crc kubenswrapper[4789]: I1122 10:10:28.016978 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68695f3e-75c7-4752-a896-dbec939c0dd6-kube-api-access-d6bgv" (OuterVolumeSpecName: "kube-api-access-d6bgv") pod "68695f3e-75c7-4752-a896-dbec939c0dd6" (UID: "68695f3e-75c7-4752-a896-dbec939c0dd6"). InnerVolumeSpecName "kube-api-access-d6bgv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 10:10:28 crc kubenswrapper[4789]: I1122 10:10:28.030687 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "test-operator-logs") pod "68695f3e-75c7-4752-a896-dbec939c0dd6" (UID: "68695f3e-75c7-4752-a896-dbec939c0dd6"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 22 10:10:28 crc kubenswrapper[4789]: I1122 10:10:28.032643 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68695f3e-75c7-4752-a896-dbec939c0dd6-ceph" (OuterVolumeSpecName: "ceph") pod "68695f3e-75c7-4752-a896-dbec939c0dd6" (UID: "68695f3e-75c7-4752-a896-dbec939c0dd6"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 10:10:28 crc kubenswrapper[4789]: I1122 10:10:28.035792 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/68695f3e-75c7-4752-a896-dbec939c0dd6-tobiko-config" (OuterVolumeSpecName: "tobiko-config") pod "68695f3e-75c7-4752-a896-dbec939c0dd6" (UID: "68695f3e-75c7-4752-a896-dbec939c0dd6"). InnerVolumeSpecName "tobiko-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 10:10:28 crc kubenswrapper[4789]: I1122 10:10:28.040676 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68695f3e-75c7-4752-a896-dbec939c0dd6-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "68695f3e-75c7-4752-a896-dbec939c0dd6" (UID: "68695f3e-75c7-4752-a896-dbec939c0dd6"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 10:10:28 crc kubenswrapper[4789]: I1122 10:10:28.062536 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/68695f3e-75c7-4752-a896-dbec939c0dd6-tobiko-public-key" (OuterVolumeSpecName: "tobiko-public-key") pod "68695f3e-75c7-4752-a896-dbec939c0dd6" (UID: "68695f3e-75c7-4752-a896-dbec939c0dd6"). InnerVolumeSpecName "tobiko-public-key". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 10:10:28 crc kubenswrapper[4789]: I1122 10:10:28.064698 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68695f3e-75c7-4752-a896-dbec939c0dd6-kubeconfig" (OuterVolumeSpecName: "kubeconfig") pod "68695f3e-75c7-4752-a896-dbec939c0dd6" (UID: "68695f3e-75c7-4752-a896-dbec939c0dd6"). InnerVolumeSpecName "kubeconfig". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 10:10:28 crc kubenswrapper[4789]: I1122 10:10:28.069627 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/68695f3e-75c7-4752-a896-dbec939c0dd6-tobiko-private-key" (OuterVolumeSpecName: "tobiko-private-key") pod "68695f3e-75c7-4752-a896-dbec939c0dd6" (UID: "68695f3e-75c7-4752-a896-dbec939c0dd6"). InnerVolumeSpecName "tobiko-private-key". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 10:10:28 crc kubenswrapper[4789]: I1122 10:10:28.086554 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/68695f3e-75c7-4752-a896-dbec939c0dd6-test-operator-clouds-config" (OuterVolumeSpecName: "test-operator-clouds-config") pod "68695f3e-75c7-4752-a896-dbec939c0dd6" (UID: "68695f3e-75c7-4752-a896-dbec939c0dd6"). InnerVolumeSpecName "test-operator-clouds-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 10:10:28 crc kubenswrapper[4789]: I1122 10:10:28.095033 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68695f3e-75c7-4752-a896-dbec939c0dd6-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "68695f3e-75c7-4752-a896-dbec939c0dd6" (UID: "68695f3e-75c7-4752-a896-dbec939c0dd6"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 10:10:28 crc kubenswrapper[4789]: I1122 10:10:28.110135 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/d3fd6f1c-a6a5-406f-9eae-b6627046b275-test-operator-ephemeral-temporary\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"d3fd6f1c-a6a5-406f-9eae-b6627046b275\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Nov 22 10:10:28 crc kubenswrapper[4789]: I1122 10:10:28.110225 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"d3fd6f1c-a6a5-406f-9eae-b6627046b275\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Nov 22 10:10:28 crc kubenswrapper[4789]: I1122 10:10:28.110273 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/d3fd6f1c-a6a5-406f-9eae-b6627046b275-test-operator-ephemeral-workdir\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"d3fd6f1c-a6a5-406f-9eae-b6627046b275\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Nov 22 10:10:28 crc kubenswrapper[4789]: I1122 10:10:28.110329 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tobiko-config\" (UniqueName: \"kubernetes.io/configmap/d3fd6f1c-a6a5-406f-9eae-b6627046b275-tobiko-config\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"d3fd6f1c-a6a5-406f-9eae-b6627046b275\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Nov 22 10:10:28 crc kubenswrapper[4789]: I1122 10:10:28.110382 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/secret/d3fd6f1c-a6a5-406f-9eae-b6627046b275-kubeconfig\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"d3fd6f1c-a6a5-406f-9eae-b6627046b275\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Nov 22 10:10:28 crc kubenswrapper[4789]: I1122 10:10:28.110418 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d3fd6f1c-a6a5-406f-9eae-b6627046b275-ceph\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"d3fd6f1c-a6a5-406f-9eae-b6627046b275\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Nov 22 10:10:28 crc kubenswrapper[4789]: I1122 10:10:28.110616 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4psw4\" (UniqueName: \"kubernetes.io/projected/d3fd6f1c-a6a5-406f-9eae-b6627046b275-kube-api-access-4psw4\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"d3fd6f1c-a6a5-406f-9eae-b6627046b275\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Nov 22 10:10:28 crc kubenswrapper[4789]: I1122 10:10:28.110820 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tobiko-private-key\" (UniqueName: \"kubernetes.io/configmap/d3fd6f1c-a6a5-406f-9eae-b6627046b275-tobiko-private-key\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"d3fd6f1c-a6a5-406f-9eae-b6627046b275\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Nov 22 10:10:28 crc kubenswrapper[4789]: I1122 10:10:28.113074 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/d3fd6f1c-a6a5-406f-9eae-b6627046b275-test-operator-clouds-config\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"d3fd6f1c-a6a5-406f-9eae-b6627046b275\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Nov 22 10:10:28 crc kubenswrapper[4789]: I1122 10:10:28.113136 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/d3fd6f1c-a6a5-406f-9eae-b6627046b275-ca-certs\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"d3fd6f1c-a6a5-406f-9eae-b6627046b275\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Nov 22 10:10:28 crc kubenswrapper[4789]: I1122 10:10:28.113174 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/d3fd6f1c-a6a5-406f-9eae-b6627046b275-openstack-config-secret\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"d3fd6f1c-a6a5-406f-9eae-b6627046b275\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Nov 22 10:10:28 crc kubenswrapper[4789]: I1122 10:10:28.113271 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tobiko-public-key\" (UniqueName: \"kubernetes.io/configmap/d3fd6f1c-a6a5-406f-9eae-b6627046b275-tobiko-public-key\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"d3fd6f1c-a6a5-406f-9eae-b6627046b275\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Nov 22 10:10:28 crc kubenswrapper[4789]: I1122 10:10:28.113387 4789 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/68695f3e-75c7-4752-a896-dbec939c0dd6-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Nov 22 10:10:28 crc kubenswrapper[4789]: I1122 10:10:28.113404 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6bgv\" (UniqueName: \"kubernetes.io/projected/68695f3e-75c7-4752-a896-dbec939c0dd6-kube-api-access-d6bgv\") on node \"crc\" DevicePath \"\"" Nov 22 10:10:28 crc kubenswrapper[4789]: I1122 10:10:28.113422 4789 reconciler_common.go:293] "Volume detached for volume \"tobiko-public-key\" (UniqueName: \"kubernetes.io/configmap/68695f3e-75c7-4752-a896-dbec939c0dd6-tobiko-public-key\") on node \"crc\" DevicePath \"\"" Nov 22 10:10:28 crc kubenswrapper[4789]: I1122 10:10:28.113433 4789 reconciler_common.go:293] "Volume detached for volume \"tobiko-private-key\" (UniqueName: \"kubernetes.io/configmap/68695f3e-75c7-4752-a896-dbec939c0dd6-tobiko-private-key\") on node \"crc\" DevicePath \"\"" Nov 22 10:10:28 crc kubenswrapper[4789]: I1122 10:10:28.113443 4789 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/68695f3e-75c7-4752-a896-dbec939c0dd6-ca-certs\") on node \"crc\" DevicePath \"\"" Nov 22 10:10:28 crc kubenswrapper[4789]: I1122 10:10:28.113456 4789 reconciler_common.go:293] "Volume detached for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/68695f3e-75c7-4752-a896-dbec939c0dd6-test-operator-clouds-config\") on node \"crc\" DevicePath \"\"" Nov 22 10:10:28 crc kubenswrapper[4789]: I1122 10:10:28.113465 4789 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/68695f3e-75c7-4752-a896-dbec939c0dd6-ceph\") on node \"crc\" DevicePath \"\"" Nov 22 10:10:28 crc kubenswrapper[4789]: I1122 10:10:28.113475 4789 reconciler_common.go:293] "Volume detached for volume \"tobiko-config\" (UniqueName: \"kubernetes.io/configmap/68695f3e-75c7-4752-a896-dbec939c0dd6-tobiko-config\") on node \"crc\" DevicePath \"\"" Nov 22 10:10:28 crc kubenswrapper[4789]: I1122 10:10:28.113488 4789 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/68695f3e-75c7-4752-a896-dbec939c0dd6-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Nov 22 10:10:28 crc kubenswrapper[4789]: I1122 10:10:28.113497 4789 reconciler_common.go:293] "Volume detached for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/secret/68695f3e-75c7-4752-a896-dbec939c0dd6-kubeconfig\") on node \"crc\" DevicePath \"\"" Nov 22 10:10:28 crc kubenswrapper[4789]: I1122 10:10:28.150851 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"d3fd6f1c-a6a5-406f-9eae-b6627046b275\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Nov 22 10:10:28 crc kubenswrapper[4789]: I1122 10:10:28.216628 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tobiko-public-key\" (UniqueName: \"kubernetes.io/configmap/d3fd6f1c-a6a5-406f-9eae-b6627046b275-tobiko-public-key\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"d3fd6f1c-a6a5-406f-9eae-b6627046b275\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Nov 22 10:10:28 crc kubenswrapper[4789]: I1122 10:10:28.216693 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/d3fd6f1c-a6a5-406f-9eae-b6627046b275-test-operator-ephemeral-temporary\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"d3fd6f1c-a6a5-406f-9eae-b6627046b275\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Nov 22 10:10:28 crc kubenswrapper[4789]: I1122 10:10:28.216777 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/d3fd6f1c-a6a5-406f-9eae-b6627046b275-test-operator-ephemeral-workdir\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"d3fd6f1c-a6a5-406f-9eae-b6627046b275\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Nov 22 10:10:28 crc kubenswrapper[4789]: I1122 10:10:28.216861 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tobiko-config\" (UniqueName: \"kubernetes.io/configmap/d3fd6f1c-a6a5-406f-9eae-b6627046b275-tobiko-config\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"d3fd6f1c-a6a5-406f-9eae-b6627046b275\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Nov 22 10:10:28 crc kubenswrapper[4789]: I1122 10:10:28.216978 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/secret/d3fd6f1c-a6a5-406f-9eae-b6627046b275-kubeconfig\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"d3fd6f1c-a6a5-406f-9eae-b6627046b275\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Nov 22 10:10:28 crc kubenswrapper[4789]: I1122 10:10:28.217037 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d3fd6f1c-a6a5-406f-9eae-b6627046b275-ceph\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"d3fd6f1c-a6a5-406f-9eae-b6627046b275\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Nov 22 10:10:28 crc kubenswrapper[4789]: I1122 10:10:28.217085 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4psw4\" (UniqueName: \"kubernetes.io/projected/d3fd6f1c-a6a5-406f-9eae-b6627046b275-kube-api-access-4psw4\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"d3fd6f1c-a6a5-406f-9eae-b6627046b275\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Nov 22 10:10:28 crc kubenswrapper[4789]: I1122 10:10:28.217138 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tobiko-private-key\" (UniqueName: \"kubernetes.io/configmap/d3fd6f1c-a6a5-406f-9eae-b6627046b275-tobiko-private-key\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"d3fd6f1c-a6a5-406f-9eae-b6627046b275\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Nov 22 10:10:28 crc kubenswrapper[4789]: I1122 10:10:28.217213 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/d3fd6f1c-a6a5-406f-9eae-b6627046b275-test-operator-clouds-config\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"d3fd6f1c-a6a5-406f-9eae-b6627046b275\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Nov 22 10:10:28 crc kubenswrapper[4789]: I1122 10:10:28.217248 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/d3fd6f1c-a6a5-406f-9eae-b6627046b275-ca-certs\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"d3fd6f1c-a6a5-406f-9eae-b6627046b275\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Nov 22 10:10:28 crc kubenswrapper[4789]: I1122 10:10:28.217273 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/d3fd6f1c-a6a5-406f-9eae-b6627046b275-openstack-config-secret\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"d3fd6f1c-a6a5-406f-9eae-b6627046b275\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Nov 22 10:10:28 crc kubenswrapper[4789]: I1122 10:10:28.217410 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/d3fd6f1c-a6a5-406f-9eae-b6627046b275-test-operator-ephemeral-temporary\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"d3fd6f1c-a6a5-406f-9eae-b6627046b275\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Nov 22 10:10:28 crc kubenswrapper[4789]: I1122 10:10:28.217486 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/d3fd6f1c-a6a5-406f-9eae-b6627046b275-test-operator-ephemeral-workdir\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"d3fd6f1c-a6a5-406f-9eae-b6627046b275\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Nov 22 10:10:28 crc kubenswrapper[4789]: I1122 10:10:28.217530 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tobiko-public-key\" (UniqueName: \"kubernetes.io/configmap/d3fd6f1c-a6a5-406f-9eae-b6627046b275-tobiko-public-key\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"d3fd6f1c-a6a5-406f-9eae-b6627046b275\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Nov 22 10:10:28 crc kubenswrapper[4789]: I1122 10:10:28.218586 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tobiko-config\" (UniqueName: \"kubernetes.io/configmap/d3fd6f1c-a6a5-406f-9eae-b6627046b275-tobiko-config\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"d3fd6f1c-a6a5-406f-9eae-b6627046b275\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Nov 22 10:10:28 crc kubenswrapper[4789]: I1122 10:10:28.219290 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tobiko-private-key\" (UniqueName: \"kubernetes.io/configmap/d3fd6f1c-a6a5-406f-9eae-b6627046b275-tobiko-private-key\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"d3fd6f1c-a6a5-406f-9eae-b6627046b275\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Nov 22 10:10:28 crc kubenswrapper[4789]: I1122 10:10:28.219633 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/d3fd6f1c-a6a5-406f-9eae-b6627046b275-test-operator-clouds-config\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"d3fd6f1c-a6a5-406f-9eae-b6627046b275\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Nov 22 10:10:28 crc kubenswrapper[4789]: I1122 10:10:28.221334 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/secret/d3fd6f1c-a6a5-406f-9eae-b6627046b275-kubeconfig\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"d3fd6f1c-a6a5-406f-9eae-b6627046b275\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Nov 22 10:10:28 crc kubenswrapper[4789]: I1122 10:10:28.221344 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/d3fd6f1c-a6a5-406f-9eae-b6627046b275-openstack-config-secret\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"d3fd6f1c-a6a5-406f-9eae-b6627046b275\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Nov 22 10:10:28 crc kubenswrapper[4789]: I1122 10:10:28.221346 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d3fd6f1c-a6a5-406f-9eae-b6627046b275-ceph\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"d3fd6f1c-a6a5-406f-9eae-b6627046b275\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Nov 22 10:10:28 crc kubenswrapper[4789]: I1122 10:10:28.223098 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/d3fd6f1c-a6a5-406f-9eae-b6627046b275-ca-certs\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"d3fd6f1c-a6a5-406f-9eae-b6627046b275\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Nov 22 10:10:28 crc kubenswrapper[4789]: I1122 10:10:28.233671 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4psw4\" (UniqueName: \"kubernetes.io/projected/d3fd6f1c-a6a5-406f-9eae-b6627046b275-kube-api-access-4psw4\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"d3fd6f1c-a6a5-406f-9eae-b6627046b275\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Nov 22 10:10:28 crc kubenswrapper[4789]: I1122 10:10:28.312426 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tobiko-tests-tobiko-s01-sanity" Nov 22 10:10:28 crc kubenswrapper[4789]: I1122 10:10:28.437256 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tobiko-tests-tobiko-s00-podified-functional" event={"ID":"68695f3e-75c7-4752-a896-dbec939c0dd6","Type":"ContainerDied","Data":"05a9180e9cacf224877958b0949b33aadd330faf8745f7d1ec035c5d77b5d9db"} Nov 22 10:10:28 crc kubenswrapper[4789]: I1122 10:10:28.437598 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="05a9180e9cacf224877958b0949b33aadd330faf8745f7d1ec035c5d77b5d9db" Nov 22 10:10:28 crc kubenswrapper[4789]: I1122 10:10:28.437312 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Nov 22 10:10:28 crc kubenswrapper[4789]: I1122 10:10:28.930999 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tobiko-tests-tobiko-s01-sanity"] Nov 22 10:10:29 crc kubenswrapper[4789]: I1122 10:10:29.304257 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/68695f3e-75c7-4752-a896-dbec939c0dd6-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "68695f3e-75c7-4752-a896-dbec939c0dd6" (UID: "68695f3e-75c7-4752-a896-dbec939c0dd6"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 10:10:29 crc kubenswrapper[4789]: I1122 10:10:29.350148 4789 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/68695f3e-75c7-4752-a896-dbec939c0dd6-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Nov 22 10:10:29 crc kubenswrapper[4789]: I1122 10:10:29.457789 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tobiko-tests-tobiko-s01-sanity" event={"ID":"d3fd6f1c-a6a5-406f-9eae-b6627046b275","Type":"ContainerStarted","Data":"bdfaab3ad8dd901dce072b82a8906d905eeb492408c396b76af3945635a2ce3a"} Nov 22 10:10:30 crc kubenswrapper[4789]: I1122 10:10:30.469126 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tobiko-tests-tobiko-s01-sanity" event={"ID":"d3fd6f1c-a6a5-406f-9eae-b6627046b275","Type":"ContainerStarted","Data":"dddf6b7c809531681c5d607f59bd710e2a6e43e92fb27fa786832ff55a5981d4"} Nov 22 10:10:30 crc kubenswrapper[4789]: I1122 10:10:30.498547 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tobiko-tests-tobiko-s01-sanity" podStartSLOduration=3.498521241 podStartE2EDuration="3.498521241s" podCreationTimestamp="2025-11-22 10:10:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 10:10:30.487138005 +0000 UTC m=+8044.721538328" watchObservedRunningTime="2025-11-22 10:10:30.498521241 +0000 UTC m=+8044.732921534" Nov 22 10:10:38 crc kubenswrapper[4789]: I1122 10:10:38.965481 4789 scope.go:117] "RemoveContainer" containerID="6e22a5a8fc3638ca6aab7b67ea550d9b289803b56baeb39c04dfbc919861ab85" Nov 22 10:10:38 crc kubenswrapper[4789]: E1122 10:10:38.966200 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 10:10:53 crc kubenswrapper[4789]: I1122 10:10:53.965832 4789 scope.go:117] "RemoveContainer" containerID="6e22a5a8fc3638ca6aab7b67ea550d9b289803b56baeb39c04dfbc919861ab85" Nov 22 10:10:53 crc kubenswrapper[4789]: E1122 10:10:53.966918 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 10:11:04 crc kubenswrapper[4789]: I1122 10:11:04.964781 4789 scope.go:117] "RemoveContainer" containerID="6e22a5a8fc3638ca6aab7b67ea550d9b289803b56baeb39c04dfbc919861ab85" Nov 22 10:11:04 crc kubenswrapper[4789]: E1122 10:11:04.966810 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 10:11:16 crc kubenswrapper[4789]: I1122 10:11:16.966993 4789 scope.go:117] "RemoveContainer" containerID="6e22a5a8fc3638ca6aab7b67ea550d9b289803b56baeb39c04dfbc919861ab85" Nov 22 10:11:17 crc kubenswrapper[4789]: E1122 10:11:16.968300 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 10:11:21 crc kubenswrapper[4789]: I1122 10:11:21.017821 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-p6bnq"] Nov 22 10:11:21 crc kubenswrapper[4789]: I1122 10:11:21.025392 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p6bnq" Nov 22 10:11:21 crc kubenswrapper[4789]: I1122 10:11:21.036227 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-p6bnq"] Nov 22 10:11:21 crc kubenswrapper[4789]: I1122 10:11:21.124901 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9hb7d\" (UniqueName: \"kubernetes.io/projected/136cff6d-0787-4921-a5f3-7c719483a110-kube-api-access-9hb7d\") pod \"redhat-operators-p6bnq\" (UID: \"136cff6d-0787-4921-a5f3-7c719483a110\") " pod="openshift-marketplace/redhat-operators-p6bnq" Nov 22 10:11:21 crc kubenswrapper[4789]: I1122 10:11:21.125292 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/136cff6d-0787-4921-a5f3-7c719483a110-catalog-content\") pod \"redhat-operators-p6bnq\" (UID: \"136cff6d-0787-4921-a5f3-7c719483a110\") " pod="openshift-marketplace/redhat-operators-p6bnq" Nov 22 10:11:21 crc kubenswrapper[4789]: I1122 10:11:21.125458 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/136cff6d-0787-4921-a5f3-7c719483a110-utilities\") pod \"redhat-operators-p6bnq\" (UID: \"136cff6d-0787-4921-a5f3-7c719483a110\") " pod="openshift-marketplace/redhat-operators-p6bnq" Nov 22 10:11:21 crc kubenswrapper[4789]: I1122 10:11:21.226953 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/136cff6d-0787-4921-a5f3-7c719483a110-utilities\") pod \"redhat-operators-p6bnq\" (UID: \"136cff6d-0787-4921-a5f3-7c719483a110\") " pod="openshift-marketplace/redhat-operators-p6bnq" Nov 22 10:11:21 crc kubenswrapper[4789]: I1122 10:11:21.227279 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/136cff6d-0787-4921-a5f3-7c719483a110-catalog-content\") pod \"redhat-operators-p6bnq\" (UID: \"136cff6d-0787-4921-a5f3-7c719483a110\") " pod="openshift-marketplace/redhat-operators-p6bnq" Nov 22 10:11:21 crc kubenswrapper[4789]: I1122 10:11:21.227416 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/136cff6d-0787-4921-a5f3-7c719483a110-utilities\") pod \"redhat-operators-p6bnq\" (UID: \"136cff6d-0787-4921-a5f3-7c719483a110\") " pod="openshift-marketplace/redhat-operators-p6bnq" Nov 22 10:11:21 crc kubenswrapper[4789]: I1122 10:11:21.227558 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9hb7d\" (UniqueName: \"kubernetes.io/projected/136cff6d-0787-4921-a5f3-7c719483a110-kube-api-access-9hb7d\") pod \"redhat-operators-p6bnq\" (UID: \"136cff6d-0787-4921-a5f3-7c719483a110\") " pod="openshift-marketplace/redhat-operators-p6bnq" Nov 22 10:11:21 crc kubenswrapper[4789]: I1122 10:11:21.227966 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/136cff6d-0787-4921-a5f3-7c719483a110-catalog-content\") pod \"redhat-operators-p6bnq\" (UID: \"136cff6d-0787-4921-a5f3-7c719483a110\") " pod="openshift-marketplace/redhat-operators-p6bnq" Nov 22 10:11:21 crc kubenswrapper[4789]: I1122 10:11:21.254297 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9hb7d\" (UniqueName: \"kubernetes.io/projected/136cff6d-0787-4921-a5f3-7c719483a110-kube-api-access-9hb7d\") pod \"redhat-operators-p6bnq\" (UID: \"136cff6d-0787-4921-a5f3-7c719483a110\") " pod="openshift-marketplace/redhat-operators-p6bnq" Nov 22 10:11:21 crc kubenswrapper[4789]: I1122 10:11:21.410546 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p6bnq" Nov 22 10:11:21 crc kubenswrapper[4789]: I1122 10:11:21.871333 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-p6bnq"] Nov 22 10:11:21 crc kubenswrapper[4789]: I1122 10:11:21.973375 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p6bnq" event={"ID":"136cff6d-0787-4921-a5f3-7c719483a110","Type":"ContainerStarted","Data":"527aedd0928dd480fefed02867a9b71273db65e5e2aa6901562b4851884ba340"} Nov 22 10:11:22 crc kubenswrapper[4789]: I1122 10:11:22.978613 4789 generic.go:334] "Generic (PLEG): container finished" podID="136cff6d-0787-4921-a5f3-7c719483a110" containerID="1bd58955a009ba1fffcca50d1d46b18159086db27eba53983f073293804af87f" exitCode=0 Nov 22 10:11:22 crc kubenswrapper[4789]: I1122 10:11:22.979107 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p6bnq" event={"ID":"136cff6d-0787-4921-a5f3-7c719483a110","Type":"ContainerDied","Data":"1bd58955a009ba1fffcca50d1d46b18159086db27eba53983f073293804af87f"} Nov 22 10:11:23 crc kubenswrapper[4789]: I1122 10:11:23.989368 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p6bnq" event={"ID":"136cff6d-0787-4921-a5f3-7c719483a110","Type":"ContainerStarted","Data":"525bd0e43b0003e77fd18701aa69e2ea1ca8c97ddb980fa320eb30ff29909929"} Nov 22 10:11:25 crc kubenswrapper[4789]: I1122 10:11:25.003033 4789 generic.go:334] "Generic (PLEG): container finished" podID="136cff6d-0787-4921-a5f3-7c719483a110" containerID="525bd0e43b0003e77fd18701aa69e2ea1ca8c97ddb980fa320eb30ff29909929" exitCode=0 Nov 22 10:11:25 crc kubenswrapper[4789]: I1122 10:11:25.003226 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p6bnq" event={"ID":"136cff6d-0787-4921-a5f3-7c719483a110","Type":"ContainerDied","Data":"525bd0e43b0003e77fd18701aa69e2ea1ca8c97ddb980fa320eb30ff29909929"} Nov 22 10:11:27 crc kubenswrapper[4789]: I1122 10:11:27.021574 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p6bnq" event={"ID":"136cff6d-0787-4921-a5f3-7c719483a110","Type":"ContainerStarted","Data":"4bb35c394c8244a7be16b7fe5b58d5238d82c8344904aab0f1804d11e6c8f23f"} Nov 22 10:11:27 crc kubenswrapper[4789]: I1122 10:11:27.043514 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-p6bnq" podStartSLOduration=3.8795085240000002 podStartE2EDuration="7.043495506s" podCreationTimestamp="2025-11-22 10:11:20 +0000 UTC" firstStartedPulling="2025-11-22 10:11:22.980965974 +0000 UTC m=+8097.215366247" lastFinishedPulling="2025-11-22 10:11:26.144952956 +0000 UTC m=+8100.379353229" observedRunningTime="2025-11-22 10:11:27.041700398 +0000 UTC m=+8101.276100681" watchObservedRunningTime="2025-11-22 10:11:27.043495506 +0000 UTC m=+8101.277895769" Nov 22 10:11:30 crc kubenswrapper[4789]: I1122 10:11:30.965649 4789 scope.go:117] "RemoveContainer" containerID="6e22a5a8fc3638ca6aab7b67ea550d9b289803b56baeb39c04dfbc919861ab85" Nov 22 10:11:30 crc kubenswrapper[4789]: E1122 10:11:30.966631 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 10:11:31 crc kubenswrapper[4789]: I1122 10:11:31.411604 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-p6bnq" Nov 22 10:11:31 crc kubenswrapper[4789]: I1122 10:11:31.412010 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-p6bnq" Nov 22 10:11:32 crc kubenswrapper[4789]: I1122 10:11:32.473254 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-p6bnq" podUID="136cff6d-0787-4921-a5f3-7c719483a110" containerName="registry-server" probeResult="failure" output=< Nov 22 10:11:32 crc kubenswrapper[4789]: timeout: failed to connect service ":50051" within 1s Nov 22 10:11:32 crc kubenswrapper[4789]: > Nov 22 10:11:41 crc kubenswrapper[4789]: I1122 10:11:41.469551 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-p6bnq" Nov 22 10:11:41 crc kubenswrapper[4789]: I1122 10:11:41.527412 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-p6bnq" Nov 22 10:11:41 crc kubenswrapper[4789]: I1122 10:11:41.722312 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-p6bnq"] Nov 22 10:11:43 crc kubenswrapper[4789]: I1122 10:11:43.186807 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-p6bnq" podUID="136cff6d-0787-4921-a5f3-7c719483a110" containerName="registry-server" containerID="cri-o://4bb35c394c8244a7be16b7fe5b58d5238d82c8344904aab0f1804d11e6c8f23f" gracePeriod=2 Nov 22 10:11:43 crc kubenswrapper[4789]: I1122 10:11:43.969023 4789 scope.go:117] "RemoveContainer" containerID="6e22a5a8fc3638ca6aab7b67ea550d9b289803b56baeb39c04dfbc919861ab85" Nov 22 10:11:44 crc kubenswrapper[4789]: I1122 10:11:44.211833 4789 generic.go:334] "Generic (PLEG): container finished" podID="136cff6d-0787-4921-a5f3-7c719483a110" containerID="4bb35c394c8244a7be16b7fe5b58d5238d82c8344904aab0f1804d11e6c8f23f" exitCode=0 Nov 22 10:11:44 crc kubenswrapper[4789]: I1122 10:11:44.211873 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p6bnq" event={"ID":"136cff6d-0787-4921-a5f3-7c719483a110","Type":"ContainerDied","Data":"4bb35c394c8244a7be16b7fe5b58d5238d82c8344904aab0f1804d11e6c8f23f"} Nov 22 10:11:44 crc kubenswrapper[4789]: I1122 10:11:44.211898 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p6bnq" event={"ID":"136cff6d-0787-4921-a5f3-7c719483a110","Type":"ContainerDied","Data":"527aedd0928dd480fefed02867a9b71273db65e5e2aa6901562b4851884ba340"} Nov 22 10:11:44 crc kubenswrapper[4789]: I1122 10:11:44.211909 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="527aedd0928dd480fefed02867a9b71273db65e5e2aa6901562b4851884ba340" Nov 22 10:11:44 crc kubenswrapper[4789]: I1122 10:11:44.213572 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p6bnq" Nov 22 10:11:44 crc kubenswrapper[4789]: I1122 10:11:44.320145 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/136cff6d-0787-4921-a5f3-7c719483a110-utilities\") pod \"136cff6d-0787-4921-a5f3-7c719483a110\" (UID: \"136cff6d-0787-4921-a5f3-7c719483a110\") " Nov 22 10:11:44 crc kubenswrapper[4789]: I1122 10:11:44.320786 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/136cff6d-0787-4921-a5f3-7c719483a110-catalog-content\") pod \"136cff6d-0787-4921-a5f3-7c719483a110\" (UID: \"136cff6d-0787-4921-a5f3-7c719483a110\") " Nov 22 10:11:44 crc kubenswrapper[4789]: I1122 10:11:44.320876 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9hb7d\" (UniqueName: \"kubernetes.io/projected/136cff6d-0787-4921-a5f3-7c719483a110-kube-api-access-9hb7d\") pod \"136cff6d-0787-4921-a5f3-7c719483a110\" (UID: \"136cff6d-0787-4921-a5f3-7c719483a110\") " Nov 22 10:11:44 crc kubenswrapper[4789]: I1122 10:11:44.321072 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/136cff6d-0787-4921-a5f3-7c719483a110-utilities" (OuterVolumeSpecName: "utilities") pod "136cff6d-0787-4921-a5f3-7c719483a110" (UID: "136cff6d-0787-4921-a5f3-7c719483a110"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 10:11:44 crc kubenswrapper[4789]: I1122 10:11:44.321620 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/136cff6d-0787-4921-a5f3-7c719483a110-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 10:11:44 crc kubenswrapper[4789]: I1122 10:11:44.326579 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/136cff6d-0787-4921-a5f3-7c719483a110-kube-api-access-9hb7d" (OuterVolumeSpecName: "kube-api-access-9hb7d") pod "136cff6d-0787-4921-a5f3-7c719483a110" (UID: "136cff6d-0787-4921-a5f3-7c719483a110"). InnerVolumeSpecName "kube-api-access-9hb7d". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 10:11:44 crc kubenswrapper[4789]: I1122 10:11:44.418494 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/136cff6d-0787-4921-a5f3-7c719483a110-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "136cff6d-0787-4921-a5f3-7c719483a110" (UID: "136cff6d-0787-4921-a5f3-7c719483a110"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 10:11:44 crc kubenswrapper[4789]: I1122 10:11:44.423526 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/136cff6d-0787-4921-a5f3-7c719483a110-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 10:11:44 crc kubenswrapper[4789]: I1122 10:11:44.423562 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9hb7d\" (UniqueName: \"kubernetes.io/projected/136cff6d-0787-4921-a5f3-7c719483a110-kube-api-access-9hb7d\") on node \"crc\" DevicePath \"\"" Nov 22 10:11:45 crc kubenswrapper[4789]: I1122 10:11:45.222411 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" event={"ID":"58aa8071-7ecc-4692-8789-2db04ec70647","Type":"ContainerStarted","Data":"4141db46cc7225c6dcef7ae55b4ce0014526d33bf629421689170714974fa6ca"} Nov 22 10:11:45 crc kubenswrapper[4789]: I1122 10:11:45.222452 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p6bnq" Nov 22 10:11:45 crc kubenswrapper[4789]: I1122 10:11:45.267740 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-p6bnq"] Nov 22 10:11:45 crc kubenswrapper[4789]: I1122 10:11:45.300322 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-p6bnq"] Nov 22 10:11:45 crc kubenswrapper[4789]: I1122 10:11:45.976915 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="136cff6d-0787-4921-a5f3-7c719483a110" path="/var/lib/kubelet/pods/136cff6d-0787-4921-a5f3-7c719483a110/volumes" Nov 22 10:12:10 crc kubenswrapper[4789]: I1122 10:12:10.459166 4789 generic.go:334] "Generic (PLEG): container finished" podID="d3fd6f1c-a6a5-406f-9eae-b6627046b275" containerID="dddf6b7c809531681c5d607f59bd710e2a6e43e92fb27fa786832ff55a5981d4" exitCode=0 Nov 22 10:12:10 crc kubenswrapper[4789]: I1122 10:12:10.459276 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tobiko-tests-tobiko-s01-sanity" event={"ID":"d3fd6f1c-a6a5-406f-9eae-b6627046b275","Type":"ContainerDied","Data":"dddf6b7c809531681c5d607f59bd710e2a6e43e92fb27fa786832ff55a5981d4"} Nov 22 10:12:12 crc kubenswrapper[4789]: I1122 10:12:12.048695 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tobiko-tests-tobiko-s01-sanity" Nov 22 10:12:12 crc kubenswrapper[4789]: I1122 10:12:12.071997 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/d3fd6f1c-a6a5-406f-9eae-b6627046b275-openstack-config-secret\") pod \"d3fd6f1c-a6a5-406f-9eae-b6627046b275\" (UID: \"d3fd6f1c-a6a5-406f-9eae-b6627046b275\") " Nov 22 10:12:12 crc kubenswrapper[4789]: I1122 10:12:12.072136 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4psw4\" (UniqueName: \"kubernetes.io/projected/d3fd6f1c-a6a5-406f-9eae-b6627046b275-kube-api-access-4psw4\") pod \"d3fd6f1c-a6a5-406f-9eae-b6627046b275\" (UID: \"d3fd6f1c-a6a5-406f-9eae-b6627046b275\") " Nov 22 10:12:12 crc kubenswrapper[4789]: I1122 10:12:12.072157 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tobiko-config\" (UniqueName: \"kubernetes.io/configmap/d3fd6f1c-a6a5-406f-9eae-b6627046b275-tobiko-config\") pod \"d3fd6f1c-a6a5-406f-9eae-b6627046b275\" (UID: \"d3fd6f1c-a6a5-406f-9eae-b6627046b275\") " Nov 22 10:12:12 crc kubenswrapper[4789]: I1122 10:12:12.073090 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tobiko-public-key\" (UniqueName: \"kubernetes.io/configmap/d3fd6f1c-a6a5-406f-9eae-b6627046b275-tobiko-public-key\") pod \"d3fd6f1c-a6a5-406f-9eae-b6627046b275\" (UID: \"d3fd6f1c-a6a5-406f-9eae-b6627046b275\") " Nov 22 10:12:12 crc kubenswrapper[4789]: I1122 10:12:12.073163 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/d3fd6f1c-a6a5-406f-9eae-b6627046b275-test-operator-ephemeral-workdir\") pod \"d3fd6f1c-a6a5-406f-9eae-b6627046b275\" (UID: \"d3fd6f1c-a6a5-406f-9eae-b6627046b275\") " Nov 22 10:12:12 crc kubenswrapper[4789]: I1122 10:12:12.073255 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/d3fd6f1c-a6a5-406f-9eae-b6627046b275-ca-certs\") pod \"d3fd6f1c-a6a5-406f-9eae-b6627046b275\" (UID: \"d3fd6f1c-a6a5-406f-9eae-b6627046b275\") " Nov 22 10:12:12 crc kubenswrapper[4789]: I1122 10:12:12.073308 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tobiko-private-key\" (UniqueName: \"kubernetes.io/configmap/d3fd6f1c-a6a5-406f-9eae-b6627046b275-tobiko-private-key\") pod \"d3fd6f1c-a6a5-406f-9eae-b6627046b275\" (UID: \"d3fd6f1c-a6a5-406f-9eae-b6627046b275\") " Nov 22 10:12:12 crc kubenswrapper[4789]: I1122 10:12:12.073420 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"d3fd6f1c-a6a5-406f-9eae-b6627046b275\" (UID: \"d3fd6f1c-a6a5-406f-9eae-b6627046b275\") " Nov 22 10:12:12 crc kubenswrapper[4789]: I1122 10:12:12.073503 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/secret/d3fd6f1c-a6a5-406f-9eae-b6627046b275-kubeconfig\") pod \"d3fd6f1c-a6a5-406f-9eae-b6627046b275\" (UID: \"d3fd6f1c-a6a5-406f-9eae-b6627046b275\") " Nov 22 10:12:12 crc kubenswrapper[4789]: I1122 10:12:12.073531 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d3fd6f1c-a6a5-406f-9eae-b6627046b275-ceph\") pod \"d3fd6f1c-a6a5-406f-9eae-b6627046b275\" (UID: \"d3fd6f1c-a6a5-406f-9eae-b6627046b275\") " Nov 22 10:12:12 crc kubenswrapper[4789]: I1122 10:12:12.073552 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/d3fd6f1c-a6a5-406f-9eae-b6627046b275-test-operator-ephemeral-temporary\") pod \"d3fd6f1c-a6a5-406f-9eae-b6627046b275\" (UID: \"d3fd6f1c-a6a5-406f-9eae-b6627046b275\") " Nov 22 10:12:12 crc kubenswrapper[4789]: I1122 10:12:12.073590 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/d3fd6f1c-a6a5-406f-9eae-b6627046b275-test-operator-clouds-config\") pod \"d3fd6f1c-a6a5-406f-9eae-b6627046b275\" (UID: \"d3fd6f1c-a6a5-406f-9eae-b6627046b275\") " Nov 22 10:12:12 crc kubenswrapper[4789]: I1122 10:12:12.075393 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d3fd6f1c-a6a5-406f-9eae-b6627046b275-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "d3fd6f1c-a6a5-406f-9eae-b6627046b275" (UID: "d3fd6f1c-a6a5-406f-9eae-b6627046b275"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 10:12:12 crc kubenswrapper[4789]: I1122 10:12:12.091377 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "test-operator-logs") pod "d3fd6f1c-a6a5-406f-9eae-b6627046b275" (UID: "d3fd6f1c-a6a5-406f-9eae-b6627046b275"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 22 10:12:12 crc kubenswrapper[4789]: I1122 10:12:12.094105 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3fd6f1c-a6a5-406f-9eae-b6627046b275-kube-api-access-4psw4" (OuterVolumeSpecName: "kube-api-access-4psw4") pod "d3fd6f1c-a6a5-406f-9eae-b6627046b275" (UID: "d3fd6f1c-a6a5-406f-9eae-b6627046b275"). InnerVolumeSpecName "kube-api-access-4psw4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 10:12:12 crc kubenswrapper[4789]: I1122 10:12:12.096196 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3fd6f1c-a6a5-406f-9eae-b6627046b275-ceph" (OuterVolumeSpecName: "ceph") pod "d3fd6f1c-a6a5-406f-9eae-b6627046b275" (UID: "d3fd6f1c-a6a5-406f-9eae-b6627046b275"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 10:12:12 crc kubenswrapper[4789]: I1122 10:12:12.110723 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d3fd6f1c-a6a5-406f-9eae-b6627046b275-tobiko-public-key" (OuterVolumeSpecName: "tobiko-public-key") pod "d3fd6f1c-a6a5-406f-9eae-b6627046b275" (UID: "d3fd6f1c-a6a5-406f-9eae-b6627046b275"). InnerVolumeSpecName "tobiko-public-key". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 10:12:12 crc kubenswrapper[4789]: I1122 10:12:12.127944 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d3fd6f1c-a6a5-406f-9eae-b6627046b275-tobiko-private-key" (OuterVolumeSpecName: "tobiko-private-key") pod "d3fd6f1c-a6a5-406f-9eae-b6627046b275" (UID: "d3fd6f1c-a6a5-406f-9eae-b6627046b275"). InnerVolumeSpecName "tobiko-private-key". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 10:12:12 crc kubenswrapper[4789]: I1122 10:12:12.131992 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3fd6f1c-a6a5-406f-9eae-b6627046b275-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "d3fd6f1c-a6a5-406f-9eae-b6627046b275" (UID: "d3fd6f1c-a6a5-406f-9eae-b6627046b275"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 10:12:12 crc kubenswrapper[4789]: I1122 10:12:12.133723 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3fd6f1c-a6a5-406f-9eae-b6627046b275-kubeconfig" (OuterVolumeSpecName: "kubeconfig") pod "d3fd6f1c-a6a5-406f-9eae-b6627046b275" (UID: "d3fd6f1c-a6a5-406f-9eae-b6627046b275"). InnerVolumeSpecName "kubeconfig". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 10:12:12 crc kubenswrapper[4789]: I1122 10:12:12.143493 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d3fd6f1c-a6a5-406f-9eae-b6627046b275-tobiko-config" (OuterVolumeSpecName: "tobiko-config") pod "d3fd6f1c-a6a5-406f-9eae-b6627046b275" (UID: "d3fd6f1c-a6a5-406f-9eae-b6627046b275"). InnerVolumeSpecName "tobiko-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 10:12:12 crc kubenswrapper[4789]: I1122 10:12:12.145211 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d3fd6f1c-a6a5-406f-9eae-b6627046b275-test-operator-clouds-config" (OuterVolumeSpecName: "test-operator-clouds-config") pod "d3fd6f1c-a6a5-406f-9eae-b6627046b275" (UID: "d3fd6f1c-a6a5-406f-9eae-b6627046b275"). InnerVolumeSpecName "test-operator-clouds-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 10:12:12 crc kubenswrapper[4789]: I1122 10:12:12.158164 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3fd6f1c-a6a5-406f-9eae-b6627046b275-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "d3fd6f1c-a6a5-406f-9eae-b6627046b275" (UID: "d3fd6f1c-a6a5-406f-9eae-b6627046b275"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 10:12:12 crc kubenswrapper[4789]: I1122 10:12:12.176290 4789 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Nov 22 10:12:12 crc kubenswrapper[4789]: I1122 10:12:12.176458 4789 reconciler_common.go:293] "Volume detached for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/secret/d3fd6f1c-a6a5-406f-9eae-b6627046b275-kubeconfig\") on node \"crc\" DevicePath \"\"" Nov 22 10:12:12 crc kubenswrapper[4789]: I1122 10:12:12.176523 4789 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d3fd6f1c-a6a5-406f-9eae-b6627046b275-ceph\") on node \"crc\" DevicePath \"\"" Nov 22 10:12:12 crc kubenswrapper[4789]: I1122 10:12:12.176579 4789 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/d3fd6f1c-a6a5-406f-9eae-b6627046b275-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Nov 22 10:12:12 crc kubenswrapper[4789]: I1122 10:12:12.176637 4789 reconciler_common.go:293] "Volume detached for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/d3fd6f1c-a6a5-406f-9eae-b6627046b275-test-operator-clouds-config\") on node \"crc\" DevicePath \"\"" Nov 22 10:12:12 crc kubenswrapper[4789]: I1122 10:12:12.176699 4789 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/d3fd6f1c-a6a5-406f-9eae-b6627046b275-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Nov 22 10:12:12 crc kubenswrapper[4789]: I1122 10:12:12.176773 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4psw4\" (UniqueName: \"kubernetes.io/projected/d3fd6f1c-a6a5-406f-9eae-b6627046b275-kube-api-access-4psw4\") on node \"crc\" DevicePath \"\"" Nov 22 10:12:12 crc kubenswrapper[4789]: I1122 10:12:12.176844 4789 reconciler_common.go:293] "Volume detached for volume \"tobiko-config\" (UniqueName: \"kubernetes.io/configmap/d3fd6f1c-a6a5-406f-9eae-b6627046b275-tobiko-config\") on node \"crc\" DevicePath \"\"" Nov 22 10:12:12 crc kubenswrapper[4789]: I1122 10:12:12.176902 4789 reconciler_common.go:293] "Volume detached for volume \"tobiko-public-key\" (UniqueName: \"kubernetes.io/configmap/d3fd6f1c-a6a5-406f-9eae-b6627046b275-tobiko-public-key\") on node \"crc\" DevicePath \"\"" Nov 22 10:12:12 crc kubenswrapper[4789]: I1122 10:12:12.176957 4789 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/d3fd6f1c-a6a5-406f-9eae-b6627046b275-ca-certs\") on node \"crc\" DevicePath \"\"" Nov 22 10:12:12 crc kubenswrapper[4789]: I1122 10:12:12.177015 4789 reconciler_common.go:293] "Volume detached for volume \"tobiko-private-key\" (UniqueName: \"kubernetes.io/configmap/d3fd6f1c-a6a5-406f-9eae-b6627046b275-tobiko-private-key\") on node \"crc\" DevicePath \"\"" Nov 22 10:12:12 crc kubenswrapper[4789]: I1122 10:12:12.196262 4789 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Nov 22 10:12:12 crc kubenswrapper[4789]: I1122 10:12:12.278985 4789 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Nov 22 10:12:12 crc kubenswrapper[4789]: I1122 10:12:12.486265 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tobiko-tests-tobiko-s01-sanity" event={"ID":"d3fd6f1c-a6a5-406f-9eae-b6627046b275","Type":"ContainerDied","Data":"bdfaab3ad8dd901dce072b82a8906d905eeb492408c396b76af3945635a2ce3a"} Nov 22 10:12:12 crc kubenswrapper[4789]: I1122 10:12:12.486307 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bdfaab3ad8dd901dce072b82a8906d905eeb492408c396b76af3945635a2ce3a" Nov 22 10:12:12 crc kubenswrapper[4789]: I1122 10:12:12.486350 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tobiko-tests-tobiko-s01-sanity" Nov 22 10:12:13 crc kubenswrapper[4789]: I1122 10:12:13.603413 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d3fd6f1c-a6a5-406f-9eae-b6627046b275-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "d3fd6f1c-a6a5-406f-9eae-b6627046b275" (UID: "d3fd6f1c-a6a5-406f-9eae-b6627046b275"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 10:12:13 crc kubenswrapper[4789]: I1122 10:12:13.613003 4789 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/d3fd6f1c-a6a5-406f-9eae-b6627046b275-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Nov 22 10:12:15 crc kubenswrapper[4789]: I1122 10:12:15.057234 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko"] Nov 22 10:12:15 crc kubenswrapper[4789]: E1122 10:12:15.057957 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="136cff6d-0787-4921-a5f3-7c719483a110" containerName="extract-content" Nov 22 10:12:15 crc kubenswrapper[4789]: I1122 10:12:15.057986 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="136cff6d-0787-4921-a5f3-7c719483a110" containerName="extract-content" Nov 22 10:12:15 crc kubenswrapper[4789]: E1122 10:12:15.058012 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3fd6f1c-a6a5-406f-9eae-b6627046b275" containerName="tobiko-tests-tobiko" Nov 22 10:12:15 crc kubenswrapper[4789]: I1122 10:12:15.058021 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3fd6f1c-a6a5-406f-9eae-b6627046b275" containerName="tobiko-tests-tobiko" Nov 22 10:12:15 crc kubenswrapper[4789]: E1122 10:12:15.058034 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="136cff6d-0787-4921-a5f3-7c719483a110" containerName="registry-server" Nov 22 10:12:15 crc kubenswrapper[4789]: I1122 10:12:15.058043 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="136cff6d-0787-4921-a5f3-7c719483a110" containerName="registry-server" Nov 22 10:12:15 crc kubenswrapper[4789]: E1122 10:12:15.058061 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="136cff6d-0787-4921-a5f3-7c719483a110" containerName="extract-utilities" Nov 22 10:12:15 crc kubenswrapper[4789]: I1122 10:12:15.058071 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="136cff6d-0787-4921-a5f3-7c719483a110" containerName="extract-utilities" Nov 22 10:12:15 crc kubenswrapper[4789]: I1122 10:12:15.058302 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3fd6f1c-a6a5-406f-9eae-b6627046b275" containerName="tobiko-tests-tobiko" Nov 22 10:12:15 crc kubenswrapper[4789]: I1122 10:12:15.058328 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="136cff6d-0787-4921-a5f3-7c719483a110" containerName="registry-server" Nov 22 10:12:15 crc kubenswrapper[4789]: I1122 10:12:15.059012 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko" Nov 22 10:12:15 crc kubenswrapper[4789]: I1122 10:12:15.069828 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko"] Nov 22 10:12:15 crc kubenswrapper[4789]: I1122 10:12:15.146379 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lb9dx\" (UniqueName: \"kubernetes.io/projected/01b8c525-b595-4f88-825a-0eda2c77c604-kube-api-access-lb9dx\") pod \"test-operator-logs-pod-tobiko-tobiko-tests-tobiko\" (UID: \"01b8c525-b595-4f88-825a-0eda2c77c604\") " pod="openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko" Nov 22 10:12:15 crc kubenswrapper[4789]: I1122 10:12:15.146447 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"test-operator-logs-pod-tobiko-tobiko-tests-tobiko\" (UID: \"01b8c525-b595-4f88-825a-0eda2c77c604\") " pod="openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko" Nov 22 10:12:15 crc kubenswrapper[4789]: I1122 10:12:15.248126 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lb9dx\" (UniqueName: \"kubernetes.io/projected/01b8c525-b595-4f88-825a-0eda2c77c604-kube-api-access-lb9dx\") pod \"test-operator-logs-pod-tobiko-tobiko-tests-tobiko\" (UID: \"01b8c525-b595-4f88-825a-0eda2c77c604\") " pod="openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko" Nov 22 10:12:15 crc kubenswrapper[4789]: I1122 10:12:15.248191 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"test-operator-logs-pod-tobiko-tobiko-tests-tobiko\" (UID: \"01b8c525-b595-4f88-825a-0eda2c77c604\") " pod="openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko" Nov 22 10:12:15 crc kubenswrapper[4789]: I1122 10:12:15.248648 4789 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"test-operator-logs-pod-tobiko-tobiko-tests-tobiko\" (UID: \"01b8c525-b595-4f88-825a-0eda2c77c604\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko" Nov 22 10:12:15 crc kubenswrapper[4789]: I1122 10:12:15.278046 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"test-operator-logs-pod-tobiko-tobiko-tests-tobiko\" (UID: \"01b8c525-b595-4f88-825a-0eda2c77c604\") " pod="openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko" Nov 22 10:12:15 crc kubenswrapper[4789]: I1122 10:12:15.287277 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lb9dx\" (UniqueName: \"kubernetes.io/projected/01b8c525-b595-4f88-825a-0eda2c77c604-kube-api-access-lb9dx\") pod \"test-operator-logs-pod-tobiko-tobiko-tests-tobiko\" (UID: \"01b8c525-b595-4f88-825a-0eda2c77c604\") " pod="openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko" Nov 22 10:12:15 crc kubenswrapper[4789]: I1122 10:12:15.382843 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko" Nov 22 10:12:15 crc kubenswrapper[4789]: I1122 10:12:15.862678 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko"] Nov 22 10:12:16 crc kubenswrapper[4789]: I1122 10:12:16.542334 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko" event={"ID":"01b8c525-b595-4f88-825a-0eda2c77c604","Type":"ContainerStarted","Data":"f3e054fdd1885a1576ba6f54152dc1375d27f301d927035b397360d8d84edab9"} Nov 22 10:12:17 crc kubenswrapper[4789]: I1122 10:12:17.569527 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko" event={"ID":"01b8c525-b595-4f88-825a-0eda2c77c604","Type":"ContainerStarted","Data":"d65bdf4e1a69689dfb54d6eccd26cb652da0edb9ffdf830a92bd49c3f5e49f0e"} Nov 22 10:12:17 crc kubenswrapper[4789]: I1122 10:12:17.592954 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko" podStartSLOduration=2.064505363 podStartE2EDuration="2.592934172s" podCreationTimestamp="2025-11-22 10:12:15 +0000 UTC" firstStartedPulling="2025-11-22 10:12:15.874886581 +0000 UTC m=+8150.109286894" lastFinishedPulling="2025-11-22 10:12:16.40331543 +0000 UTC m=+8150.637715703" observedRunningTime="2025-11-22 10:12:17.58467828 +0000 UTC m=+8151.819078593" watchObservedRunningTime="2025-11-22 10:12:17.592934172 +0000 UTC m=+8151.827334455" Nov 22 10:12:29 crc kubenswrapper[4789]: I1122 10:12:29.579705 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ansibletest-ansibletest"] Nov 22 10:12:29 crc kubenswrapper[4789]: I1122 10:12:29.582856 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ansibletest-ansibletest" Nov 22 10:12:29 crc kubenswrapper[4789]: I1122 10:12:29.588315 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Nov 22 10:12:29 crc kubenswrapper[4789]: I1122 10:12:29.588775 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 22 10:12:29 crc kubenswrapper[4789]: I1122 10:12:29.589929 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ansibletest-ansibletest"] Nov 22 10:12:29 crc kubenswrapper[4789]: I1122 10:12:29.740324 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/0e385574-644b-4047-a329-a392a52c663f-ca-certs\") pod \"ansibletest-ansibletest\" (UID: \"0e385574-644b-4047-a329-a392a52c663f\") " pod="openstack/ansibletest-ansibletest" Nov 22 10:12:29 crc kubenswrapper[4789]: I1122 10:12:29.740635 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/0e385574-644b-4047-a329-a392a52c663f-openstack-config-secret\") pod \"ansibletest-ansibletest\" (UID: \"0e385574-644b-4047-a329-a392a52c663f\") " pod="openstack/ansibletest-ansibletest" Nov 22 10:12:29 crc kubenswrapper[4789]: I1122 10:12:29.740806 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"compute-ssh-secret\" (UniqueName: \"kubernetes.io/secret/0e385574-644b-4047-a329-a392a52c663f-compute-ssh-secret\") pod \"ansibletest-ansibletest\" (UID: \"0e385574-644b-4047-a329-a392a52c663f\") " pod="openstack/ansibletest-ansibletest" Nov 22 10:12:29 crc kubenswrapper[4789]: I1122 10:12:29.740966 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0e385574-644b-4047-a329-a392a52c663f-ceph\") pod \"ansibletest-ansibletest\" (UID: \"0e385574-644b-4047-a329-a392a52c663f\") " pod="openstack/ansibletest-ansibletest" Nov 22 10:12:29 crc kubenswrapper[4789]: I1122 10:12:29.741085 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"workload-ssh-secret\" (UniqueName: \"kubernetes.io/secret/0e385574-644b-4047-a329-a392a52c663f-workload-ssh-secret\") pod \"ansibletest-ansibletest\" (UID: \"0e385574-644b-4047-a329-a392a52c663f\") " pod="openstack/ansibletest-ansibletest" Nov 22 10:12:29 crc kubenswrapper[4789]: I1122 10:12:29.741166 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/0e385574-644b-4047-a329-a392a52c663f-openstack-config\") pod \"ansibletest-ansibletest\" (UID: \"0e385574-644b-4047-a329-a392a52c663f\") " pod="openstack/ansibletest-ansibletest" Nov 22 10:12:29 crc kubenswrapper[4789]: I1122 10:12:29.741375 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/0e385574-644b-4047-a329-a392a52c663f-test-operator-ephemeral-workdir\") pod \"ansibletest-ansibletest\" (UID: \"0e385574-644b-4047-a329-a392a52c663f\") " pod="openstack/ansibletest-ansibletest" Nov 22 10:12:29 crc kubenswrapper[4789]: I1122 10:12:29.741430 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bhvnp\" (UniqueName: \"kubernetes.io/projected/0e385574-644b-4047-a329-a392a52c663f-kube-api-access-bhvnp\") pod \"ansibletest-ansibletest\" (UID: \"0e385574-644b-4047-a329-a392a52c663f\") " pod="openstack/ansibletest-ansibletest" Nov 22 10:12:29 crc kubenswrapper[4789]: I1122 10:12:29.741456 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/0e385574-644b-4047-a329-a392a52c663f-test-operator-ephemeral-temporary\") pod \"ansibletest-ansibletest\" (UID: \"0e385574-644b-4047-a329-a392a52c663f\") " pod="openstack/ansibletest-ansibletest" Nov 22 10:12:29 crc kubenswrapper[4789]: I1122 10:12:29.741508 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ansibletest-ansibletest\" (UID: \"0e385574-644b-4047-a329-a392a52c663f\") " pod="openstack/ansibletest-ansibletest" Nov 22 10:12:29 crc kubenswrapper[4789]: I1122 10:12:29.843461 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ansibletest-ansibletest\" (UID: \"0e385574-644b-4047-a329-a392a52c663f\") " pod="openstack/ansibletest-ansibletest" Nov 22 10:12:29 crc kubenswrapper[4789]: I1122 10:12:29.843579 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/0e385574-644b-4047-a329-a392a52c663f-ca-certs\") pod \"ansibletest-ansibletest\" (UID: \"0e385574-644b-4047-a329-a392a52c663f\") " pod="openstack/ansibletest-ansibletest" Nov 22 10:12:29 crc kubenswrapper[4789]: I1122 10:12:29.843683 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/0e385574-644b-4047-a329-a392a52c663f-openstack-config-secret\") pod \"ansibletest-ansibletest\" (UID: \"0e385574-644b-4047-a329-a392a52c663f\") " pod="openstack/ansibletest-ansibletest" Nov 22 10:12:29 crc kubenswrapper[4789]: I1122 10:12:29.843787 4789 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ansibletest-ansibletest\" (UID: \"0e385574-644b-4047-a329-a392a52c663f\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/ansibletest-ansibletest" Nov 22 10:12:29 crc kubenswrapper[4789]: I1122 10:12:29.843815 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"compute-ssh-secret\" (UniqueName: \"kubernetes.io/secret/0e385574-644b-4047-a329-a392a52c663f-compute-ssh-secret\") pod \"ansibletest-ansibletest\" (UID: \"0e385574-644b-4047-a329-a392a52c663f\") " pod="openstack/ansibletest-ansibletest" Nov 22 10:12:29 crc kubenswrapper[4789]: I1122 10:12:29.843876 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0e385574-644b-4047-a329-a392a52c663f-ceph\") pod \"ansibletest-ansibletest\" (UID: \"0e385574-644b-4047-a329-a392a52c663f\") " pod="openstack/ansibletest-ansibletest" Nov 22 10:12:29 crc kubenswrapper[4789]: I1122 10:12:29.843970 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"workload-ssh-secret\" (UniqueName: \"kubernetes.io/secret/0e385574-644b-4047-a329-a392a52c663f-workload-ssh-secret\") pod \"ansibletest-ansibletest\" (UID: \"0e385574-644b-4047-a329-a392a52c663f\") " pod="openstack/ansibletest-ansibletest" Nov 22 10:12:29 crc kubenswrapper[4789]: I1122 10:12:29.844035 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/0e385574-644b-4047-a329-a392a52c663f-openstack-config\") pod \"ansibletest-ansibletest\" (UID: \"0e385574-644b-4047-a329-a392a52c663f\") " pod="openstack/ansibletest-ansibletest" Nov 22 10:12:29 crc kubenswrapper[4789]: I1122 10:12:29.844232 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/0e385574-644b-4047-a329-a392a52c663f-test-operator-ephemeral-workdir\") pod \"ansibletest-ansibletest\" (UID: \"0e385574-644b-4047-a329-a392a52c663f\") " pod="openstack/ansibletest-ansibletest" Nov 22 10:12:29 crc kubenswrapper[4789]: I1122 10:12:29.844302 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bhvnp\" (UniqueName: \"kubernetes.io/projected/0e385574-644b-4047-a329-a392a52c663f-kube-api-access-bhvnp\") pod \"ansibletest-ansibletest\" (UID: \"0e385574-644b-4047-a329-a392a52c663f\") " pod="openstack/ansibletest-ansibletest" Nov 22 10:12:29 crc kubenswrapper[4789]: I1122 10:12:29.844351 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/0e385574-644b-4047-a329-a392a52c663f-test-operator-ephemeral-temporary\") pod \"ansibletest-ansibletest\" (UID: \"0e385574-644b-4047-a329-a392a52c663f\") " pod="openstack/ansibletest-ansibletest" Nov 22 10:12:29 crc kubenswrapper[4789]: I1122 10:12:29.845248 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/0e385574-644b-4047-a329-a392a52c663f-test-operator-ephemeral-temporary\") pod \"ansibletest-ansibletest\" (UID: \"0e385574-644b-4047-a329-a392a52c663f\") " pod="openstack/ansibletest-ansibletest" Nov 22 10:12:29 crc kubenswrapper[4789]: I1122 10:12:29.845291 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/0e385574-644b-4047-a329-a392a52c663f-test-operator-ephemeral-workdir\") pod \"ansibletest-ansibletest\" (UID: \"0e385574-644b-4047-a329-a392a52c663f\") " pod="openstack/ansibletest-ansibletest" Nov 22 10:12:29 crc kubenswrapper[4789]: I1122 10:12:29.846400 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/0e385574-644b-4047-a329-a392a52c663f-openstack-config\") pod \"ansibletest-ansibletest\" (UID: \"0e385574-644b-4047-a329-a392a52c663f\") " pod="openstack/ansibletest-ansibletest" Nov 22 10:12:29 crc kubenswrapper[4789]: I1122 10:12:29.850999 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"compute-ssh-secret\" (UniqueName: \"kubernetes.io/secret/0e385574-644b-4047-a329-a392a52c663f-compute-ssh-secret\") pod \"ansibletest-ansibletest\" (UID: \"0e385574-644b-4047-a329-a392a52c663f\") " pod="openstack/ansibletest-ansibletest" Nov 22 10:12:29 crc kubenswrapper[4789]: I1122 10:12:29.852208 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/0e385574-644b-4047-a329-a392a52c663f-ca-certs\") pod \"ansibletest-ansibletest\" (UID: \"0e385574-644b-4047-a329-a392a52c663f\") " pod="openstack/ansibletest-ansibletest" Nov 22 10:12:29 crc kubenswrapper[4789]: I1122 10:12:29.856655 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/0e385574-644b-4047-a329-a392a52c663f-openstack-config-secret\") pod \"ansibletest-ansibletest\" (UID: \"0e385574-644b-4047-a329-a392a52c663f\") " pod="openstack/ansibletest-ansibletest" Nov 22 10:12:29 crc kubenswrapper[4789]: I1122 10:12:29.857901 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"workload-ssh-secret\" (UniqueName: \"kubernetes.io/secret/0e385574-644b-4047-a329-a392a52c663f-workload-ssh-secret\") pod \"ansibletest-ansibletest\" (UID: \"0e385574-644b-4047-a329-a392a52c663f\") " pod="openstack/ansibletest-ansibletest" Nov 22 10:12:29 crc kubenswrapper[4789]: I1122 10:12:29.858937 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0e385574-644b-4047-a329-a392a52c663f-ceph\") pod \"ansibletest-ansibletest\" (UID: \"0e385574-644b-4047-a329-a392a52c663f\") " pod="openstack/ansibletest-ansibletest" Nov 22 10:12:29 crc kubenswrapper[4789]: I1122 10:12:29.865515 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bhvnp\" (UniqueName: \"kubernetes.io/projected/0e385574-644b-4047-a329-a392a52c663f-kube-api-access-bhvnp\") pod \"ansibletest-ansibletest\" (UID: \"0e385574-644b-4047-a329-a392a52c663f\") " pod="openstack/ansibletest-ansibletest" Nov 22 10:12:29 crc kubenswrapper[4789]: I1122 10:12:29.873865 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ansibletest-ansibletest\" (UID: \"0e385574-644b-4047-a329-a392a52c663f\") " pod="openstack/ansibletest-ansibletest" Nov 22 10:12:29 crc kubenswrapper[4789]: I1122 10:12:29.915235 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ansibletest-ansibletest" Nov 22 10:12:30 crc kubenswrapper[4789]: I1122 10:12:30.393697 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ansibletest-ansibletest"] Nov 22 10:12:30 crc kubenswrapper[4789]: I1122 10:12:30.703574 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ansibletest-ansibletest" event={"ID":"0e385574-644b-4047-a329-a392a52c663f","Type":"ContainerStarted","Data":"d6be2f2ef07d33a8e3ad38af54f6e2111297daef6f70e1f479944f45b525b91d"} Nov 22 10:12:44 crc kubenswrapper[4789]: E1122 10:12:44.185843 4789 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ansible-tests:current-podified" Nov 22 10:12:44 crc kubenswrapper[4789]: E1122 10:12:44.187310 4789 kuberuntime_manager.go:1274] "Unhandled Error" err=< Nov 22 10:12:44 crc kubenswrapper[4789]: container &Container{Name:ansibletest-ansibletest,Image:quay.io/podified-antelope-centos9/openstack-ansible-tests:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:POD_ANSIBLE_EXTRA_VARS,Value:-e manual_run=false,ValueFrom:nil,},EnvVar{Name:POD_ANSIBLE_FILE_EXTRA_VARS,Value:--- Nov 22 10:12:44 crc kubenswrapper[4789]: foo: bar Nov 22 10:12:44 crc kubenswrapper[4789]: ,ValueFrom:nil,},EnvVar{Name:POD_ANSIBLE_GIT_BRANCH,Value:,ValueFrom:nil,},EnvVar{Name:POD_ANSIBLE_GIT_REPO,Value:https://github.com/ansible/test-playbooks,ValueFrom:nil,},EnvVar{Name:POD_ANSIBLE_INVENTORY,Value:localhost ansible_connection=local ansible_python_interpreter=python3 Nov 22 10:12:44 crc kubenswrapper[4789]: ,ValueFrom:nil,},EnvVar{Name:POD_ANSIBLE_PLAYBOOK,Value:./debug.yml,ValueFrom:nil,},EnvVar{Name:POD_DEBUG,Value:false,ValueFrom:nil,},EnvVar{Name:POD_INSTALL_COLLECTIONS,Value:,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{4 0} {} 4 DecimalSI},memory: {{4294967296 0} {} 4Gi BinarySI},},Requests:ResourceList{cpu: {{2 0} {} 2 DecimalSI},memory: {{2147483648 0} {} 2Gi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/ansible,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/AnsibleTests/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/ansible/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/var/lib/ansible/.config/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ca-bundle.trust.crt,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:workload-ssh-secret,ReadOnly:true,MountPath:/var/lib/ansible/test_keypair.key,SubPath:ssh-privatekey,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:compute-ssh-secret,ReadOnly:true,MountPath:/var/lib/ansible/.ssh/compute_id,SubPath:ssh-privatekey,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ceph,ReadOnly:true,MountPath:/etc/ceph,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bhvnp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[NET_ADMIN NET_RAW],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*227,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*227,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ansibletest-ansibletest_openstack(0e385574-644b-4047-a329-a392a52c663f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled Nov 22 10:12:44 crc kubenswrapper[4789]: > logger="UnhandledError" Nov 22 10:12:44 crc kubenswrapper[4789]: E1122 10:12:44.189359 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ansibletest-ansibletest\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ansibletest-ansibletest" podUID="0e385574-644b-4047-a329-a392a52c663f" Nov 22 10:12:44 crc kubenswrapper[4789]: E1122 10:12:44.848160 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ansibletest-ansibletest\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-ansible-tests:current-podified\\\"\"" pod="openstack/ansibletest-ansibletest" podUID="0e385574-644b-4047-a329-a392a52c663f" Nov 22 10:12:57 crc kubenswrapper[4789]: I1122 10:12:57.974618 4789 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 22 10:13:00 crc kubenswrapper[4789]: I1122 10:13:00.020830 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ansibletest-ansibletest" event={"ID":"0e385574-644b-4047-a329-a392a52c663f","Type":"ContainerStarted","Data":"9faa1307f0ca55629eda9ea3a72581553ae1d3463dbcdbcffc63754d248fda7a"} Nov 22 10:13:00 crc kubenswrapper[4789]: I1122 10:13:00.054932 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ansibletest-ansibletest" podStartSLOduration=3.950876159 podStartE2EDuration="32.054910326s" podCreationTimestamp="2025-11-22 10:12:28 +0000 UTC" firstStartedPulling="2025-11-22 10:12:30.407375491 +0000 UTC m=+8164.641775774" lastFinishedPulling="2025-11-22 10:12:58.511409628 +0000 UTC m=+8192.745809941" observedRunningTime="2025-11-22 10:13:00.043222252 +0000 UTC m=+8194.277622525" watchObservedRunningTime="2025-11-22 10:13:00.054910326 +0000 UTC m=+8194.289310599" Nov 22 10:13:01 crc kubenswrapper[4789]: I1122 10:13:01.033621 4789 generic.go:334] "Generic (PLEG): container finished" podID="0e385574-644b-4047-a329-a392a52c663f" containerID="9faa1307f0ca55629eda9ea3a72581553ae1d3463dbcdbcffc63754d248fda7a" exitCode=0 Nov 22 10:13:01 crc kubenswrapper[4789]: I1122 10:13:01.033735 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ansibletest-ansibletest" event={"ID":"0e385574-644b-4047-a329-a392a52c663f","Type":"ContainerDied","Data":"9faa1307f0ca55629eda9ea3a72581553ae1d3463dbcdbcffc63754d248fda7a"} Nov 22 10:13:02 crc kubenswrapper[4789]: I1122 10:13:02.440233 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ansibletest-ansibletest" Nov 22 10:13:02 crc kubenswrapper[4789]: I1122 10:13:02.544425 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"compute-ssh-secret\" (UniqueName: \"kubernetes.io/secret/0e385574-644b-4047-a329-a392a52c663f-compute-ssh-secret\") pod \"0e385574-644b-4047-a329-a392a52c663f\" (UID: \"0e385574-644b-4047-a329-a392a52c663f\") " Nov 22 10:13:02 crc kubenswrapper[4789]: I1122 10:13:02.544545 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/0e385574-644b-4047-a329-a392a52c663f-openstack-config\") pod \"0e385574-644b-4047-a329-a392a52c663f\" (UID: \"0e385574-644b-4047-a329-a392a52c663f\") " Nov 22 10:13:02 crc kubenswrapper[4789]: I1122 10:13:02.544629 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/0e385574-644b-4047-a329-a392a52c663f-openstack-config-secret\") pod \"0e385574-644b-4047-a329-a392a52c663f\" (UID: \"0e385574-644b-4047-a329-a392a52c663f\") " Nov 22 10:13:02 crc kubenswrapper[4789]: I1122 10:13:02.544693 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/0e385574-644b-4047-a329-a392a52c663f-ca-certs\") pod \"0e385574-644b-4047-a329-a392a52c663f\" (UID: \"0e385574-644b-4047-a329-a392a52c663f\") " Nov 22 10:13:02 crc kubenswrapper[4789]: I1122 10:13:02.544831 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/0e385574-644b-4047-a329-a392a52c663f-test-operator-ephemeral-temporary\") pod \"0e385574-644b-4047-a329-a392a52c663f\" (UID: \"0e385574-644b-4047-a329-a392a52c663f\") " Nov 22 10:13:02 crc kubenswrapper[4789]: I1122 10:13:02.544873 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"0e385574-644b-4047-a329-a392a52c663f\" (UID: \"0e385574-644b-4047-a329-a392a52c663f\") " Nov 22 10:13:02 crc kubenswrapper[4789]: I1122 10:13:02.545412 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0e385574-644b-4047-a329-a392a52c663f-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "0e385574-644b-4047-a329-a392a52c663f" (UID: "0e385574-644b-4047-a329-a392a52c663f"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 10:13:02 crc kubenswrapper[4789]: I1122 10:13:02.545610 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bhvnp\" (UniqueName: \"kubernetes.io/projected/0e385574-644b-4047-a329-a392a52c663f-kube-api-access-bhvnp\") pod \"0e385574-644b-4047-a329-a392a52c663f\" (UID: \"0e385574-644b-4047-a329-a392a52c663f\") " Nov 22 10:13:02 crc kubenswrapper[4789]: I1122 10:13:02.545682 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/0e385574-644b-4047-a329-a392a52c663f-test-operator-ephemeral-workdir\") pod \"0e385574-644b-4047-a329-a392a52c663f\" (UID: \"0e385574-644b-4047-a329-a392a52c663f\") " Nov 22 10:13:02 crc kubenswrapper[4789]: I1122 10:13:02.545744 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0e385574-644b-4047-a329-a392a52c663f-ceph\") pod \"0e385574-644b-4047-a329-a392a52c663f\" (UID: \"0e385574-644b-4047-a329-a392a52c663f\") " Nov 22 10:13:02 crc kubenswrapper[4789]: I1122 10:13:02.545802 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"workload-ssh-secret\" (UniqueName: \"kubernetes.io/secret/0e385574-644b-4047-a329-a392a52c663f-workload-ssh-secret\") pod \"0e385574-644b-4047-a329-a392a52c663f\" (UID: \"0e385574-644b-4047-a329-a392a52c663f\") " Nov 22 10:13:02 crc kubenswrapper[4789]: I1122 10:13:02.546313 4789 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/0e385574-644b-4047-a329-a392a52c663f-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Nov 22 10:13:02 crc kubenswrapper[4789]: I1122 10:13:02.558938 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "test-operator-logs") pod "0e385574-644b-4047-a329-a392a52c663f" (UID: "0e385574-644b-4047-a329-a392a52c663f"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 22 10:13:02 crc kubenswrapper[4789]: I1122 10:13:02.558955 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e385574-644b-4047-a329-a392a52c663f-ceph" (OuterVolumeSpecName: "ceph") pod "0e385574-644b-4047-a329-a392a52c663f" (UID: "0e385574-644b-4047-a329-a392a52c663f"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 10:13:02 crc kubenswrapper[4789]: I1122 10:13:02.566590 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e385574-644b-4047-a329-a392a52c663f-kube-api-access-bhvnp" (OuterVolumeSpecName: "kube-api-access-bhvnp") pod "0e385574-644b-4047-a329-a392a52c663f" (UID: "0e385574-644b-4047-a329-a392a52c663f"). InnerVolumeSpecName "kube-api-access-bhvnp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 10:13:02 crc kubenswrapper[4789]: I1122 10:13:02.579028 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0e385574-644b-4047-a329-a392a52c663f-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "0e385574-644b-4047-a329-a392a52c663f" (UID: "0e385574-644b-4047-a329-a392a52c663f"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 10:13:02 crc kubenswrapper[4789]: I1122 10:13:02.580201 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e385574-644b-4047-a329-a392a52c663f-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "0e385574-644b-4047-a329-a392a52c663f" (UID: "0e385574-644b-4047-a329-a392a52c663f"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 10:13:02 crc kubenswrapper[4789]: I1122 10:13:02.586531 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e385574-644b-4047-a329-a392a52c663f-compute-ssh-secret" (OuterVolumeSpecName: "compute-ssh-secret") pod "0e385574-644b-4047-a329-a392a52c663f" (UID: "0e385574-644b-4047-a329-a392a52c663f"). InnerVolumeSpecName "compute-ssh-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 10:13:02 crc kubenswrapper[4789]: I1122 10:13:02.597073 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e385574-644b-4047-a329-a392a52c663f-workload-ssh-secret" (OuterVolumeSpecName: "workload-ssh-secret") pod "0e385574-644b-4047-a329-a392a52c663f" (UID: "0e385574-644b-4047-a329-a392a52c663f"). InnerVolumeSpecName "workload-ssh-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 10:13:02 crc kubenswrapper[4789]: I1122 10:13:02.615872 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e385574-644b-4047-a329-a392a52c663f-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "0e385574-644b-4047-a329-a392a52c663f" (UID: "0e385574-644b-4047-a329-a392a52c663f"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 10:13:02 crc kubenswrapper[4789]: I1122 10:13:02.624018 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0e385574-644b-4047-a329-a392a52c663f-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "0e385574-644b-4047-a329-a392a52c663f" (UID: "0e385574-644b-4047-a329-a392a52c663f"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 10:13:02 crc kubenswrapper[4789]: I1122 10:13:02.648237 4789 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/0e385574-644b-4047-a329-a392a52c663f-ca-certs\") on node \"crc\" DevicePath \"\"" Nov 22 10:13:02 crc kubenswrapper[4789]: I1122 10:13:02.648498 4789 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Nov 22 10:13:02 crc kubenswrapper[4789]: I1122 10:13:02.648583 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bhvnp\" (UniqueName: \"kubernetes.io/projected/0e385574-644b-4047-a329-a392a52c663f-kube-api-access-bhvnp\") on node \"crc\" DevicePath \"\"" Nov 22 10:13:02 crc kubenswrapper[4789]: I1122 10:13:02.648681 4789 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/0e385574-644b-4047-a329-a392a52c663f-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Nov 22 10:13:02 crc kubenswrapper[4789]: I1122 10:13:02.648789 4789 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0e385574-644b-4047-a329-a392a52c663f-ceph\") on node \"crc\" DevicePath \"\"" Nov 22 10:13:02 crc kubenswrapper[4789]: I1122 10:13:02.648875 4789 reconciler_common.go:293] "Volume detached for volume \"workload-ssh-secret\" (UniqueName: \"kubernetes.io/secret/0e385574-644b-4047-a329-a392a52c663f-workload-ssh-secret\") on node \"crc\" DevicePath \"\"" Nov 22 10:13:02 crc kubenswrapper[4789]: I1122 10:13:02.648947 4789 reconciler_common.go:293] "Volume detached for volume \"compute-ssh-secret\" (UniqueName: \"kubernetes.io/secret/0e385574-644b-4047-a329-a392a52c663f-compute-ssh-secret\") on node \"crc\" DevicePath \"\"" Nov 22 10:13:02 crc kubenswrapper[4789]: I1122 10:13:02.649023 4789 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/0e385574-644b-4047-a329-a392a52c663f-openstack-config\") on node \"crc\" DevicePath \"\"" Nov 22 10:13:02 crc kubenswrapper[4789]: I1122 10:13:02.649101 4789 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/0e385574-644b-4047-a329-a392a52c663f-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Nov 22 10:13:02 crc kubenswrapper[4789]: I1122 10:13:02.667940 4789 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Nov 22 10:13:02 crc kubenswrapper[4789]: I1122 10:13:02.750873 4789 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Nov 22 10:13:03 crc kubenswrapper[4789]: I1122 10:13:03.057831 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ansibletest-ansibletest" event={"ID":"0e385574-644b-4047-a329-a392a52c663f","Type":"ContainerDied","Data":"d6be2f2ef07d33a8e3ad38af54f6e2111297daef6f70e1f479944f45b525b91d"} Nov 22 10:13:03 crc kubenswrapper[4789]: I1122 10:13:03.057925 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d6be2f2ef07d33a8e3ad38af54f6e2111297daef6f70e1f479944f45b525b91d" Nov 22 10:13:03 crc kubenswrapper[4789]: I1122 10:13:03.058006 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ansibletest-ansibletest" Nov 22 10:13:14 crc kubenswrapper[4789]: I1122 10:13:14.481795 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest"] Nov 22 10:13:14 crc kubenswrapper[4789]: E1122 10:13:14.482671 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e385574-644b-4047-a329-a392a52c663f" containerName="ansibletest-ansibletest" Nov 22 10:13:14 crc kubenswrapper[4789]: I1122 10:13:14.482683 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e385574-644b-4047-a329-a392a52c663f" containerName="ansibletest-ansibletest" Nov 22 10:13:14 crc kubenswrapper[4789]: I1122 10:13:14.482926 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e385574-644b-4047-a329-a392a52c663f" containerName="ansibletest-ansibletest" Nov 22 10:13:14 crc kubenswrapper[4789]: I1122 10:13:14.483532 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest" Nov 22 10:13:14 crc kubenswrapper[4789]: I1122 10:13:14.519764 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest"] Nov 22 10:13:14 crc kubenswrapper[4789]: I1122 10:13:14.612983 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"test-operator-logs-pod-ansibletest-ansibletest-ansibletest\" (UID: \"7108f947-1001-4cb3-992f-7e6edf860db4\") " pod="openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest" Nov 22 10:13:14 crc kubenswrapper[4789]: I1122 10:13:14.613112 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g2h6x\" (UniqueName: \"kubernetes.io/projected/7108f947-1001-4cb3-992f-7e6edf860db4-kube-api-access-g2h6x\") pod \"test-operator-logs-pod-ansibletest-ansibletest-ansibletest\" (UID: \"7108f947-1001-4cb3-992f-7e6edf860db4\") " pod="openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest" Nov 22 10:13:14 crc kubenswrapper[4789]: I1122 10:13:14.715705 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g2h6x\" (UniqueName: \"kubernetes.io/projected/7108f947-1001-4cb3-992f-7e6edf860db4-kube-api-access-g2h6x\") pod \"test-operator-logs-pod-ansibletest-ansibletest-ansibletest\" (UID: \"7108f947-1001-4cb3-992f-7e6edf860db4\") " pod="openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest" Nov 22 10:13:14 crc kubenswrapper[4789]: I1122 10:13:14.716004 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"test-operator-logs-pod-ansibletest-ansibletest-ansibletest\" (UID: \"7108f947-1001-4cb3-992f-7e6edf860db4\") " pod="openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest" Nov 22 10:13:14 crc kubenswrapper[4789]: I1122 10:13:14.716626 4789 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"test-operator-logs-pod-ansibletest-ansibletest-ansibletest\" (UID: \"7108f947-1001-4cb3-992f-7e6edf860db4\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest" Nov 22 10:13:14 crc kubenswrapper[4789]: I1122 10:13:14.737637 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g2h6x\" (UniqueName: \"kubernetes.io/projected/7108f947-1001-4cb3-992f-7e6edf860db4-kube-api-access-g2h6x\") pod \"test-operator-logs-pod-ansibletest-ansibletest-ansibletest\" (UID: \"7108f947-1001-4cb3-992f-7e6edf860db4\") " pod="openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest" Nov 22 10:13:14 crc kubenswrapper[4789]: I1122 10:13:14.755804 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"test-operator-logs-pod-ansibletest-ansibletest-ansibletest\" (UID: \"7108f947-1001-4cb3-992f-7e6edf860db4\") " pod="openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest" Nov 22 10:13:14 crc kubenswrapper[4789]: I1122 10:13:14.828278 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest" Nov 22 10:13:15 crc kubenswrapper[4789]: I1122 10:13:15.301858 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest"] Nov 22 10:13:16 crc kubenswrapper[4789]: I1122 10:13:16.212469 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest" event={"ID":"7108f947-1001-4cb3-992f-7e6edf860db4","Type":"ContainerStarted","Data":"7ba620f3c8983a099974cd997afe93874b806286edd0fa7f50c4be3f8d314036"} Nov 22 10:13:16 crc kubenswrapper[4789]: I1122 10:13:16.212842 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest" event={"ID":"7108f947-1001-4cb3-992f-7e6edf860db4","Type":"ContainerStarted","Data":"bb19675eecbba7722d0df4fbdf50da6a6c41da79a80506ee0311c5b9d914b119"} Nov 22 10:13:16 crc kubenswrapper[4789]: I1122 10:13:16.237122 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest" podStartSLOduration=1.796188897 podStartE2EDuration="2.237099717s" podCreationTimestamp="2025-11-22 10:13:14 +0000 UTC" firstStartedPulling="2025-11-22 10:13:15.311158988 +0000 UTC m=+8209.545559271" lastFinishedPulling="2025-11-22 10:13:15.752069808 +0000 UTC m=+8209.986470091" observedRunningTime="2025-11-22 10:13:16.227122049 +0000 UTC m=+8210.461522372" watchObservedRunningTime="2025-11-22 10:13:16.237099717 +0000 UTC m=+8210.471499990" Nov 22 10:13:25 crc kubenswrapper[4789]: I1122 10:13:25.098831 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-gxzxz"] Nov 22 10:13:25 crc kubenswrapper[4789]: I1122 10:13:25.102935 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gxzxz" Nov 22 10:13:25 crc kubenswrapper[4789]: I1122 10:13:25.110005 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gxzxz"] Nov 22 10:13:25 crc kubenswrapper[4789]: I1122 10:13:25.238350 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6740b49c-a49b-4560-a75e-d5ba0c804885-utilities\") pod \"community-operators-gxzxz\" (UID: \"6740b49c-a49b-4560-a75e-d5ba0c804885\") " pod="openshift-marketplace/community-operators-gxzxz" Nov 22 10:13:25 crc kubenswrapper[4789]: I1122 10:13:25.238479 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5trht\" (UniqueName: \"kubernetes.io/projected/6740b49c-a49b-4560-a75e-d5ba0c804885-kube-api-access-5trht\") pod \"community-operators-gxzxz\" (UID: \"6740b49c-a49b-4560-a75e-d5ba0c804885\") " pod="openshift-marketplace/community-operators-gxzxz" Nov 22 10:13:25 crc kubenswrapper[4789]: I1122 10:13:25.238654 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6740b49c-a49b-4560-a75e-d5ba0c804885-catalog-content\") pod \"community-operators-gxzxz\" (UID: \"6740b49c-a49b-4560-a75e-d5ba0c804885\") " pod="openshift-marketplace/community-operators-gxzxz" Nov 22 10:13:25 crc kubenswrapper[4789]: I1122 10:13:25.340240 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6740b49c-a49b-4560-a75e-d5ba0c804885-catalog-content\") pod \"community-operators-gxzxz\" (UID: \"6740b49c-a49b-4560-a75e-d5ba0c804885\") " pod="openshift-marketplace/community-operators-gxzxz" Nov 22 10:13:25 crc kubenswrapper[4789]: I1122 10:13:25.340362 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6740b49c-a49b-4560-a75e-d5ba0c804885-utilities\") pod \"community-operators-gxzxz\" (UID: \"6740b49c-a49b-4560-a75e-d5ba0c804885\") " pod="openshift-marketplace/community-operators-gxzxz" Nov 22 10:13:25 crc kubenswrapper[4789]: I1122 10:13:25.340423 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5trht\" (UniqueName: \"kubernetes.io/projected/6740b49c-a49b-4560-a75e-d5ba0c804885-kube-api-access-5trht\") pod \"community-operators-gxzxz\" (UID: \"6740b49c-a49b-4560-a75e-d5ba0c804885\") " pod="openshift-marketplace/community-operators-gxzxz" Nov 22 10:13:25 crc kubenswrapper[4789]: I1122 10:13:25.341310 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6740b49c-a49b-4560-a75e-d5ba0c804885-catalog-content\") pod \"community-operators-gxzxz\" (UID: \"6740b49c-a49b-4560-a75e-d5ba0c804885\") " pod="openshift-marketplace/community-operators-gxzxz" Nov 22 10:13:25 crc kubenswrapper[4789]: I1122 10:13:25.341382 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6740b49c-a49b-4560-a75e-d5ba0c804885-utilities\") pod \"community-operators-gxzxz\" (UID: \"6740b49c-a49b-4560-a75e-d5ba0c804885\") " pod="openshift-marketplace/community-operators-gxzxz" Nov 22 10:13:25 crc kubenswrapper[4789]: I1122 10:13:25.359280 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5trht\" (UniqueName: \"kubernetes.io/projected/6740b49c-a49b-4560-a75e-d5ba0c804885-kube-api-access-5trht\") pod \"community-operators-gxzxz\" (UID: \"6740b49c-a49b-4560-a75e-d5ba0c804885\") " pod="openshift-marketplace/community-operators-gxzxz" Nov 22 10:13:25 crc kubenswrapper[4789]: I1122 10:13:25.440098 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gxzxz" Nov 22 10:13:25 crc kubenswrapper[4789]: I1122 10:13:25.951209 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gxzxz"] Nov 22 10:13:26 crc kubenswrapper[4789]: I1122 10:13:26.328243 4789 generic.go:334] "Generic (PLEG): container finished" podID="6740b49c-a49b-4560-a75e-d5ba0c804885" containerID="760ee3d0341a97d6f4953997b1fd92530875b3c0d70afbfc2b34b6ad8fc2d323" exitCode=0 Nov 22 10:13:26 crc kubenswrapper[4789]: I1122 10:13:26.328284 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gxzxz" event={"ID":"6740b49c-a49b-4560-a75e-d5ba0c804885","Type":"ContainerDied","Data":"760ee3d0341a97d6f4953997b1fd92530875b3c0d70afbfc2b34b6ad8fc2d323"} Nov 22 10:13:26 crc kubenswrapper[4789]: I1122 10:13:26.328307 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gxzxz" event={"ID":"6740b49c-a49b-4560-a75e-d5ba0c804885","Type":"ContainerStarted","Data":"d5b615f94639c9af5b6481e928c46d3f56afaa4853dd71741a193a6f6657fb69"} Nov 22 10:13:27 crc kubenswrapper[4789]: I1122 10:13:27.340890 4789 generic.go:334] "Generic (PLEG): container finished" podID="6740b49c-a49b-4560-a75e-d5ba0c804885" containerID="606b5374f5ecbef34c618e2378e6c692df99d21850806c0d9ac1b99e66c63368" exitCode=0 Nov 22 10:13:27 crc kubenswrapper[4789]: I1122 10:13:27.340955 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gxzxz" event={"ID":"6740b49c-a49b-4560-a75e-d5ba0c804885","Type":"ContainerDied","Data":"606b5374f5ecbef34c618e2378e6c692df99d21850806c0d9ac1b99e66c63368"} Nov 22 10:13:28 crc kubenswrapper[4789]: I1122 10:13:28.352273 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gxzxz" event={"ID":"6740b49c-a49b-4560-a75e-d5ba0c804885","Type":"ContainerStarted","Data":"209b92c1c1ca55d0e6dcef7c8a49e1e199fdf89fb15558a3089ccd86a42c6a52"} Nov 22 10:13:28 crc kubenswrapper[4789]: I1122 10:13:28.376315 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-gxzxz" podStartSLOduration=1.862924654 podStartE2EDuration="3.376298041s" podCreationTimestamp="2025-11-22 10:13:25 +0000 UTC" firstStartedPulling="2025-11-22 10:13:26.330257951 +0000 UTC m=+8220.564658224" lastFinishedPulling="2025-11-22 10:13:27.843631338 +0000 UTC m=+8222.078031611" observedRunningTime="2025-11-22 10:13:28.370194636 +0000 UTC m=+8222.604594919" watchObservedRunningTime="2025-11-22 10:13:28.376298041 +0000 UTC m=+8222.610698314" Nov 22 10:13:28 crc kubenswrapper[4789]: I1122 10:13:28.927504 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizontest-tests-horizontest"] Nov 22 10:13:28 crc kubenswrapper[4789]: I1122 10:13:28.929430 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizontest-tests-horizontest" Nov 22 10:13:28 crc kubenswrapper[4789]: W1122 10:13:28.932281 4789 reflector.go:561] object-"openstack"/"horizontest-tests-horizontesthorizontest-config": failed to list *v1.ConfigMap: configmaps "horizontest-tests-horizontesthorizontest-config" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openstack": no relationship found between node 'crc' and this object Nov 22 10:13:28 crc kubenswrapper[4789]: E1122 10:13:28.932325 4789 reflector.go:158] "Unhandled Error" err="object-\"openstack\"/\"horizontest-tests-horizontesthorizontest-config\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"horizontest-tests-horizontesthorizontest-config\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openstack\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 22 10:13:28 crc kubenswrapper[4789]: I1122 10:13:28.933429 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"test-operator-clouds-config" Nov 22 10:13:28 crc kubenswrapper[4789]: I1122 10:13:28.944474 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizontest-tests-horizontest"] Nov 22 10:13:29 crc kubenswrapper[4789]: I1122 10:13:29.119751 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341-ceph\") pod \"horizontest-tests-horizontest\" (UID: \"e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341\") " pod="openstack/horizontest-tests-horizontest" Nov 22 10:13:29 crc kubenswrapper[4789]: I1122 10:13:29.119888 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341-ca-certs\") pod \"horizontest-tests-horizontest\" (UID: \"e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341\") " pod="openstack/horizontest-tests-horizontest" Nov 22 10:13:29 crc kubenswrapper[4789]: I1122 10:13:29.120037 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341-openstack-config-secret\") pod \"horizontest-tests-horizontest\" (UID: \"e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341\") " pod="openstack/horizontest-tests-horizontest" Nov 22 10:13:29 crc kubenswrapper[4789]: I1122 10:13:29.120109 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341-test-operator-clouds-config\") pod \"horizontest-tests-horizontest\" (UID: \"e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341\") " pod="openstack/horizontest-tests-horizontest" Nov 22 10:13:29 crc kubenswrapper[4789]: I1122 10:13:29.120138 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"horizontest-tests-horizontest\" (UID: \"e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341\") " pod="openstack/horizontest-tests-horizontest" Nov 22 10:13:29 crc kubenswrapper[4789]: I1122 10:13:29.120245 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341-test-operator-ephemeral-temporary\") pod \"horizontest-tests-horizontest\" (UID: \"e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341\") " pod="openstack/horizontest-tests-horizontest" Nov 22 10:13:29 crc kubenswrapper[4789]: I1122 10:13:29.120287 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341-test-operator-ephemeral-workdir\") pod \"horizontest-tests-horizontest\" (UID: \"e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341\") " pod="openstack/horizontest-tests-horizontest" Nov 22 10:13:29 crc kubenswrapper[4789]: I1122 10:13:29.120307 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wsbcr\" (UniqueName: \"kubernetes.io/projected/e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341-kube-api-access-wsbcr\") pod \"horizontest-tests-horizontest\" (UID: \"e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341\") " pod="openstack/horizontest-tests-horizontest" Nov 22 10:13:29 crc kubenswrapper[4789]: I1122 10:13:29.222107 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341-test-operator-ephemeral-temporary\") pod \"horizontest-tests-horizontest\" (UID: \"e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341\") " pod="openstack/horizontest-tests-horizontest" Nov 22 10:13:29 crc kubenswrapper[4789]: I1122 10:13:29.222170 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341-test-operator-ephemeral-workdir\") pod \"horizontest-tests-horizontest\" (UID: \"e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341\") " pod="openstack/horizontest-tests-horizontest" Nov 22 10:13:29 crc kubenswrapper[4789]: I1122 10:13:29.222193 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wsbcr\" (UniqueName: \"kubernetes.io/projected/e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341-kube-api-access-wsbcr\") pod \"horizontest-tests-horizontest\" (UID: \"e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341\") " pod="openstack/horizontest-tests-horizontest" Nov 22 10:13:29 crc kubenswrapper[4789]: I1122 10:13:29.222243 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341-ceph\") pod \"horizontest-tests-horizontest\" (UID: \"e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341\") " pod="openstack/horizontest-tests-horizontest" Nov 22 10:13:29 crc kubenswrapper[4789]: I1122 10:13:29.222261 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341-ca-certs\") pod \"horizontest-tests-horizontest\" (UID: \"e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341\") " pod="openstack/horizontest-tests-horizontest" Nov 22 10:13:29 crc kubenswrapper[4789]: I1122 10:13:29.222329 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341-openstack-config-secret\") pod \"horizontest-tests-horizontest\" (UID: \"e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341\") " pod="openstack/horizontest-tests-horizontest" Nov 22 10:13:29 crc kubenswrapper[4789]: I1122 10:13:29.222365 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341-test-operator-clouds-config\") pod \"horizontest-tests-horizontest\" (UID: \"e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341\") " pod="openstack/horizontest-tests-horizontest" Nov 22 10:13:29 crc kubenswrapper[4789]: I1122 10:13:29.222383 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"horizontest-tests-horizontest\" (UID: \"e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341\") " pod="openstack/horizontest-tests-horizontest" Nov 22 10:13:29 crc kubenswrapper[4789]: I1122 10:13:29.222827 4789 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"horizontest-tests-horizontest\" (UID: \"e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/horizontest-tests-horizontest" Nov 22 10:13:29 crc kubenswrapper[4789]: I1122 10:13:29.223036 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341-test-operator-ephemeral-temporary\") pod \"horizontest-tests-horizontest\" (UID: \"e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341\") " pod="openstack/horizontest-tests-horizontest" Nov 22 10:13:29 crc kubenswrapper[4789]: I1122 10:13:29.223651 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341-test-operator-ephemeral-workdir\") pod \"horizontest-tests-horizontest\" (UID: \"e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341\") " pod="openstack/horizontest-tests-horizontest" Nov 22 10:13:29 crc kubenswrapper[4789]: I1122 10:13:29.224500 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341-test-operator-clouds-config\") pod \"horizontest-tests-horizontest\" (UID: \"e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341\") " pod="openstack/horizontest-tests-horizontest" Nov 22 10:13:29 crc kubenswrapper[4789]: I1122 10:13:29.230450 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341-ca-certs\") pod \"horizontest-tests-horizontest\" (UID: \"e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341\") " pod="openstack/horizontest-tests-horizontest" Nov 22 10:13:29 crc kubenswrapper[4789]: I1122 10:13:29.230461 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341-ceph\") pod \"horizontest-tests-horizontest\" (UID: \"e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341\") " pod="openstack/horizontest-tests-horizontest" Nov 22 10:13:29 crc kubenswrapper[4789]: I1122 10:13:29.230728 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341-openstack-config-secret\") pod \"horizontest-tests-horizontest\" (UID: \"e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341\") " pod="openstack/horizontest-tests-horizontest" Nov 22 10:13:29 crc kubenswrapper[4789]: I1122 10:13:29.244184 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wsbcr\" (UniqueName: \"kubernetes.io/projected/e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341-kube-api-access-wsbcr\") pod \"horizontest-tests-horizontest\" (UID: \"e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341\") " pod="openstack/horizontest-tests-horizontest" Nov 22 10:13:29 crc kubenswrapper[4789]: I1122 10:13:29.254791 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"horizontest-tests-horizontest\" (UID: \"e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341\") " pod="openstack/horizontest-tests-horizontest" Nov 22 10:13:29 crc kubenswrapper[4789]: I1122 10:13:29.305928 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizontest-tests-horizontest" Nov 22 10:13:29 crc kubenswrapper[4789]: I1122 10:13:29.804821 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizontest-tests-horizontest"] Nov 22 10:13:29 crc kubenswrapper[4789]: W1122 10:13:29.812953 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode0ed751e_8e37_4e1f_9aa6_7fb4d9b31341.slice/crio-870eb0e97c92bceacfda69643478b287df4b4f0955f59fd00e4340f690a59a98 WatchSource:0}: Error finding container 870eb0e97c92bceacfda69643478b287df4b4f0955f59fd00e4340f690a59a98: Status 404 returned error can't find the container with id 870eb0e97c92bceacfda69643478b287df4b4f0955f59fd00e4340f690a59a98 Nov 22 10:13:30 crc kubenswrapper[4789]: I1122 10:13:30.125311 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizontest-tests-horizontesthorizontest-config" Nov 22 10:13:30 crc kubenswrapper[4789]: I1122 10:13:30.377109 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizontest-tests-horizontest" event={"ID":"e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341","Type":"ContainerStarted","Data":"870eb0e97c92bceacfda69643478b287df4b4f0955f59fd00e4340f690a59a98"} Nov 22 10:13:35 crc kubenswrapper[4789]: I1122 10:13:35.440702 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-gxzxz" Nov 22 10:13:35 crc kubenswrapper[4789]: I1122 10:13:35.441214 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-gxzxz" Nov 22 10:13:35 crc kubenswrapper[4789]: I1122 10:13:35.489603 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-gxzxz" Nov 22 10:13:36 crc kubenswrapper[4789]: I1122 10:13:36.520236 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-gxzxz" Nov 22 10:13:36 crc kubenswrapper[4789]: I1122 10:13:36.586828 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gxzxz"] Nov 22 10:13:38 crc kubenswrapper[4789]: I1122 10:13:38.458432 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-gxzxz" podUID="6740b49c-a49b-4560-a75e-d5ba0c804885" containerName="registry-server" containerID="cri-o://209b92c1c1ca55d0e6dcef7c8a49e1e199fdf89fb15558a3089ccd86a42c6a52" gracePeriod=2 Nov 22 10:13:39 crc kubenswrapper[4789]: I1122 10:13:39.471017 4789 generic.go:334] "Generic (PLEG): container finished" podID="6740b49c-a49b-4560-a75e-d5ba0c804885" containerID="209b92c1c1ca55d0e6dcef7c8a49e1e199fdf89fb15558a3089ccd86a42c6a52" exitCode=0 Nov 22 10:13:39 crc kubenswrapper[4789]: I1122 10:13:39.473435 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gxzxz" event={"ID":"6740b49c-a49b-4560-a75e-d5ba0c804885","Type":"ContainerDied","Data":"209b92c1c1ca55d0e6dcef7c8a49e1e199fdf89fb15558a3089ccd86a42c6a52"} Nov 22 10:13:45 crc kubenswrapper[4789]: E1122 10:13:45.441394 4789 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 209b92c1c1ca55d0e6dcef7c8a49e1e199fdf89fb15558a3089ccd86a42c6a52 is running failed: container process not found" containerID="209b92c1c1ca55d0e6dcef7c8a49e1e199fdf89fb15558a3089ccd86a42c6a52" cmd=["grpc_health_probe","-addr=:50051"] Nov 22 10:13:45 crc kubenswrapper[4789]: E1122 10:13:45.442229 4789 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 209b92c1c1ca55d0e6dcef7c8a49e1e199fdf89fb15558a3089ccd86a42c6a52 is running failed: container process not found" containerID="209b92c1c1ca55d0e6dcef7c8a49e1e199fdf89fb15558a3089ccd86a42c6a52" cmd=["grpc_health_probe","-addr=:50051"] Nov 22 10:13:45 crc kubenswrapper[4789]: E1122 10:13:45.442634 4789 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 209b92c1c1ca55d0e6dcef7c8a49e1e199fdf89fb15558a3089ccd86a42c6a52 is running failed: container process not found" containerID="209b92c1c1ca55d0e6dcef7c8a49e1e199fdf89fb15558a3089ccd86a42c6a52" cmd=["grpc_health_probe","-addr=:50051"] Nov 22 10:13:45 crc kubenswrapper[4789]: E1122 10:13:45.442656 4789 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 209b92c1c1ca55d0e6dcef7c8a49e1e199fdf89fb15558a3089ccd86a42c6a52 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/community-operators-gxzxz" podUID="6740b49c-a49b-4560-a75e-d5ba0c804885" containerName="registry-server" Nov 22 10:13:46 crc kubenswrapper[4789]: E1122 10:13:46.320868 4789 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizontest:current-podified" Nov 22 10:13:46 crc kubenswrapper[4789]: E1122 10:13:46.321362 4789 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizontest-tests-horizontest,Image:quay.io/podified-antelope-centos9/openstack-horizontest:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:ADMIN_PASSWORD,Value:12345678,ValueFrom:nil,},EnvVar{Name:ADMIN_USERNAME,Value:admin,ValueFrom:nil,},EnvVar{Name:AUTH_URL,Value:https://keystone-public-openstack.apps-crc.testing,ValueFrom:nil,},EnvVar{Name:DASHBOARD_URL,Value:https://horizon-openstack.apps-crc.testing/,ValueFrom:nil,},EnvVar{Name:EXTRA_FLAG,Value:not pagination and test_users.py,ValueFrom:nil,},EnvVar{Name:FLAVOR_NAME,Value:m1.tiny,ValueFrom:nil,},EnvVar{Name:HORIZONTEST_DEBUG_MODE,Value:false,ValueFrom:nil,},EnvVar{Name:HORIZON_KEYS_FOLDER,Value:/etc/test_operator,ValueFrom:nil,},EnvVar{Name:HORIZON_LOGS_DIR_NAME,Value:horizon,ValueFrom:nil,},EnvVar{Name:HORIZON_REPO_BRANCH,Value:master,ValueFrom:nil,},EnvVar{Name:IMAGE_FILE,Value:/var/lib/horizontest/cirros-0.6.2-x86_64-disk.img,ValueFrom:nil,},EnvVar{Name:IMAGE_FILE_NAME,Value:cirros-0.6.2-x86_64-disk,ValueFrom:nil,},EnvVar{Name:IMAGE_URL,Value:http://download.cirros-cloud.net/0.6.2/cirros-0.6.2-x86_64-disk.img,ValueFrom:nil,},EnvVar{Name:PASSWORD,Value:horizontest,ValueFrom:nil,},EnvVar{Name:PROJECT_NAME,Value:horizontest,ValueFrom:nil,},EnvVar{Name:PROJECT_NAME_XPATH,Value://*[@class=\"context-project\"]//ancestor::ul,ValueFrom:nil,},EnvVar{Name:REPO_URL,Value:https://review.opendev.org/openstack/horizon,ValueFrom:nil,},EnvVar{Name:USER_NAME,Value:horizontest,ValueFrom:nil,},EnvVar{Name:USE_EXTERNAL_FILES,Value:True,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{2 0} {} 2 DecimalSI},memory: {{4294967296 0} {} 4Gi BinarySI},},Requests:ResourceList{cpu: {{1 0} {} 1 DecimalSI},memory: {{2147483648 0} {} 2Gi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/horizontest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/horizontest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-clouds-config,ReadOnly:true,MountPath:/var/lib/horizontest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-clouds-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ca-bundle.trust.crt,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ceph,ReadOnly:true,MountPath:/etc/ceph,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wsbcr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[NET_ADMIN NET_RAW],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42455,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42455,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizontest-tests-horizontest_openstack(e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 22 10:13:46 crc kubenswrapper[4789]: E1122 10:13:46.322983 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"horizontest-tests-horizontest\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/horizontest-tests-horizontest" podUID="e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341" Nov 22 10:13:46 crc kubenswrapper[4789]: E1122 10:13:46.540567 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"horizontest-tests-horizontest\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizontest:current-podified\\\"\"" pod="openstack/horizontest-tests-horizontest" podUID="e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341" Nov 22 10:13:46 crc kubenswrapper[4789]: I1122 10:13:46.680228 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gxzxz" Nov 22 10:13:46 crc kubenswrapper[4789]: I1122 10:13:46.800021 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6740b49c-a49b-4560-a75e-d5ba0c804885-catalog-content\") pod \"6740b49c-a49b-4560-a75e-d5ba0c804885\" (UID: \"6740b49c-a49b-4560-a75e-d5ba0c804885\") " Nov 22 10:13:46 crc kubenswrapper[4789]: I1122 10:13:46.800342 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6740b49c-a49b-4560-a75e-d5ba0c804885-utilities\") pod \"6740b49c-a49b-4560-a75e-d5ba0c804885\" (UID: \"6740b49c-a49b-4560-a75e-d5ba0c804885\") " Nov 22 10:13:46 crc kubenswrapper[4789]: I1122 10:13:46.800461 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5trht\" (UniqueName: \"kubernetes.io/projected/6740b49c-a49b-4560-a75e-d5ba0c804885-kube-api-access-5trht\") pod \"6740b49c-a49b-4560-a75e-d5ba0c804885\" (UID: \"6740b49c-a49b-4560-a75e-d5ba0c804885\") " Nov 22 10:13:46 crc kubenswrapper[4789]: I1122 10:13:46.801029 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6740b49c-a49b-4560-a75e-d5ba0c804885-utilities" (OuterVolumeSpecName: "utilities") pod "6740b49c-a49b-4560-a75e-d5ba0c804885" (UID: "6740b49c-a49b-4560-a75e-d5ba0c804885"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 10:13:46 crc kubenswrapper[4789]: I1122 10:13:46.809007 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6740b49c-a49b-4560-a75e-d5ba0c804885-kube-api-access-5trht" (OuterVolumeSpecName: "kube-api-access-5trht") pod "6740b49c-a49b-4560-a75e-d5ba0c804885" (UID: "6740b49c-a49b-4560-a75e-d5ba0c804885"). InnerVolumeSpecName "kube-api-access-5trht". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 10:13:46 crc kubenswrapper[4789]: I1122 10:13:46.858474 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6740b49c-a49b-4560-a75e-d5ba0c804885-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6740b49c-a49b-4560-a75e-d5ba0c804885" (UID: "6740b49c-a49b-4560-a75e-d5ba0c804885"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 10:13:46 crc kubenswrapper[4789]: I1122 10:13:46.902548 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6740b49c-a49b-4560-a75e-d5ba0c804885-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 10:13:46 crc kubenswrapper[4789]: I1122 10:13:46.902578 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6740b49c-a49b-4560-a75e-d5ba0c804885-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 10:13:46 crc kubenswrapper[4789]: I1122 10:13:46.902589 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5trht\" (UniqueName: \"kubernetes.io/projected/6740b49c-a49b-4560-a75e-d5ba0c804885-kube-api-access-5trht\") on node \"crc\" DevicePath \"\"" Nov 22 10:13:47 crc kubenswrapper[4789]: I1122 10:13:47.550346 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gxzxz" event={"ID":"6740b49c-a49b-4560-a75e-d5ba0c804885","Type":"ContainerDied","Data":"d5b615f94639c9af5b6481e928c46d3f56afaa4853dd71741a193a6f6657fb69"} Nov 22 10:13:47 crc kubenswrapper[4789]: I1122 10:13:47.550411 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gxzxz" Nov 22 10:13:47 crc kubenswrapper[4789]: I1122 10:13:47.550445 4789 scope.go:117] "RemoveContainer" containerID="209b92c1c1ca55d0e6dcef7c8a49e1e199fdf89fb15558a3089ccd86a42c6a52" Nov 22 10:13:47 crc kubenswrapper[4789]: I1122 10:13:47.594728 4789 scope.go:117] "RemoveContainer" containerID="606b5374f5ecbef34c618e2378e6c692df99d21850806c0d9ac1b99e66c63368" Nov 22 10:13:47 crc kubenswrapper[4789]: I1122 10:13:47.595967 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gxzxz"] Nov 22 10:13:47 crc kubenswrapper[4789]: I1122 10:13:47.607571 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-gxzxz"] Nov 22 10:13:47 crc kubenswrapper[4789]: I1122 10:13:47.621109 4789 scope.go:117] "RemoveContainer" containerID="760ee3d0341a97d6f4953997b1fd92530875b3c0d70afbfc2b34b6ad8fc2d323" Nov 22 10:13:47 crc kubenswrapper[4789]: I1122 10:13:47.986101 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6740b49c-a49b-4560-a75e-d5ba0c804885" path="/var/lib/kubelet/pods/6740b49c-a49b-4560-a75e-d5ba0c804885/volumes" Nov 22 10:14:02 crc kubenswrapper[4789]: I1122 10:14:02.722918 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizontest-tests-horizontest" event={"ID":"e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341","Type":"ContainerStarted","Data":"b9ff06bad945cdfbefbfa29f9d65499896a9a0965f5a5d3d04804b7eb0f798c4"} Nov 22 10:14:02 crc kubenswrapper[4789]: I1122 10:14:02.752931 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizontest-tests-horizontest" podStartSLOduration=3.680047994 podStartE2EDuration="35.75290962s" podCreationTimestamp="2025-11-22 10:13:27 +0000 UTC" firstStartedPulling="2025-11-22 10:13:29.816120786 +0000 UTC m=+8224.050521059" lastFinishedPulling="2025-11-22 10:14:01.888982402 +0000 UTC m=+8256.123382685" observedRunningTime="2025-11-22 10:14:02.739611361 +0000 UTC m=+8256.974011684" watchObservedRunningTime="2025-11-22 10:14:02.75290962 +0000 UTC m=+8256.987309913" Nov 22 10:14:05 crc kubenswrapper[4789]: I1122 10:14:05.372068 4789 patch_prober.go:28] interesting pod/machine-config-daemon-pxvrm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 10:14:05 crc kubenswrapper[4789]: I1122 10:14:05.372559 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 10:14:35 crc kubenswrapper[4789]: I1122 10:14:35.373283 4789 patch_prober.go:28] interesting pod/machine-config-daemon-pxvrm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 10:14:35 crc kubenswrapper[4789]: I1122 10:14:35.374063 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 10:15:00 crc kubenswrapper[4789]: I1122 10:15:00.164856 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396775-mqm9v"] Nov 22 10:15:00 crc kubenswrapper[4789]: E1122 10:15:00.165927 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6740b49c-a49b-4560-a75e-d5ba0c804885" containerName="extract-utilities" Nov 22 10:15:00 crc kubenswrapper[4789]: I1122 10:15:00.165944 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="6740b49c-a49b-4560-a75e-d5ba0c804885" containerName="extract-utilities" Nov 22 10:15:00 crc kubenswrapper[4789]: E1122 10:15:00.165957 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6740b49c-a49b-4560-a75e-d5ba0c804885" containerName="extract-content" Nov 22 10:15:00 crc kubenswrapper[4789]: I1122 10:15:00.165964 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="6740b49c-a49b-4560-a75e-d5ba0c804885" containerName="extract-content" Nov 22 10:15:00 crc kubenswrapper[4789]: E1122 10:15:00.165982 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6740b49c-a49b-4560-a75e-d5ba0c804885" containerName="registry-server" Nov 22 10:15:00 crc kubenswrapper[4789]: I1122 10:15:00.165991 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="6740b49c-a49b-4560-a75e-d5ba0c804885" containerName="registry-server" Nov 22 10:15:00 crc kubenswrapper[4789]: I1122 10:15:00.166235 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="6740b49c-a49b-4560-a75e-d5ba0c804885" containerName="registry-server" Nov 22 10:15:00 crc kubenswrapper[4789]: I1122 10:15:00.167117 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396775-mqm9v" Nov 22 10:15:00 crc kubenswrapper[4789]: I1122 10:15:00.169497 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 22 10:15:00 crc kubenswrapper[4789]: I1122 10:15:00.169674 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 22 10:15:00 crc kubenswrapper[4789]: I1122 10:15:00.194573 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396775-mqm9v"] Nov 22 10:15:00 crc kubenswrapper[4789]: I1122 10:15:00.291963 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bb9a1e9a-8883-4958-817c-95f891782743-config-volume\") pod \"collect-profiles-29396775-mqm9v\" (UID: \"bb9a1e9a-8883-4958-817c-95f891782743\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396775-mqm9v" Nov 22 10:15:00 crc kubenswrapper[4789]: I1122 10:15:00.292023 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-svlpg\" (UniqueName: \"kubernetes.io/projected/bb9a1e9a-8883-4958-817c-95f891782743-kube-api-access-svlpg\") pod \"collect-profiles-29396775-mqm9v\" (UID: \"bb9a1e9a-8883-4958-817c-95f891782743\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396775-mqm9v" Nov 22 10:15:00 crc kubenswrapper[4789]: I1122 10:15:00.292178 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bb9a1e9a-8883-4958-817c-95f891782743-secret-volume\") pod \"collect-profiles-29396775-mqm9v\" (UID: \"bb9a1e9a-8883-4958-817c-95f891782743\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396775-mqm9v" Nov 22 10:15:00 crc kubenswrapper[4789]: I1122 10:15:00.394357 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-svlpg\" (UniqueName: \"kubernetes.io/projected/bb9a1e9a-8883-4958-817c-95f891782743-kube-api-access-svlpg\") pod \"collect-profiles-29396775-mqm9v\" (UID: \"bb9a1e9a-8883-4958-817c-95f891782743\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396775-mqm9v" Nov 22 10:15:00 crc kubenswrapper[4789]: I1122 10:15:00.394595 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bb9a1e9a-8883-4958-817c-95f891782743-secret-volume\") pod \"collect-profiles-29396775-mqm9v\" (UID: \"bb9a1e9a-8883-4958-817c-95f891782743\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396775-mqm9v" Nov 22 10:15:00 crc kubenswrapper[4789]: I1122 10:15:00.394715 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bb9a1e9a-8883-4958-817c-95f891782743-config-volume\") pod \"collect-profiles-29396775-mqm9v\" (UID: \"bb9a1e9a-8883-4958-817c-95f891782743\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396775-mqm9v" Nov 22 10:15:00 crc kubenswrapper[4789]: I1122 10:15:00.395965 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bb9a1e9a-8883-4958-817c-95f891782743-config-volume\") pod \"collect-profiles-29396775-mqm9v\" (UID: \"bb9a1e9a-8883-4958-817c-95f891782743\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396775-mqm9v" Nov 22 10:15:00 crc kubenswrapper[4789]: I1122 10:15:00.410808 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bb9a1e9a-8883-4958-817c-95f891782743-secret-volume\") pod \"collect-profiles-29396775-mqm9v\" (UID: \"bb9a1e9a-8883-4958-817c-95f891782743\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396775-mqm9v" Nov 22 10:15:00 crc kubenswrapper[4789]: I1122 10:15:00.414856 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-svlpg\" (UniqueName: \"kubernetes.io/projected/bb9a1e9a-8883-4958-817c-95f891782743-kube-api-access-svlpg\") pod \"collect-profiles-29396775-mqm9v\" (UID: \"bb9a1e9a-8883-4958-817c-95f891782743\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396775-mqm9v" Nov 22 10:15:00 crc kubenswrapper[4789]: I1122 10:15:00.493741 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396775-mqm9v" Nov 22 10:15:00 crc kubenswrapper[4789]: I1122 10:15:00.957384 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396775-mqm9v"] Nov 22 10:15:01 crc kubenswrapper[4789]: I1122 10:15:01.339094 4789 generic.go:334] "Generic (PLEG): container finished" podID="bb9a1e9a-8883-4958-817c-95f891782743" containerID="94092649c274d1bc05d8a5812dfc99722354d3fdf36061450f73f2f5566d7042" exitCode=0 Nov 22 10:15:01 crc kubenswrapper[4789]: I1122 10:15:01.339163 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29396775-mqm9v" event={"ID":"bb9a1e9a-8883-4958-817c-95f891782743","Type":"ContainerDied","Data":"94092649c274d1bc05d8a5812dfc99722354d3fdf36061450f73f2f5566d7042"} Nov 22 10:15:01 crc kubenswrapper[4789]: I1122 10:15:01.339451 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29396775-mqm9v" event={"ID":"bb9a1e9a-8883-4958-817c-95f891782743","Type":"ContainerStarted","Data":"be98e0771f71bd69429b08da58eb9735786235b2ebca8d2c6601b135b15e8fc3"} Nov 22 10:15:02 crc kubenswrapper[4789]: I1122 10:15:02.752067 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396775-mqm9v" Nov 22 10:15:02 crc kubenswrapper[4789]: I1122 10:15:02.857212 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bb9a1e9a-8883-4958-817c-95f891782743-secret-volume\") pod \"bb9a1e9a-8883-4958-817c-95f891782743\" (UID: \"bb9a1e9a-8883-4958-817c-95f891782743\") " Nov 22 10:15:02 crc kubenswrapper[4789]: I1122 10:15:02.857399 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bb9a1e9a-8883-4958-817c-95f891782743-config-volume\") pod \"bb9a1e9a-8883-4958-817c-95f891782743\" (UID: \"bb9a1e9a-8883-4958-817c-95f891782743\") " Nov 22 10:15:02 crc kubenswrapper[4789]: I1122 10:15:02.857430 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-svlpg\" (UniqueName: \"kubernetes.io/projected/bb9a1e9a-8883-4958-817c-95f891782743-kube-api-access-svlpg\") pod \"bb9a1e9a-8883-4958-817c-95f891782743\" (UID: \"bb9a1e9a-8883-4958-817c-95f891782743\") " Nov 22 10:15:02 crc kubenswrapper[4789]: I1122 10:15:02.858587 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb9a1e9a-8883-4958-817c-95f891782743-config-volume" (OuterVolumeSpecName: "config-volume") pod "bb9a1e9a-8883-4958-817c-95f891782743" (UID: "bb9a1e9a-8883-4958-817c-95f891782743"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 10:15:02 crc kubenswrapper[4789]: I1122 10:15:02.864285 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb9a1e9a-8883-4958-817c-95f891782743-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "bb9a1e9a-8883-4958-817c-95f891782743" (UID: "bb9a1e9a-8883-4958-817c-95f891782743"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 10:15:02 crc kubenswrapper[4789]: I1122 10:15:02.867687 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb9a1e9a-8883-4958-817c-95f891782743-kube-api-access-svlpg" (OuterVolumeSpecName: "kube-api-access-svlpg") pod "bb9a1e9a-8883-4958-817c-95f891782743" (UID: "bb9a1e9a-8883-4958-817c-95f891782743"). InnerVolumeSpecName "kube-api-access-svlpg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 10:15:02 crc kubenswrapper[4789]: I1122 10:15:02.959445 4789 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bb9a1e9a-8883-4958-817c-95f891782743-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 22 10:15:02 crc kubenswrapper[4789]: I1122 10:15:02.959484 4789 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bb9a1e9a-8883-4958-817c-95f891782743-config-volume\") on node \"crc\" DevicePath \"\"" Nov 22 10:15:02 crc kubenswrapper[4789]: I1122 10:15:02.959495 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-svlpg\" (UniqueName: \"kubernetes.io/projected/bb9a1e9a-8883-4958-817c-95f891782743-kube-api-access-svlpg\") on node \"crc\" DevicePath \"\"" Nov 22 10:15:03 crc kubenswrapper[4789]: I1122 10:15:03.361966 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29396775-mqm9v" event={"ID":"bb9a1e9a-8883-4958-817c-95f891782743","Type":"ContainerDied","Data":"be98e0771f71bd69429b08da58eb9735786235b2ebca8d2c6601b135b15e8fc3"} Nov 22 10:15:03 crc kubenswrapper[4789]: I1122 10:15:03.362004 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="be98e0771f71bd69429b08da58eb9735786235b2ebca8d2c6601b135b15e8fc3" Nov 22 10:15:03 crc kubenswrapper[4789]: I1122 10:15:03.362021 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396775-mqm9v" Nov 22 10:15:03 crc kubenswrapper[4789]: I1122 10:15:03.851717 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396730-t8nxp"] Nov 22 10:15:03 crc kubenswrapper[4789]: I1122 10:15:03.863216 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396730-t8nxp"] Nov 22 10:15:03 crc kubenswrapper[4789]: I1122 10:15:03.978880 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2bf75c3b-b28f-4205-b47b-eb2b7bd7c70e" path="/var/lib/kubelet/pods/2bf75c3b-b28f-4205-b47b-eb2b7bd7c70e/volumes" Nov 22 10:15:05 crc kubenswrapper[4789]: I1122 10:15:05.372328 4789 patch_prober.go:28] interesting pod/machine-config-daemon-pxvrm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 10:15:05 crc kubenswrapper[4789]: I1122 10:15:05.372378 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 10:15:05 crc kubenswrapper[4789]: I1122 10:15:05.372422 4789 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" Nov 22 10:15:05 crc kubenswrapper[4789]: I1122 10:15:05.373187 4789 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4141db46cc7225c6dcef7ae55b4ce0014526d33bf629421689170714974fa6ca"} pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 22 10:15:05 crc kubenswrapper[4789]: I1122 10:15:05.373243 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" containerID="cri-o://4141db46cc7225c6dcef7ae55b4ce0014526d33bf629421689170714974fa6ca" gracePeriod=600 Nov 22 10:15:06 crc kubenswrapper[4789]: I1122 10:15:06.393904 4789 generic.go:334] "Generic (PLEG): container finished" podID="58aa8071-7ecc-4692-8789-2db04ec70647" containerID="4141db46cc7225c6dcef7ae55b4ce0014526d33bf629421689170714974fa6ca" exitCode=0 Nov 22 10:15:06 crc kubenswrapper[4789]: I1122 10:15:06.394038 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" event={"ID":"58aa8071-7ecc-4692-8789-2db04ec70647","Type":"ContainerDied","Data":"4141db46cc7225c6dcef7ae55b4ce0014526d33bf629421689170714974fa6ca"} Nov 22 10:15:06 crc kubenswrapper[4789]: I1122 10:15:06.394542 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" event={"ID":"58aa8071-7ecc-4692-8789-2db04ec70647","Type":"ContainerStarted","Data":"9e4b9858fc97b3d62c7338b8fc7e6fb3cc0a7ad892259a591c7f692014f9eac6"} Nov 22 10:15:06 crc kubenswrapper[4789]: I1122 10:15:06.394585 4789 scope.go:117] "RemoveContainer" containerID="6e22a5a8fc3638ca6aab7b67ea550d9b289803b56baeb39c04dfbc919861ab85" Nov 22 10:15:30 crc kubenswrapper[4789]: I1122 10:15:30.751925 4789 scope.go:117] "RemoveContainer" containerID="2dfa9b64332408b98d2c174d7d17576e468a64aec10ff9d95e11b756a14fa78d" Nov 22 10:17:05 crc kubenswrapper[4789]: I1122 10:17:05.372650 4789 patch_prober.go:28] interesting pod/machine-config-daemon-pxvrm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 10:17:05 crc kubenswrapper[4789]: I1122 10:17:05.373402 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 10:17:30 crc kubenswrapper[4789]: I1122 10:17:30.827328 4789 scope.go:117] "RemoveContainer" containerID="4bb35c394c8244a7be16b7fe5b58d5238d82c8344904aab0f1804d11e6c8f23f" Nov 22 10:17:30 crc kubenswrapper[4789]: I1122 10:17:30.864607 4789 scope.go:117] "RemoveContainer" containerID="525bd0e43b0003e77fd18701aa69e2ea1ca8c97ddb980fa320eb30ff29909929" Nov 22 10:17:30 crc kubenswrapper[4789]: I1122 10:17:30.908035 4789 scope.go:117] "RemoveContainer" containerID="1bd58955a009ba1fffcca50d1d46b18159086db27eba53983f073293804af87f" Nov 22 10:17:35 crc kubenswrapper[4789]: I1122 10:17:35.372481 4789 patch_prober.go:28] interesting pod/machine-config-daemon-pxvrm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 10:17:35 crc kubenswrapper[4789]: I1122 10:17:35.373161 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 10:18:05 crc kubenswrapper[4789]: I1122 10:18:05.372743 4789 patch_prober.go:28] interesting pod/machine-config-daemon-pxvrm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 10:18:05 crc kubenswrapper[4789]: I1122 10:18:05.373285 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 10:18:05 crc kubenswrapper[4789]: I1122 10:18:05.373342 4789 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" Nov 22 10:18:05 crc kubenswrapper[4789]: I1122 10:18:05.374217 4789 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9e4b9858fc97b3d62c7338b8fc7e6fb3cc0a7ad892259a591c7f692014f9eac6"} pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 22 10:18:05 crc kubenswrapper[4789]: I1122 10:18:05.374282 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" containerID="cri-o://9e4b9858fc97b3d62c7338b8fc7e6fb3cc0a7ad892259a591c7f692014f9eac6" gracePeriod=600 Nov 22 10:18:05 crc kubenswrapper[4789]: E1122 10:18:05.508827 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 10:18:06 crc kubenswrapper[4789]: I1122 10:18:06.498702 4789 generic.go:334] "Generic (PLEG): container finished" podID="58aa8071-7ecc-4692-8789-2db04ec70647" containerID="9e4b9858fc97b3d62c7338b8fc7e6fb3cc0a7ad892259a591c7f692014f9eac6" exitCode=0 Nov 22 10:18:06 crc kubenswrapper[4789]: I1122 10:18:06.498815 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" event={"ID":"58aa8071-7ecc-4692-8789-2db04ec70647","Type":"ContainerDied","Data":"9e4b9858fc97b3d62c7338b8fc7e6fb3cc0a7ad892259a591c7f692014f9eac6"} Nov 22 10:18:06 crc kubenswrapper[4789]: I1122 10:18:06.499278 4789 scope.go:117] "RemoveContainer" containerID="4141db46cc7225c6dcef7ae55b4ce0014526d33bf629421689170714974fa6ca" Nov 22 10:18:06 crc kubenswrapper[4789]: I1122 10:18:06.500773 4789 scope.go:117] "RemoveContainer" containerID="9e4b9858fc97b3d62c7338b8fc7e6fb3cc0a7ad892259a591c7f692014f9eac6" Nov 22 10:18:06 crc kubenswrapper[4789]: E1122 10:18:06.501785 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 10:18:10 crc kubenswrapper[4789]: I1122 10:18:10.004683 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-xd8ss"] Nov 22 10:18:10 crc kubenswrapper[4789]: E1122 10:18:10.006449 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb9a1e9a-8883-4958-817c-95f891782743" containerName="collect-profiles" Nov 22 10:18:10 crc kubenswrapper[4789]: I1122 10:18:10.006482 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb9a1e9a-8883-4958-817c-95f891782743" containerName="collect-profiles" Nov 22 10:18:10 crc kubenswrapper[4789]: I1122 10:18:10.006944 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb9a1e9a-8883-4958-817c-95f891782743" containerName="collect-profiles" Nov 22 10:18:10 crc kubenswrapper[4789]: I1122 10:18:10.012382 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xd8ss"] Nov 22 10:18:10 crc kubenswrapper[4789]: I1122 10:18:10.012493 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xd8ss" Nov 22 10:18:10 crc kubenswrapper[4789]: I1122 10:18:10.129212 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ac48b63-d114-419a-ba2f-005c8a583370-catalog-content\") pod \"certified-operators-xd8ss\" (UID: \"4ac48b63-d114-419a-ba2f-005c8a583370\") " pod="openshift-marketplace/certified-operators-xd8ss" Nov 22 10:18:10 crc kubenswrapper[4789]: I1122 10:18:10.129399 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lm5xs\" (UniqueName: \"kubernetes.io/projected/4ac48b63-d114-419a-ba2f-005c8a583370-kube-api-access-lm5xs\") pod \"certified-operators-xd8ss\" (UID: \"4ac48b63-d114-419a-ba2f-005c8a583370\") " pod="openshift-marketplace/certified-operators-xd8ss" Nov 22 10:18:10 crc kubenswrapper[4789]: I1122 10:18:10.129874 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ac48b63-d114-419a-ba2f-005c8a583370-utilities\") pod \"certified-operators-xd8ss\" (UID: \"4ac48b63-d114-419a-ba2f-005c8a583370\") " pod="openshift-marketplace/certified-operators-xd8ss" Nov 22 10:18:10 crc kubenswrapper[4789]: I1122 10:18:10.231821 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lm5xs\" (UniqueName: \"kubernetes.io/projected/4ac48b63-d114-419a-ba2f-005c8a583370-kube-api-access-lm5xs\") pod \"certified-operators-xd8ss\" (UID: \"4ac48b63-d114-419a-ba2f-005c8a583370\") " pod="openshift-marketplace/certified-operators-xd8ss" Nov 22 10:18:10 crc kubenswrapper[4789]: I1122 10:18:10.231910 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ac48b63-d114-419a-ba2f-005c8a583370-utilities\") pod \"certified-operators-xd8ss\" (UID: \"4ac48b63-d114-419a-ba2f-005c8a583370\") " pod="openshift-marketplace/certified-operators-xd8ss" Nov 22 10:18:10 crc kubenswrapper[4789]: I1122 10:18:10.231998 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ac48b63-d114-419a-ba2f-005c8a583370-catalog-content\") pod \"certified-operators-xd8ss\" (UID: \"4ac48b63-d114-419a-ba2f-005c8a583370\") " pod="openshift-marketplace/certified-operators-xd8ss" Nov 22 10:18:10 crc kubenswrapper[4789]: I1122 10:18:10.232451 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ac48b63-d114-419a-ba2f-005c8a583370-catalog-content\") pod \"certified-operators-xd8ss\" (UID: \"4ac48b63-d114-419a-ba2f-005c8a583370\") " pod="openshift-marketplace/certified-operators-xd8ss" Nov 22 10:18:10 crc kubenswrapper[4789]: I1122 10:18:10.232585 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ac48b63-d114-419a-ba2f-005c8a583370-utilities\") pod \"certified-operators-xd8ss\" (UID: \"4ac48b63-d114-419a-ba2f-005c8a583370\") " pod="openshift-marketplace/certified-operators-xd8ss" Nov 22 10:18:10 crc kubenswrapper[4789]: I1122 10:18:10.255980 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lm5xs\" (UniqueName: \"kubernetes.io/projected/4ac48b63-d114-419a-ba2f-005c8a583370-kube-api-access-lm5xs\") pod \"certified-operators-xd8ss\" (UID: \"4ac48b63-d114-419a-ba2f-005c8a583370\") " pod="openshift-marketplace/certified-operators-xd8ss" Nov 22 10:18:10 crc kubenswrapper[4789]: I1122 10:18:10.332220 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xd8ss" Nov 22 10:18:10 crc kubenswrapper[4789]: I1122 10:18:10.798158 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xd8ss"] Nov 22 10:18:11 crc kubenswrapper[4789]: I1122 10:18:11.559820 4789 generic.go:334] "Generic (PLEG): container finished" podID="4ac48b63-d114-419a-ba2f-005c8a583370" containerID="843dc4f17cb636fb0b6a21d102fdf6bfe1771afe2d9cae2cc0df2867a92868ef" exitCode=0 Nov 22 10:18:11 crc kubenswrapper[4789]: I1122 10:18:11.559873 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xd8ss" event={"ID":"4ac48b63-d114-419a-ba2f-005c8a583370","Type":"ContainerDied","Data":"843dc4f17cb636fb0b6a21d102fdf6bfe1771afe2d9cae2cc0df2867a92868ef"} Nov 22 10:18:11 crc kubenswrapper[4789]: I1122 10:18:11.559903 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xd8ss" event={"ID":"4ac48b63-d114-419a-ba2f-005c8a583370","Type":"ContainerStarted","Data":"41f41a7d9d1125b6a8dc56c784333e7c6f7fe2c18b599111ea209c00b38afe9e"} Nov 22 10:18:11 crc kubenswrapper[4789]: I1122 10:18:11.562006 4789 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 22 10:18:12 crc kubenswrapper[4789]: I1122 10:18:12.575155 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xd8ss" event={"ID":"4ac48b63-d114-419a-ba2f-005c8a583370","Type":"ContainerStarted","Data":"6bdc3225f94ab3377e15fc4af24b9acd8e790fc4bd4314229709090d4a10896f"} Nov 22 10:18:13 crc kubenswrapper[4789]: I1122 10:18:13.590491 4789 generic.go:334] "Generic (PLEG): container finished" podID="4ac48b63-d114-419a-ba2f-005c8a583370" containerID="6bdc3225f94ab3377e15fc4af24b9acd8e790fc4bd4314229709090d4a10896f" exitCode=0 Nov 22 10:18:13 crc kubenswrapper[4789]: I1122 10:18:13.590562 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xd8ss" event={"ID":"4ac48b63-d114-419a-ba2f-005c8a583370","Type":"ContainerDied","Data":"6bdc3225f94ab3377e15fc4af24b9acd8e790fc4bd4314229709090d4a10896f"} Nov 22 10:18:13 crc kubenswrapper[4789]: I1122 10:18:13.749451 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-9hr7p"] Nov 22 10:18:13 crc kubenswrapper[4789]: I1122 10:18:13.753505 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9hr7p" Nov 22 10:18:13 crc kubenswrapper[4789]: I1122 10:18:13.777254 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9hr7p"] Nov 22 10:18:13 crc kubenswrapper[4789]: I1122 10:18:13.805472 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1202c7f8-4cc4-40e0-96a0-5f5de25cf845-catalog-content\") pod \"redhat-marketplace-9hr7p\" (UID: \"1202c7f8-4cc4-40e0-96a0-5f5de25cf845\") " pod="openshift-marketplace/redhat-marketplace-9hr7p" Nov 22 10:18:13 crc kubenswrapper[4789]: I1122 10:18:13.805551 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rwvbx\" (UniqueName: \"kubernetes.io/projected/1202c7f8-4cc4-40e0-96a0-5f5de25cf845-kube-api-access-rwvbx\") pod \"redhat-marketplace-9hr7p\" (UID: \"1202c7f8-4cc4-40e0-96a0-5f5de25cf845\") " pod="openshift-marketplace/redhat-marketplace-9hr7p" Nov 22 10:18:13 crc kubenswrapper[4789]: I1122 10:18:13.805805 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1202c7f8-4cc4-40e0-96a0-5f5de25cf845-utilities\") pod \"redhat-marketplace-9hr7p\" (UID: \"1202c7f8-4cc4-40e0-96a0-5f5de25cf845\") " pod="openshift-marketplace/redhat-marketplace-9hr7p" Nov 22 10:18:13 crc kubenswrapper[4789]: I1122 10:18:13.908696 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1202c7f8-4cc4-40e0-96a0-5f5de25cf845-catalog-content\") pod \"redhat-marketplace-9hr7p\" (UID: \"1202c7f8-4cc4-40e0-96a0-5f5de25cf845\") " pod="openshift-marketplace/redhat-marketplace-9hr7p" Nov 22 10:18:13 crc kubenswrapper[4789]: I1122 10:18:13.908816 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rwvbx\" (UniqueName: \"kubernetes.io/projected/1202c7f8-4cc4-40e0-96a0-5f5de25cf845-kube-api-access-rwvbx\") pod \"redhat-marketplace-9hr7p\" (UID: \"1202c7f8-4cc4-40e0-96a0-5f5de25cf845\") " pod="openshift-marketplace/redhat-marketplace-9hr7p" Nov 22 10:18:13 crc kubenswrapper[4789]: I1122 10:18:13.908929 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1202c7f8-4cc4-40e0-96a0-5f5de25cf845-utilities\") pod \"redhat-marketplace-9hr7p\" (UID: \"1202c7f8-4cc4-40e0-96a0-5f5de25cf845\") " pod="openshift-marketplace/redhat-marketplace-9hr7p" Nov 22 10:18:13 crc kubenswrapper[4789]: I1122 10:18:13.909468 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1202c7f8-4cc4-40e0-96a0-5f5de25cf845-catalog-content\") pod \"redhat-marketplace-9hr7p\" (UID: \"1202c7f8-4cc4-40e0-96a0-5f5de25cf845\") " pod="openshift-marketplace/redhat-marketplace-9hr7p" Nov 22 10:18:13 crc kubenswrapper[4789]: I1122 10:18:13.909726 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1202c7f8-4cc4-40e0-96a0-5f5de25cf845-utilities\") pod \"redhat-marketplace-9hr7p\" (UID: \"1202c7f8-4cc4-40e0-96a0-5f5de25cf845\") " pod="openshift-marketplace/redhat-marketplace-9hr7p" Nov 22 10:18:13 crc kubenswrapper[4789]: I1122 10:18:13.948619 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rwvbx\" (UniqueName: \"kubernetes.io/projected/1202c7f8-4cc4-40e0-96a0-5f5de25cf845-kube-api-access-rwvbx\") pod \"redhat-marketplace-9hr7p\" (UID: \"1202c7f8-4cc4-40e0-96a0-5f5de25cf845\") " pod="openshift-marketplace/redhat-marketplace-9hr7p" Nov 22 10:18:14 crc kubenswrapper[4789]: I1122 10:18:14.084606 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9hr7p" Nov 22 10:18:14 crc kubenswrapper[4789]: I1122 10:18:14.569289 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9hr7p"] Nov 22 10:18:14 crc kubenswrapper[4789]: I1122 10:18:14.610686 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9hr7p" event={"ID":"1202c7f8-4cc4-40e0-96a0-5f5de25cf845","Type":"ContainerStarted","Data":"5a8cdb31b78c72b5e08cbe7de231f958daf30f6e33d57da0f0bb1c030ceb5e84"} Nov 22 10:18:14 crc kubenswrapper[4789]: I1122 10:18:14.613543 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xd8ss" event={"ID":"4ac48b63-d114-419a-ba2f-005c8a583370","Type":"ContainerStarted","Data":"945a58ab2ab48663f0a9dc900f1739b3b5e6760572db2c65ae0d8c31f9348adb"} Nov 22 10:18:14 crc kubenswrapper[4789]: I1122 10:18:14.629617 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-xd8ss" podStartSLOduration=3.193107688 podStartE2EDuration="5.629581458s" podCreationTimestamp="2025-11-22 10:18:09 +0000 UTC" firstStartedPulling="2025-11-22 10:18:11.561649494 +0000 UTC m=+8505.796049777" lastFinishedPulling="2025-11-22 10:18:13.998123234 +0000 UTC m=+8508.232523547" observedRunningTime="2025-11-22 10:18:14.628794657 +0000 UTC m=+8508.863194940" watchObservedRunningTime="2025-11-22 10:18:14.629581458 +0000 UTC m=+8508.863981781" Nov 22 10:18:15 crc kubenswrapper[4789]: I1122 10:18:15.627637 4789 generic.go:334] "Generic (PLEG): container finished" podID="1202c7f8-4cc4-40e0-96a0-5f5de25cf845" containerID="101275ce5171256cf6c68dba6abc0a5c2e8a8cec3d2c58050e5487dbac57ede4" exitCode=0 Nov 22 10:18:15 crc kubenswrapper[4789]: I1122 10:18:15.627697 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9hr7p" event={"ID":"1202c7f8-4cc4-40e0-96a0-5f5de25cf845","Type":"ContainerDied","Data":"101275ce5171256cf6c68dba6abc0a5c2e8a8cec3d2c58050e5487dbac57ede4"} Nov 22 10:18:17 crc kubenswrapper[4789]: I1122 10:18:17.657178 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9hr7p" event={"ID":"1202c7f8-4cc4-40e0-96a0-5f5de25cf845","Type":"ContainerStarted","Data":"7bd8b37682dc0fac7d0b761c0781c66b49eb6c6d03a243c0089b5daa95464459"} Nov 22 10:18:18 crc kubenswrapper[4789]: I1122 10:18:18.667314 4789 generic.go:334] "Generic (PLEG): container finished" podID="1202c7f8-4cc4-40e0-96a0-5f5de25cf845" containerID="7bd8b37682dc0fac7d0b761c0781c66b49eb6c6d03a243c0089b5daa95464459" exitCode=0 Nov 22 10:18:18 crc kubenswrapper[4789]: I1122 10:18:18.667439 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9hr7p" event={"ID":"1202c7f8-4cc4-40e0-96a0-5f5de25cf845","Type":"ContainerDied","Data":"7bd8b37682dc0fac7d0b761c0781c66b49eb6c6d03a243c0089b5daa95464459"} Nov 22 10:18:18 crc kubenswrapper[4789]: I1122 10:18:18.667672 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9hr7p" event={"ID":"1202c7f8-4cc4-40e0-96a0-5f5de25cf845","Type":"ContainerStarted","Data":"10c8163d8635796bc734a57819a0aa93c6d1ec410e74a2f3563e4872ed0ab1da"} Nov 22 10:18:18 crc kubenswrapper[4789]: I1122 10:18:18.703744 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-9hr7p" podStartSLOduration=3.234233964 podStartE2EDuration="5.703716253s" podCreationTimestamp="2025-11-22 10:18:13 +0000 UTC" firstStartedPulling="2025-11-22 10:18:15.630169188 +0000 UTC m=+8509.864569471" lastFinishedPulling="2025-11-22 10:18:18.099651487 +0000 UTC m=+8512.334051760" observedRunningTime="2025-11-22 10:18:18.689897031 +0000 UTC m=+8512.924297324" watchObservedRunningTime="2025-11-22 10:18:18.703716253 +0000 UTC m=+8512.938116566" Nov 22 10:18:19 crc kubenswrapper[4789]: I1122 10:18:19.965793 4789 scope.go:117] "RemoveContainer" containerID="9e4b9858fc97b3d62c7338b8fc7e6fb3cc0a7ad892259a591c7f692014f9eac6" Nov 22 10:18:19 crc kubenswrapper[4789]: E1122 10:18:19.966065 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 10:18:20 crc kubenswrapper[4789]: I1122 10:18:20.333289 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-xd8ss" Nov 22 10:18:20 crc kubenswrapper[4789]: I1122 10:18:20.333704 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-xd8ss" Nov 22 10:18:20 crc kubenswrapper[4789]: I1122 10:18:20.399008 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-xd8ss" Nov 22 10:18:20 crc kubenswrapper[4789]: I1122 10:18:20.762877 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-xd8ss" Nov 22 10:18:21 crc kubenswrapper[4789]: I1122 10:18:21.933097 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xd8ss"] Nov 22 10:18:23 crc kubenswrapper[4789]: I1122 10:18:23.719082 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-xd8ss" podUID="4ac48b63-d114-419a-ba2f-005c8a583370" containerName="registry-server" containerID="cri-o://945a58ab2ab48663f0a9dc900f1739b3b5e6760572db2c65ae0d8c31f9348adb" gracePeriod=2 Nov 22 10:18:24 crc kubenswrapper[4789]: I1122 10:18:24.085696 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-9hr7p" Nov 22 10:18:24 crc kubenswrapper[4789]: I1122 10:18:24.086020 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-9hr7p" Nov 22 10:18:24 crc kubenswrapper[4789]: I1122 10:18:24.163094 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-9hr7p" Nov 22 10:18:24 crc kubenswrapper[4789]: I1122 10:18:24.260380 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xd8ss" Nov 22 10:18:24 crc kubenswrapper[4789]: I1122 10:18:24.332987 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ac48b63-d114-419a-ba2f-005c8a583370-utilities\") pod \"4ac48b63-d114-419a-ba2f-005c8a583370\" (UID: \"4ac48b63-d114-419a-ba2f-005c8a583370\") " Nov 22 10:18:24 crc kubenswrapper[4789]: I1122 10:18:24.333186 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ac48b63-d114-419a-ba2f-005c8a583370-catalog-content\") pod \"4ac48b63-d114-419a-ba2f-005c8a583370\" (UID: \"4ac48b63-d114-419a-ba2f-005c8a583370\") " Nov 22 10:18:24 crc kubenswrapper[4789]: I1122 10:18:24.333366 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lm5xs\" (UniqueName: \"kubernetes.io/projected/4ac48b63-d114-419a-ba2f-005c8a583370-kube-api-access-lm5xs\") pod \"4ac48b63-d114-419a-ba2f-005c8a583370\" (UID: \"4ac48b63-d114-419a-ba2f-005c8a583370\") " Nov 22 10:18:24 crc kubenswrapper[4789]: I1122 10:18:24.334295 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4ac48b63-d114-419a-ba2f-005c8a583370-utilities" (OuterVolumeSpecName: "utilities") pod "4ac48b63-d114-419a-ba2f-005c8a583370" (UID: "4ac48b63-d114-419a-ba2f-005c8a583370"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 10:18:24 crc kubenswrapper[4789]: I1122 10:18:24.343229 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ac48b63-d114-419a-ba2f-005c8a583370-kube-api-access-lm5xs" (OuterVolumeSpecName: "kube-api-access-lm5xs") pod "4ac48b63-d114-419a-ba2f-005c8a583370" (UID: "4ac48b63-d114-419a-ba2f-005c8a583370"). InnerVolumeSpecName "kube-api-access-lm5xs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 10:18:24 crc kubenswrapper[4789]: I1122 10:18:24.390971 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4ac48b63-d114-419a-ba2f-005c8a583370-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4ac48b63-d114-419a-ba2f-005c8a583370" (UID: "4ac48b63-d114-419a-ba2f-005c8a583370"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 10:18:24 crc kubenswrapper[4789]: I1122 10:18:24.436284 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ac48b63-d114-419a-ba2f-005c8a583370-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 10:18:24 crc kubenswrapper[4789]: I1122 10:18:24.436329 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ac48b63-d114-419a-ba2f-005c8a583370-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 10:18:24 crc kubenswrapper[4789]: I1122 10:18:24.436349 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lm5xs\" (UniqueName: \"kubernetes.io/projected/4ac48b63-d114-419a-ba2f-005c8a583370-kube-api-access-lm5xs\") on node \"crc\" DevicePath \"\"" Nov 22 10:18:24 crc kubenswrapper[4789]: I1122 10:18:24.736568 4789 generic.go:334] "Generic (PLEG): container finished" podID="4ac48b63-d114-419a-ba2f-005c8a583370" containerID="945a58ab2ab48663f0a9dc900f1739b3b5e6760572db2c65ae0d8c31f9348adb" exitCode=0 Nov 22 10:18:24 crc kubenswrapper[4789]: I1122 10:18:24.736680 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xd8ss" event={"ID":"4ac48b63-d114-419a-ba2f-005c8a583370","Type":"ContainerDied","Data":"945a58ab2ab48663f0a9dc900f1739b3b5e6760572db2c65ae0d8c31f9348adb"} Nov 22 10:18:24 crc kubenswrapper[4789]: I1122 10:18:24.737105 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xd8ss" event={"ID":"4ac48b63-d114-419a-ba2f-005c8a583370","Type":"ContainerDied","Data":"41f41a7d9d1125b6a8dc56c784333e7c6f7fe2c18b599111ea209c00b38afe9e"} Nov 22 10:18:24 crc kubenswrapper[4789]: I1122 10:18:24.737159 4789 scope.go:117] "RemoveContainer" containerID="945a58ab2ab48663f0a9dc900f1739b3b5e6760572db2c65ae0d8c31f9348adb" Nov 22 10:18:24 crc kubenswrapper[4789]: I1122 10:18:24.736705 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xd8ss" Nov 22 10:18:24 crc kubenswrapper[4789]: I1122 10:18:24.799308 4789 scope.go:117] "RemoveContainer" containerID="6bdc3225f94ab3377e15fc4af24b9acd8e790fc4bd4314229709090d4a10896f" Nov 22 10:18:24 crc kubenswrapper[4789]: I1122 10:18:24.805878 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xd8ss"] Nov 22 10:18:24 crc kubenswrapper[4789]: I1122 10:18:24.829044 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-xd8ss"] Nov 22 10:18:24 crc kubenswrapper[4789]: I1122 10:18:24.831599 4789 scope.go:117] "RemoveContainer" containerID="843dc4f17cb636fb0b6a21d102fdf6bfe1771afe2d9cae2cc0df2867a92868ef" Nov 22 10:18:24 crc kubenswrapper[4789]: I1122 10:18:24.832122 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-9hr7p" Nov 22 10:18:24 crc kubenswrapper[4789]: I1122 10:18:24.884463 4789 scope.go:117] "RemoveContainer" containerID="945a58ab2ab48663f0a9dc900f1739b3b5e6760572db2c65ae0d8c31f9348adb" Nov 22 10:18:24 crc kubenswrapper[4789]: E1122 10:18:24.885284 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"945a58ab2ab48663f0a9dc900f1739b3b5e6760572db2c65ae0d8c31f9348adb\": container with ID starting with 945a58ab2ab48663f0a9dc900f1739b3b5e6760572db2c65ae0d8c31f9348adb not found: ID does not exist" containerID="945a58ab2ab48663f0a9dc900f1739b3b5e6760572db2c65ae0d8c31f9348adb" Nov 22 10:18:24 crc kubenswrapper[4789]: I1122 10:18:24.885340 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"945a58ab2ab48663f0a9dc900f1739b3b5e6760572db2c65ae0d8c31f9348adb"} err="failed to get container status \"945a58ab2ab48663f0a9dc900f1739b3b5e6760572db2c65ae0d8c31f9348adb\": rpc error: code = NotFound desc = could not find container \"945a58ab2ab48663f0a9dc900f1739b3b5e6760572db2c65ae0d8c31f9348adb\": container with ID starting with 945a58ab2ab48663f0a9dc900f1739b3b5e6760572db2c65ae0d8c31f9348adb not found: ID does not exist" Nov 22 10:18:24 crc kubenswrapper[4789]: I1122 10:18:24.885376 4789 scope.go:117] "RemoveContainer" containerID="6bdc3225f94ab3377e15fc4af24b9acd8e790fc4bd4314229709090d4a10896f" Nov 22 10:18:24 crc kubenswrapper[4789]: E1122 10:18:24.885828 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6bdc3225f94ab3377e15fc4af24b9acd8e790fc4bd4314229709090d4a10896f\": container with ID starting with 6bdc3225f94ab3377e15fc4af24b9acd8e790fc4bd4314229709090d4a10896f not found: ID does not exist" containerID="6bdc3225f94ab3377e15fc4af24b9acd8e790fc4bd4314229709090d4a10896f" Nov 22 10:18:24 crc kubenswrapper[4789]: I1122 10:18:24.885887 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6bdc3225f94ab3377e15fc4af24b9acd8e790fc4bd4314229709090d4a10896f"} err="failed to get container status \"6bdc3225f94ab3377e15fc4af24b9acd8e790fc4bd4314229709090d4a10896f\": rpc error: code = NotFound desc = could not find container \"6bdc3225f94ab3377e15fc4af24b9acd8e790fc4bd4314229709090d4a10896f\": container with ID starting with 6bdc3225f94ab3377e15fc4af24b9acd8e790fc4bd4314229709090d4a10896f not found: ID does not exist" Nov 22 10:18:24 crc kubenswrapper[4789]: I1122 10:18:24.885929 4789 scope.go:117] "RemoveContainer" containerID="843dc4f17cb636fb0b6a21d102fdf6bfe1771afe2d9cae2cc0df2867a92868ef" Nov 22 10:18:24 crc kubenswrapper[4789]: E1122 10:18:24.886285 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"843dc4f17cb636fb0b6a21d102fdf6bfe1771afe2d9cae2cc0df2867a92868ef\": container with ID starting with 843dc4f17cb636fb0b6a21d102fdf6bfe1771afe2d9cae2cc0df2867a92868ef not found: ID does not exist" containerID="843dc4f17cb636fb0b6a21d102fdf6bfe1771afe2d9cae2cc0df2867a92868ef" Nov 22 10:18:24 crc kubenswrapper[4789]: I1122 10:18:24.886317 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"843dc4f17cb636fb0b6a21d102fdf6bfe1771afe2d9cae2cc0df2867a92868ef"} err="failed to get container status \"843dc4f17cb636fb0b6a21d102fdf6bfe1771afe2d9cae2cc0df2867a92868ef\": rpc error: code = NotFound desc = could not find container \"843dc4f17cb636fb0b6a21d102fdf6bfe1771afe2d9cae2cc0df2867a92868ef\": container with ID starting with 843dc4f17cb636fb0b6a21d102fdf6bfe1771afe2d9cae2cc0df2867a92868ef not found: ID does not exist" Nov 22 10:18:25 crc kubenswrapper[4789]: I1122 10:18:25.989170 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ac48b63-d114-419a-ba2f-005c8a583370" path="/var/lib/kubelet/pods/4ac48b63-d114-419a-ba2f-005c8a583370/volumes" Nov 22 10:18:26 crc kubenswrapper[4789]: I1122 10:18:26.534429 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9hr7p"] Nov 22 10:18:26 crc kubenswrapper[4789]: I1122 10:18:26.779264 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-9hr7p" podUID="1202c7f8-4cc4-40e0-96a0-5f5de25cf845" containerName="registry-server" containerID="cri-o://10c8163d8635796bc734a57819a0aa93c6d1ec410e74a2f3563e4872ed0ab1da" gracePeriod=2 Nov 22 10:18:27 crc kubenswrapper[4789]: E1122 10:18:27.063216 4789 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1202c7f8_4cc4_40e0_96a0_5f5de25cf845.slice/crio-conmon-10c8163d8635796bc734a57819a0aa93c6d1ec410e74a2f3563e4872ed0ab1da.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1202c7f8_4cc4_40e0_96a0_5f5de25cf845.slice/crio-10c8163d8635796bc734a57819a0aa93c6d1ec410e74a2f3563e4872ed0ab1da.scope\": RecentStats: unable to find data in memory cache]" Nov 22 10:18:27 crc kubenswrapper[4789]: I1122 10:18:27.316917 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9hr7p" Nov 22 10:18:27 crc kubenswrapper[4789]: I1122 10:18:27.399572 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1202c7f8-4cc4-40e0-96a0-5f5de25cf845-catalog-content\") pod \"1202c7f8-4cc4-40e0-96a0-5f5de25cf845\" (UID: \"1202c7f8-4cc4-40e0-96a0-5f5de25cf845\") " Nov 22 10:18:27 crc kubenswrapper[4789]: I1122 10:18:27.399649 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1202c7f8-4cc4-40e0-96a0-5f5de25cf845-utilities\") pod \"1202c7f8-4cc4-40e0-96a0-5f5de25cf845\" (UID: \"1202c7f8-4cc4-40e0-96a0-5f5de25cf845\") " Nov 22 10:18:27 crc kubenswrapper[4789]: I1122 10:18:27.399810 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rwvbx\" (UniqueName: \"kubernetes.io/projected/1202c7f8-4cc4-40e0-96a0-5f5de25cf845-kube-api-access-rwvbx\") pod \"1202c7f8-4cc4-40e0-96a0-5f5de25cf845\" (UID: \"1202c7f8-4cc4-40e0-96a0-5f5de25cf845\") " Nov 22 10:18:27 crc kubenswrapper[4789]: I1122 10:18:27.400568 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1202c7f8-4cc4-40e0-96a0-5f5de25cf845-utilities" (OuterVolumeSpecName: "utilities") pod "1202c7f8-4cc4-40e0-96a0-5f5de25cf845" (UID: "1202c7f8-4cc4-40e0-96a0-5f5de25cf845"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 10:18:27 crc kubenswrapper[4789]: I1122 10:18:27.407015 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1202c7f8-4cc4-40e0-96a0-5f5de25cf845-kube-api-access-rwvbx" (OuterVolumeSpecName: "kube-api-access-rwvbx") pod "1202c7f8-4cc4-40e0-96a0-5f5de25cf845" (UID: "1202c7f8-4cc4-40e0-96a0-5f5de25cf845"). InnerVolumeSpecName "kube-api-access-rwvbx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 10:18:27 crc kubenswrapper[4789]: I1122 10:18:27.413996 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1202c7f8-4cc4-40e0-96a0-5f5de25cf845-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1202c7f8-4cc4-40e0-96a0-5f5de25cf845" (UID: "1202c7f8-4cc4-40e0-96a0-5f5de25cf845"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 10:18:27 crc kubenswrapper[4789]: I1122 10:18:27.501772 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1202c7f8-4cc4-40e0-96a0-5f5de25cf845-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 10:18:27 crc kubenswrapper[4789]: I1122 10:18:27.501800 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1202c7f8-4cc4-40e0-96a0-5f5de25cf845-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 10:18:27 crc kubenswrapper[4789]: I1122 10:18:27.501810 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rwvbx\" (UniqueName: \"kubernetes.io/projected/1202c7f8-4cc4-40e0-96a0-5f5de25cf845-kube-api-access-rwvbx\") on node \"crc\" DevicePath \"\"" Nov 22 10:18:27 crc kubenswrapper[4789]: I1122 10:18:27.796067 4789 generic.go:334] "Generic (PLEG): container finished" podID="1202c7f8-4cc4-40e0-96a0-5f5de25cf845" containerID="10c8163d8635796bc734a57819a0aa93c6d1ec410e74a2f3563e4872ed0ab1da" exitCode=0 Nov 22 10:18:27 crc kubenswrapper[4789]: I1122 10:18:27.796134 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9hr7p" event={"ID":"1202c7f8-4cc4-40e0-96a0-5f5de25cf845","Type":"ContainerDied","Data":"10c8163d8635796bc734a57819a0aa93c6d1ec410e74a2f3563e4872ed0ab1da"} Nov 22 10:18:27 crc kubenswrapper[4789]: I1122 10:18:27.796179 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9hr7p" Nov 22 10:18:27 crc kubenswrapper[4789]: I1122 10:18:27.796206 4789 scope.go:117] "RemoveContainer" containerID="10c8163d8635796bc734a57819a0aa93c6d1ec410e74a2f3563e4872ed0ab1da" Nov 22 10:18:27 crc kubenswrapper[4789]: I1122 10:18:27.796192 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9hr7p" event={"ID":"1202c7f8-4cc4-40e0-96a0-5f5de25cf845","Type":"ContainerDied","Data":"5a8cdb31b78c72b5e08cbe7de231f958daf30f6e33d57da0f0bb1c030ceb5e84"} Nov 22 10:18:27 crc kubenswrapper[4789]: I1122 10:18:27.835457 4789 scope.go:117] "RemoveContainer" containerID="7bd8b37682dc0fac7d0b761c0781c66b49eb6c6d03a243c0089b5daa95464459" Nov 22 10:18:27 crc kubenswrapper[4789]: I1122 10:18:27.845070 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9hr7p"] Nov 22 10:18:27 crc kubenswrapper[4789]: I1122 10:18:27.854576 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-9hr7p"] Nov 22 10:18:27 crc kubenswrapper[4789]: I1122 10:18:27.867965 4789 scope.go:117] "RemoveContainer" containerID="101275ce5171256cf6c68dba6abc0a5c2e8a8cec3d2c58050e5487dbac57ede4" Nov 22 10:18:27 crc kubenswrapper[4789]: I1122 10:18:27.906863 4789 scope.go:117] "RemoveContainer" containerID="10c8163d8635796bc734a57819a0aa93c6d1ec410e74a2f3563e4872ed0ab1da" Nov 22 10:18:27 crc kubenswrapper[4789]: E1122 10:18:27.907447 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"10c8163d8635796bc734a57819a0aa93c6d1ec410e74a2f3563e4872ed0ab1da\": container with ID starting with 10c8163d8635796bc734a57819a0aa93c6d1ec410e74a2f3563e4872ed0ab1da not found: ID does not exist" containerID="10c8163d8635796bc734a57819a0aa93c6d1ec410e74a2f3563e4872ed0ab1da" Nov 22 10:18:27 crc kubenswrapper[4789]: I1122 10:18:27.907504 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"10c8163d8635796bc734a57819a0aa93c6d1ec410e74a2f3563e4872ed0ab1da"} err="failed to get container status \"10c8163d8635796bc734a57819a0aa93c6d1ec410e74a2f3563e4872ed0ab1da\": rpc error: code = NotFound desc = could not find container \"10c8163d8635796bc734a57819a0aa93c6d1ec410e74a2f3563e4872ed0ab1da\": container with ID starting with 10c8163d8635796bc734a57819a0aa93c6d1ec410e74a2f3563e4872ed0ab1da not found: ID does not exist" Nov 22 10:18:27 crc kubenswrapper[4789]: I1122 10:18:27.907538 4789 scope.go:117] "RemoveContainer" containerID="7bd8b37682dc0fac7d0b761c0781c66b49eb6c6d03a243c0089b5daa95464459" Nov 22 10:18:27 crc kubenswrapper[4789]: E1122 10:18:27.907978 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7bd8b37682dc0fac7d0b761c0781c66b49eb6c6d03a243c0089b5daa95464459\": container with ID starting with 7bd8b37682dc0fac7d0b761c0781c66b49eb6c6d03a243c0089b5daa95464459 not found: ID does not exist" containerID="7bd8b37682dc0fac7d0b761c0781c66b49eb6c6d03a243c0089b5daa95464459" Nov 22 10:18:27 crc kubenswrapper[4789]: I1122 10:18:27.908114 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7bd8b37682dc0fac7d0b761c0781c66b49eb6c6d03a243c0089b5daa95464459"} err="failed to get container status \"7bd8b37682dc0fac7d0b761c0781c66b49eb6c6d03a243c0089b5daa95464459\": rpc error: code = NotFound desc = could not find container \"7bd8b37682dc0fac7d0b761c0781c66b49eb6c6d03a243c0089b5daa95464459\": container with ID starting with 7bd8b37682dc0fac7d0b761c0781c66b49eb6c6d03a243c0089b5daa95464459 not found: ID does not exist" Nov 22 10:18:27 crc kubenswrapper[4789]: I1122 10:18:27.908153 4789 scope.go:117] "RemoveContainer" containerID="101275ce5171256cf6c68dba6abc0a5c2e8a8cec3d2c58050e5487dbac57ede4" Nov 22 10:18:27 crc kubenswrapper[4789]: E1122 10:18:27.908501 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"101275ce5171256cf6c68dba6abc0a5c2e8a8cec3d2c58050e5487dbac57ede4\": container with ID starting with 101275ce5171256cf6c68dba6abc0a5c2e8a8cec3d2c58050e5487dbac57ede4 not found: ID does not exist" containerID="101275ce5171256cf6c68dba6abc0a5c2e8a8cec3d2c58050e5487dbac57ede4" Nov 22 10:18:27 crc kubenswrapper[4789]: I1122 10:18:27.908592 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"101275ce5171256cf6c68dba6abc0a5c2e8a8cec3d2c58050e5487dbac57ede4"} err="failed to get container status \"101275ce5171256cf6c68dba6abc0a5c2e8a8cec3d2c58050e5487dbac57ede4\": rpc error: code = NotFound desc = could not find container \"101275ce5171256cf6c68dba6abc0a5c2e8a8cec3d2c58050e5487dbac57ede4\": container with ID starting with 101275ce5171256cf6c68dba6abc0a5c2e8a8cec3d2c58050e5487dbac57ede4 not found: ID does not exist" Nov 22 10:18:27 crc kubenswrapper[4789]: I1122 10:18:27.983790 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1202c7f8-4cc4-40e0-96a0-5f5de25cf845" path="/var/lib/kubelet/pods/1202c7f8-4cc4-40e0-96a0-5f5de25cf845/volumes" Nov 22 10:18:32 crc kubenswrapper[4789]: I1122 10:18:32.966316 4789 scope.go:117] "RemoveContainer" containerID="9e4b9858fc97b3d62c7338b8fc7e6fb3cc0a7ad892259a591c7f692014f9eac6" Nov 22 10:18:32 crc kubenswrapper[4789]: E1122 10:18:32.967854 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 10:18:37 crc kubenswrapper[4789]: E1122 10:18:37.359985 4789 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1202c7f8_4cc4_40e0_96a0_5f5de25cf845.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1202c7f8_4cc4_40e0_96a0_5f5de25cf845.slice/crio-5a8cdb31b78c72b5e08cbe7de231f958daf30f6e33d57da0f0bb1c030ceb5e84\": RecentStats: unable to find data in memory cache]" Nov 22 10:18:44 crc kubenswrapper[4789]: I1122 10:18:44.966055 4789 scope.go:117] "RemoveContainer" containerID="9e4b9858fc97b3d62c7338b8fc7e6fb3cc0a7ad892259a591c7f692014f9eac6" Nov 22 10:18:44 crc kubenswrapper[4789]: E1122 10:18:44.967227 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 10:18:47 crc kubenswrapper[4789]: E1122 10:18:47.591920 4789 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1202c7f8_4cc4_40e0_96a0_5f5de25cf845.slice/crio-5a8cdb31b78c72b5e08cbe7de231f958daf30f6e33d57da0f0bb1c030ceb5e84\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1202c7f8_4cc4_40e0_96a0_5f5de25cf845.slice\": RecentStats: unable to find data in memory cache]" Nov 22 10:18:55 crc kubenswrapper[4789]: I1122 10:18:55.965554 4789 scope.go:117] "RemoveContainer" containerID="9e4b9858fc97b3d62c7338b8fc7e6fb3cc0a7ad892259a591c7f692014f9eac6" Nov 22 10:18:55 crc kubenswrapper[4789]: E1122 10:18:55.966498 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 10:18:57 crc kubenswrapper[4789]: E1122 10:18:57.826839 4789 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1202c7f8_4cc4_40e0_96a0_5f5de25cf845.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1202c7f8_4cc4_40e0_96a0_5f5de25cf845.slice/crio-5a8cdb31b78c72b5e08cbe7de231f958daf30f6e33d57da0f0bb1c030ceb5e84\": RecentStats: unable to find data in memory cache]" Nov 22 10:19:08 crc kubenswrapper[4789]: E1122 10:19:08.136545 4789 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1202c7f8_4cc4_40e0_96a0_5f5de25cf845.slice/crio-5a8cdb31b78c72b5e08cbe7de231f958daf30f6e33d57da0f0bb1c030ceb5e84\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1202c7f8_4cc4_40e0_96a0_5f5de25cf845.slice\": RecentStats: unable to find data in memory cache]" Nov 22 10:19:08 crc kubenswrapper[4789]: I1122 10:19:08.966494 4789 scope.go:117] "RemoveContainer" containerID="9e4b9858fc97b3d62c7338b8fc7e6fb3cc0a7ad892259a591c7f692014f9eac6" Nov 22 10:19:08 crc kubenswrapper[4789]: E1122 10:19:08.967154 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 10:19:18 crc kubenswrapper[4789]: E1122 10:19:18.417344 4789 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1202c7f8_4cc4_40e0_96a0_5f5de25cf845.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1202c7f8_4cc4_40e0_96a0_5f5de25cf845.slice/crio-5a8cdb31b78c72b5e08cbe7de231f958daf30f6e33d57da0f0bb1c030ceb5e84\": RecentStats: unable to find data in memory cache]" Nov 22 10:19:23 crc kubenswrapper[4789]: I1122 10:19:23.965934 4789 scope.go:117] "RemoveContainer" containerID="9e4b9858fc97b3d62c7338b8fc7e6fb3cc0a7ad892259a591c7f692014f9eac6" Nov 22 10:19:23 crc kubenswrapper[4789]: E1122 10:19:23.967262 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 10:19:37 crc kubenswrapper[4789]: I1122 10:19:37.977981 4789 scope.go:117] "RemoveContainer" containerID="9e4b9858fc97b3d62c7338b8fc7e6fb3cc0a7ad892259a591c7f692014f9eac6" Nov 22 10:19:37 crc kubenswrapper[4789]: E1122 10:19:37.979051 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 10:19:52 crc kubenswrapper[4789]: I1122 10:19:52.965546 4789 scope.go:117] "RemoveContainer" containerID="9e4b9858fc97b3d62c7338b8fc7e6fb3cc0a7ad892259a591c7f692014f9eac6" Nov 22 10:19:52 crc kubenswrapper[4789]: E1122 10:19:52.966531 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 10:20:06 crc kubenswrapper[4789]: I1122 10:20:06.967479 4789 scope.go:117] "RemoveContainer" containerID="9e4b9858fc97b3d62c7338b8fc7e6fb3cc0a7ad892259a591c7f692014f9eac6" Nov 22 10:20:06 crc kubenswrapper[4789]: E1122 10:20:06.968932 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 10:20:18 crc kubenswrapper[4789]: I1122 10:20:18.966687 4789 scope.go:117] "RemoveContainer" containerID="9e4b9858fc97b3d62c7338b8fc7e6fb3cc0a7ad892259a591c7f692014f9eac6" Nov 22 10:20:18 crc kubenswrapper[4789]: E1122 10:20:18.968392 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 10:20:32 crc kubenswrapper[4789]: I1122 10:20:32.965580 4789 scope.go:117] "RemoveContainer" containerID="9e4b9858fc97b3d62c7338b8fc7e6fb3cc0a7ad892259a591c7f692014f9eac6" Nov 22 10:20:32 crc kubenswrapper[4789]: E1122 10:20:32.967087 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 10:20:47 crc kubenswrapper[4789]: I1122 10:20:47.988275 4789 scope.go:117] "RemoveContainer" containerID="9e4b9858fc97b3d62c7338b8fc7e6fb3cc0a7ad892259a591c7f692014f9eac6" Nov 22 10:20:47 crc kubenswrapper[4789]: E1122 10:20:47.989325 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 10:20:54 crc kubenswrapper[4789]: I1122 10:20:54.563236 4789 generic.go:334] "Generic (PLEG): container finished" podID="e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341" containerID="b9ff06bad945cdfbefbfa29f9d65499896a9a0965f5a5d3d04804b7eb0f798c4" exitCode=0 Nov 22 10:20:54 crc kubenswrapper[4789]: I1122 10:20:54.563375 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizontest-tests-horizontest" event={"ID":"e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341","Type":"ContainerDied","Data":"b9ff06bad945cdfbefbfa29f9d65499896a9a0965f5a5d3d04804b7eb0f798c4"} Nov 22 10:20:55 crc kubenswrapper[4789]: I1122 10:20:55.969858 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizontest-tests-horizontest" Nov 22 10:20:56 crc kubenswrapper[4789]: I1122 10:20:56.081665 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341-test-operator-clouds-config\") pod \"e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341\" (UID: \"e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341\") " Nov 22 10:20:56 crc kubenswrapper[4789]: I1122 10:20:56.081733 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wsbcr\" (UniqueName: \"kubernetes.io/projected/e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341-kube-api-access-wsbcr\") pod \"e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341\" (UID: \"e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341\") " Nov 22 10:20:56 crc kubenswrapper[4789]: I1122 10:20:56.081793 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341-test-operator-ephemeral-temporary\") pod \"e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341\" (UID: \"e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341\") " Nov 22 10:20:56 crc kubenswrapper[4789]: I1122 10:20:56.081887 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341-test-operator-ephemeral-workdir\") pod \"e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341\" (UID: \"e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341\") " Nov 22 10:20:56 crc kubenswrapper[4789]: I1122 10:20:56.081909 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341\" (UID: \"e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341\") " Nov 22 10:20:56 crc kubenswrapper[4789]: I1122 10:20:56.081955 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341-ceph\") pod \"e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341\" (UID: \"e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341\") " Nov 22 10:20:56 crc kubenswrapper[4789]: I1122 10:20:56.082051 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341-openstack-config-secret\") pod \"e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341\" (UID: \"e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341\") " Nov 22 10:20:56 crc kubenswrapper[4789]: I1122 10:20:56.082171 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341-ca-certs\") pod \"e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341\" (UID: \"e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341\") " Nov 22 10:20:56 crc kubenswrapper[4789]: I1122 10:20:56.082587 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341" (UID: "e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 10:20:56 crc kubenswrapper[4789]: I1122 10:20:56.082722 4789 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Nov 22 10:20:56 crc kubenswrapper[4789]: I1122 10:20:56.088215 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "test-operator-logs") pod "e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341" (UID: "e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 22 10:20:56 crc kubenswrapper[4789]: I1122 10:20:56.088964 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341-ceph" (OuterVolumeSpecName: "ceph") pod "e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341" (UID: "e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 10:20:56 crc kubenswrapper[4789]: I1122 10:20:56.099386 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341-kube-api-access-wsbcr" (OuterVolumeSpecName: "kube-api-access-wsbcr") pod "e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341" (UID: "e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341"). InnerVolumeSpecName "kube-api-access-wsbcr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 10:20:56 crc kubenswrapper[4789]: I1122 10:20:56.114783 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341" (UID: "e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 10:20:56 crc kubenswrapper[4789]: I1122 10:20:56.157267 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341" (UID: "e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 10:20:56 crc kubenswrapper[4789]: I1122 10:20:56.175818 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341-test-operator-clouds-config" (OuterVolumeSpecName: "test-operator-clouds-config") pod "e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341" (UID: "e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341"). InnerVolumeSpecName "test-operator-clouds-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 10:20:56 crc kubenswrapper[4789]: I1122 10:20:56.185074 4789 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Nov 22 10:20:56 crc kubenswrapper[4789]: I1122 10:20:56.185244 4789 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341-ca-certs\") on node \"crc\" DevicePath \"\"" Nov 22 10:20:56 crc kubenswrapper[4789]: I1122 10:20:56.185441 4789 reconciler_common.go:293] "Volume detached for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341-test-operator-clouds-config\") on node \"crc\" DevicePath \"\"" Nov 22 10:20:56 crc kubenswrapper[4789]: I1122 10:20:56.185584 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wsbcr\" (UniqueName: \"kubernetes.io/projected/e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341-kube-api-access-wsbcr\") on node \"crc\" DevicePath \"\"" Nov 22 10:20:56 crc kubenswrapper[4789]: I1122 10:20:56.185731 4789 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Nov 22 10:20:56 crc kubenswrapper[4789]: I1122 10:20:56.185878 4789 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341-ceph\") on node \"crc\" DevicePath \"\"" Nov 22 10:20:56 crc kubenswrapper[4789]: I1122 10:20:56.215595 4789 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Nov 22 10:20:56 crc kubenswrapper[4789]: I1122 10:20:56.289729 4789 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Nov 22 10:20:56 crc kubenswrapper[4789]: I1122 10:20:56.347223 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341" (UID: "e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 10:20:56 crc kubenswrapper[4789]: I1122 10:20:56.392155 4789 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Nov 22 10:20:56 crc kubenswrapper[4789]: I1122 10:20:56.589601 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizontest-tests-horizontest" event={"ID":"e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341","Type":"ContainerDied","Data":"870eb0e97c92bceacfda69643478b287df4b4f0955f59fd00e4340f690a59a98"} Nov 22 10:20:56 crc kubenswrapper[4789]: I1122 10:20:56.589664 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="870eb0e97c92bceacfda69643478b287df4b4f0955f59fd00e4340f690a59a98" Nov 22 10:20:56 crc kubenswrapper[4789]: I1122 10:20:56.589850 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizontest-tests-horizontest" Nov 22 10:21:02 crc kubenswrapper[4789]: I1122 10:21:02.213370 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest"] Nov 22 10:21:02 crc kubenswrapper[4789]: E1122 10:21:02.214585 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1202c7f8-4cc4-40e0-96a0-5f5de25cf845" containerName="registry-server" Nov 22 10:21:02 crc kubenswrapper[4789]: I1122 10:21:02.214614 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="1202c7f8-4cc4-40e0-96a0-5f5de25cf845" containerName="registry-server" Nov 22 10:21:02 crc kubenswrapper[4789]: E1122 10:21:02.214691 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ac48b63-d114-419a-ba2f-005c8a583370" containerName="extract-utilities" Nov 22 10:21:02 crc kubenswrapper[4789]: I1122 10:21:02.214705 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ac48b63-d114-419a-ba2f-005c8a583370" containerName="extract-utilities" Nov 22 10:21:02 crc kubenswrapper[4789]: E1122 10:21:02.214814 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ac48b63-d114-419a-ba2f-005c8a583370" containerName="registry-server" Nov 22 10:21:02 crc kubenswrapper[4789]: I1122 10:21:02.214832 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ac48b63-d114-419a-ba2f-005c8a583370" containerName="registry-server" Nov 22 10:21:02 crc kubenswrapper[4789]: E1122 10:21:02.214902 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341" containerName="horizontest-tests-horizontest" Nov 22 10:21:02 crc kubenswrapper[4789]: I1122 10:21:02.214917 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341" containerName="horizontest-tests-horizontest" Nov 22 10:21:02 crc kubenswrapper[4789]: E1122 10:21:02.214941 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ac48b63-d114-419a-ba2f-005c8a583370" containerName="extract-content" Nov 22 10:21:02 crc kubenswrapper[4789]: I1122 10:21:02.214994 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ac48b63-d114-419a-ba2f-005c8a583370" containerName="extract-content" Nov 22 10:21:02 crc kubenswrapper[4789]: E1122 10:21:02.215031 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1202c7f8-4cc4-40e0-96a0-5f5de25cf845" containerName="extract-utilities" Nov 22 10:21:02 crc kubenswrapper[4789]: I1122 10:21:02.215080 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="1202c7f8-4cc4-40e0-96a0-5f5de25cf845" containerName="extract-utilities" Nov 22 10:21:02 crc kubenswrapper[4789]: E1122 10:21:02.215096 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1202c7f8-4cc4-40e0-96a0-5f5de25cf845" containerName="extract-content" Nov 22 10:21:02 crc kubenswrapper[4789]: I1122 10:21:02.215108 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="1202c7f8-4cc4-40e0-96a0-5f5de25cf845" containerName="extract-content" Nov 22 10:21:02 crc kubenswrapper[4789]: I1122 10:21:02.215695 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341" containerName="horizontest-tests-horizontest" Nov 22 10:21:02 crc kubenswrapper[4789]: I1122 10:21:02.215837 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="1202c7f8-4cc4-40e0-96a0-5f5de25cf845" containerName="registry-server" Nov 22 10:21:02 crc kubenswrapper[4789]: I1122 10:21:02.215921 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ac48b63-d114-419a-ba2f-005c8a583370" containerName="registry-server" Nov 22 10:21:02 crc kubenswrapper[4789]: I1122 10:21:02.217793 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest" Nov 22 10:21:02 crc kubenswrapper[4789]: I1122 10:21:02.230895 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest"] Nov 22 10:21:02 crc kubenswrapper[4789]: I1122 10:21:02.347565 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wfbgj\" (UniqueName: \"kubernetes.io/projected/7d0c8264-b357-4277-ba5e-924871a1d088-kube-api-access-wfbgj\") pod \"test-operator-logs-pod-horizontest-horizontest-tests-horizontest\" (UID: \"7d0c8264-b357-4277-ba5e-924871a1d088\") " pod="openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest" Nov 22 10:21:02 crc kubenswrapper[4789]: I1122 10:21:02.347662 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"test-operator-logs-pod-horizontest-horizontest-tests-horizontest\" (UID: \"7d0c8264-b357-4277-ba5e-924871a1d088\") " pod="openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest" Nov 22 10:21:02 crc kubenswrapper[4789]: I1122 10:21:02.450054 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wfbgj\" (UniqueName: \"kubernetes.io/projected/7d0c8264-b357-4277-ba5e-924871a1d088-kube-api-access-wfbgj\") pod \"test-operator-logs-pod-horizontest-horizontest-tests-horizontest\" (UID: \"7d0c8264-b357-4277-ba5e-924871a1d088\") " pod="openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest" Nov 22 10:21:02 crc kubenswrapper[4789]: I1122 10:21:02.450135 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"test-operator-logs-pod-horizontest-horizontest-tests-horizontest\" (UID: \"7d0c8264-b357-4277-ba5e-924871a1d088\") " pod="openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest" Nov 22 10:21:02 crc kubenswrapper[4789]: I1122 10:21:02.450769 4789 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"test-operator-logs-pod-horizontest-horizontest-tests-horizontest\" (UID: \"7d0c8264-b357-4277-ba5e-924871a1d088\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest" Nov 22 10:21:02 crc kubenswrapper[4789]: I1122 10:21:02.481980 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wfbgj\" (UniqueName: \"kubernetes.io/projected/7d0c8264-b357-4277-ba5e-924871a1d088-kube-api-access-wfbgj\") pod \"test-operator-logs-pod-horizontest-horizontest-tests-horizontest\" (UID: \"7d0c8264-b357-4277-ba5e-924871a1d088\") " pod="openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest" Nov 22 10:21:02 crc kubenswrapper[4789]: I1122 10:21:02.486546 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"test-operator-logs-pod-horizontest-horizontest-tests-horizontest\" (UID: \"7d0c8264-b357-4277-ba5e-924871a1d088\") " pod="openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest" Nov 22 10:21:02 crc kubenswrapper[4789]: I1122 10:21:02.554771 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest" Nov 22 10:21:02 crc kubenswrapper[4789]: E1122 10:21:02.554958 4789 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Nov 22 10:21:02 crc kubenswrapper[4789]: I1122 10:21:02.967462 4789 scope.go:117] "RemoveContainer" containerID="9e4b9858fc97b3d62c7338b8fc7e6fb3cc0a7ad892259a591c7f692014f9eac6" Nov 22 10:21:02 crc kubenswrapper[4789]: E1122 10:21:02.969044 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 10:21:03 crc kubenswrapper[4789]: I1122 10:21:03.431640 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest"] Nov 22 10:21:03 crc kubenswrapper[4789]: E1122 10:21:03.433261 4789 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Nov 22 10:21:03 crc kubenswrapper[4789]: I1122 10:21:03.680298 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest" event={"ID":"7d0c8264-b357-4277-ba5e-924871a1d088","Type":"ContainerStarted","Data":"f8c36a7874de2692f6be1d23bad5bd115662cc7ca3b5dfe22ff69bf222c21aa9"} Nov 22 10:21:03 crc kubenswrapper[4789]: E1122 10:21:03.845266 4789 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Nov 22 10:21:04 crc kubenswrapper[4789]: I1122 10:21:04.695783 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest" event={"ID":"7d0c8264-b357-4277-ba5e-924871a1d088","Type":"ContainerStarted","Data":"1fbd490a87ba7ccb547a2f10461acb5ca19ca01185724289511b6e83a3c537fe"} Nov 22 10:21:04 crc kubenswrapper[4789]: E1122 10:21:04.696767 4789 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Nov 22 10:21:04 crc kubenswrapper[4789]: I1122 10:21:04.728437 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest" podStartSLOduration=2.318418329 podStartE2EDuration="2.728409846s" podCreationTimestamp="2025-11-22 10:21:02 +0000 UTC" firstStartedPulling="2025-11-22 10:21:03.435185211 +0000 UTC m=+8677.669585494" lastFinishedPulling="2025-11-22 10:21:03.845176718 +0000 UTC m=+8678.079577011" observedRunningTime="2025-11-22 10:21:04.712009284 +0000 UTC m=+8678.946409587" watchObservedRunningTime="2025-11-22 10:21:04.728409846 +0000 UTC m=+8678.962810159" Nov 22 10:21:05 crc kubenswrapper[4789]: E1122 10:21:05.709061 4789 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Nov 22 10:21:13 crc kubenswrapper[4789]: I1122 10:21:13.965966 4789 scope.go:117] "RemoveContainer" containerID="9e4b9858fc97b3d62c7338b8fc7e6fb3cc0a7ad892259a591c7f692014f9eac6" Nov 22 10:21:13 crc kubenswrapper[4789]: E1122 10:21:13.967380 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 10:21:26 crc kubenswrapper[4789]: I1122 10:21:26.964964 4789 scope.go:117] "RemoveContainer" containerID="9e4b9858fc97b3d62c7338b8fc7e6fb3cc0a7ad892259a591c7f692014f9eac6" Nov 22 10:21:26 crc kubenswrapper[4789]: E1122 10:21:26.965745 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 10:21:39 crc kubenswrapper[4789]: I1122 10:21:39.965004 4789 scope.go:117] "RemoveContainer" containerID="9e4b9858fc97b3d62c7338b8fc7e6fb3cc0a7ad892259a591c7f692014f9eac6" Nov 22 10:21:39 crc kubenswrapper[4789]: E1122 10:21:39.965897 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 10:21:42 crc kubenswrapper[4789]: I1122 10:21:42.497000 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-nv7kv"] Nov 22 10:21:42 crc kubenswrapper[4789]: I1122 10:21:42.500900 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nv7kv" Nov 22 10:21:42 crc kubenswrapper[4789]: I1122 10:21:42.519251 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nv7kv"] Nov 22 10:21:42 crc kubenswrapper[4789]: I1122 10:21:42.599090 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e9fbb55-57e6-4fb3-8edd-8d91a82f5b65-catalog-content\") pod \"redhat-operators-nv7kv\" (UID: \"7e9fbb55-57e6-4fb3-8edd-8d91a82f5b65\") " pod="openshift-marketplace/redhat-operators-nv7kv" Nov 22 10:21:42 crc kubenswrapper[4789]: I1122 10:21:42.599583 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e9fbb55-57e6-4fb3-8edd-8d91a82f5b65-utilities\") pod \"redhat-operators-nv7kv\" (UID: \"7e9fbb55-57e6-4fb3-8edd-8d91a82f5b65\") " pod="openshift-marketplace/redhat-operators-nv7kv" Nov 22 10:21:42 crc kubenswrapper[4789]: I1122 10:21:42.599662 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b5zxd\" (UniqueName: \"kubernetes.io/projected/7e9fbb55-57e6-4fb3-8edd-8d91a82f5b65-kube-api-access-b5zxd\") pod \"redhat-operators-nv7kv\" (UID: \"7e9fbb55-57e6-4fb3-8edd-8d91a82f5b65\") " pod="openshift-marketplace/redhat-operators-nv7kv" Nov 22 10:21:42 crc kubenswrapper[4789]: I1122 10:21:42.701170 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e9fbb55-57e6-4fb3-8edd-8d91a82f5b65-utilities\") pod \"redhat-operators-nv7kv\" (UID: \"7e9fbb55-57e6-4fb3-8edd-8d91a82f5b65\") " pod="openshift-marketplace/redhat-operators-nv7kv" Nov 22 10:21:42 crc kubenswrapper[4789]: I1122 10:21:42.701462 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b5zxd\" (UniqueName: \"kubernetes.io/projected/7e9fbb55-57e6-4fb3-8edd-8d91a82f5b65-kube-api-access-b5zxd\") pod \"redhat-operators-nv7kv\" (UID: \"7e9fbb55-57e6-4fb3-8edd-8d91a82f5b65\") " pod="openshift-marketplace/redhat-operators-nv7kv" Nov 22 10:21:42 crc kubenswrapper[4789]: I1122 10:21:42.701644 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e9fbb55-57e6-4fb3-8edd-8d91a82f5b65-catalog-content\") pod \"redhat-operators-nv7kv\" (UID: \"7e9fbb55-57e6-4fb3-8edd-8d91a82f5b65\") " pod="openshift-marketplace/redhat-operators-nv7kv" Nov 22 10:21:42 crc kubenswrapper[4789]: I1122 10:21:42.701976 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e9fbb55-57e6-4fb3-8edd-8d91a82f5b65-utilities\") pod \"redhat-operators-nv7kv\" (UID: \"7e9fbb55-57e6-4fb3-8edd-8d91a82f5b65\") " pod="openshift-marketplace/redhat-operators-nv7kv" Nov 22 10:21:42 crc kubenswrapper[4789]: I1122 10:21:42.702075 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e9fbb55-57e6-4fb3-8edd-8d91a82f5b65-catalog-content\") pod \"redhat-operators-nv7kv\" (UID: \"7e9fbb55-57e6-4fb3-8edd-8d91a82f5b65\") " pod="openshift-marketplace/redhat-operators-nv7kv" Nov 22 10:21:42 crc kubenswrapper[4789]: I1122 10:21:42.721000 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b5zxd\" (UniqueName: \"kubernetes.io/projected/7e9fbb55-57e6-4fb3-8edd-8d91a82f5b65-kube-api-access-b5zxd\") pod \"redhat-operators-nv7kv\" (UID: \"7e9fbb55-57e6-4fb3-8edd-8d91a82f5b65\") " pod="openshift-marketplace/redhat-operators-nv7kv" Nov 22 10:21:42 crc kubenswrapper[4789]: I1122 10:21:42.843746 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nv7kv" Nov 22 10:21:43 crc kubenswrapper[4789]: I1122 10:21:43.332180 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nv7kv"] Nov 22 10:21:44 crc kubenswrapper[4789]: I1122 10:21:44.192566 4789 generic.go:334] "Generic (PLEG): container finished" podID="7e9fbb55-57e6-4fb3-8edd-8d91a82f5b65" containerID="2a461ff89617e136fd19118efde1787cf7bc60f48f39d26428e0b99ebc8b81a4" exitCode=0 Nov 22 10:21:44 crc kubenswrapper[4789]: I1122 10:21:44.192780 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nv7kv" event={"ID":"7e9fbb55-57e6-4fb3-8edd-8d91a82f5b65","Type":"ContainerDied","Data":"2a461ff89617e136fd19118efde1787cf7bc60f48f39d26428e0b99ebc8b81a4"} Nov 22 10:21:44 crc kubenswrapper[4789]: I1122 10:21:44.192978 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nv7kv" event={"ID":"7e9fbb55-57e6-4fb3-8edd-8d91a82f5b65","Type":"ContainerStarted","Data":"f80c2d544141798aca1408d05aec767cb91e5b6de9077b7b3e479db3405abde3"} Nov 22 10:21:45 crc kubenswrapper[4789]: I1122 10:21:45.208228 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nv7kv" event={"ID":"7e9fbb55-57e6-4fb3-8edd-8d91a82f5b65","Type":"ContainerStarted","Data":"763d3389dfd7436dbb6e4dc93df6025f4d45e05839ae79b84a46674a713e39cd"} Nov 22 10:21:46 crc kubenswrapper[4789]: I1122 10:21:46.223404 4789 generic.go:334] "Generic (PLEG): container finished" podID="7e9fbb55-57e6-4fb3-8edd-8d91a82f5b65" containerID="763d3389dfd7436dbb6e4dc93df6025f4d45e05839ae79b84a46674a713e39cd" exitCode=0 Nov 22 10:21:46 crc kubenswrapper[4789]: I1122 10:21:46.223472 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nv7kv" event={"ID":"7e9fbb55-57e6-4fb3-8edd-8d91a82f5b65","Type":"ContainerDied","Data":"763d3389dfd7436dbb6e4dc93df6025f4d45e05839ae79b84a46674a713e39cd"} Nov 22 10:21:47 crc kubenswrapper[4789]: I1122 10:21:47.238512 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nv7kv" event={"ID":"7e9fbb55-57e6-4fb3-8edd-8d91a82f5b65","Type":"ContainerStarted","Data":"751ce5a094ebba478a0dc7f88d2920aec21faa3f9d29f84bfacbc17b6a7e1cab"} Nov 22 10:21:47 crc kubenswrapper[4789]: I1122 10:21:47.272702 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-nv7kv" podStartSLOduration=2.853215756 podStartE2EDuration="5.272683457s" podCreationTimestamp="2025-11-22 10:21:42 +0000 UTC" firstStartedPulling="2025-11-22 10:21:44.195256499 +0000 UTC m=+8718.429656772" lastFinishedPulling="2025-11-22 10:21:46.61472416 +0000 UTC m=+8720.849124473" observedRunningTime="2025-11-22 10:21:47.265872755 +0000 UTC m=+8721.500273058" watchObservedRunningTime="2025-11-22 10:21:47.272683457 +0000 UTC m=+8721.507083720" Nov 22 10:21:52 crc kubenswrapper[4789]: I1122 10:21:52.844069 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-nv7kv" Nov 22 10:21:52 crc kubenswrapper[4789]: I1122 10:21:52.846222 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-nv7kv" Nov 22 10:21:53 crc kubenswrapper[4789]: I1122 10:21:53.918270 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-nv7kv" podUID="7e9fbb55-57e6-4fb3-8edd-8d91a82f5b65" containerName="registry-server" probeResult="failure" output=< Nov 22 10:21:53 crc kubenswrapper[4789]: timeout: failed to connect service ":50051" within 1s Nov 22 10:21:53 crc kubenswrapper[4789]: > Nov 22 10:21:54 crc kubenswrapper[4789]: I1122 10:21:54.966972 4789 scope.go:117] "RemoveContainer" containerID="9e4b9858fc97b3d62c7338b8fc7e6fb3cc0a7ad892259a591c7f692014f9eac6" Nov 22 10:21:54 crc kubenswrapper[4789]: E1122 10:21:54.967837 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 10:21:55 crc kubenswrapper[4789]: I1122 10:21:55.829012 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-69pj5/must-gather-shff9"] Nov 22 10:21:55 crc kubenswrapper[4789]: I1122 10:21:55.830603 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-69pj5/must-gather-shff9" Nov 22 10:21:55 crc kubenswrapper[4789]: I1122 10:21:55.832663 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-69pj5"/"default-dockercfg-grhfs" Nov 22 10:21:55 crc kubenswrapper[4789]: I1122 10:21:55.832796 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-69pj5"/"kube-root-ca.crt" Nov 22 10:21:55 crc kubenswrapper[4789]: I1122 10:21:55.837517 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-69pj5"/"openshift-service-ca.crt" Nov 22 10:21:55 crc kubenswrapper[4789]: I1122 10:21:55.837786 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-69pj5/must-gather-shff9"] Nov 22 10:21:55 crc kubenswrapper[4789]: I1122 10:21:55.838362 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wt9r5\" (UniqueName: \"kubernetes.io/projected/3d58dd27-067a-4510-b219-340217ca2ecb-kube-api-access-wt9r5\") pod \"must-gather-shff9\" (UID: \"3d58dd27-067a-4510-b219-340217ca2ecb\") " pod="openshift-must-gather-69pj5/must-gather-shff9" Nov 22 10:21:55 crc kubenswrapper[4789]: I1122 10:21:55.838456 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/3d58dd27-067a-4510-b219-340217ca2ecb-must-gather-output\") pod \"must-gather-shff9\" (UID: \"3d58dd27-067a-4510-b219-340217ca2ecb\") " pod="openshift-must-gather-69pj5/must-gather-shff9" Nov 22 10:21:55 crc kubenswrapper[4789]: I1122 10:21:55.939801 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wt9r5\" (UniqueName: \"kubernetes.io/projected/3d58dd27-067a-4510-b219-340217ca2ecb-kube-api-access-wt9r5\") pod \"must-gather-shff9\" (UID: \"3d58dd27-067a-4510-b219-340217ca2ecb\") " pod="openshift-must-gather-69pj5/must-gather-shff9" Nov 22 10:21:55 crc kubenswrapper[4789]: I1122 10:21:55.939892 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/3d58dd27-067a-4510-b219-340217ca2ecb-must-gather-output\") pod \"must-gather-shff9\" (UID: \"3d58dd27-067a-4510-b219-340217ca2ecb\") " pod="openshift-must-gather-69pj5/must-gather-shff9" Nov 22 10:21:55 crc kubenswrapper[4789]: I1122 10:21:55.940365 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/3d58dd27-067a-4510-b219-340217ca2ecb-must-gather-output\") pod \"must-gather-shff9\" (UID: \"3d58dd27-067a-4510-b219-340217ca2ecb\") " pod="openshift-must-gather-69pj5/must-gather-shff9" Nov 22 10:21:55 crc kubenswrapper[4789]: I1122 10:21:55.976352 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wt9r5\" (UniqueName: \"kubernetes.io/projected/3d58dd27-067a-4510-b219-340217ca2ecb-kube-api-access-wt9r5\") pod \"must-gather-shff9\" (UID: \"3d58dd27-067a-4510-b219-340217ca2ecb\") " pod="openshift-must-gather-69pj5/must-gather-shff9" Nov 22 10:21:56 crc kubenswrapper[4789]: I1122 10:21:56.147137 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-69pj5/must-gather-shff9" Nov 22 10:21:56 crc kubenswrapper[4789]: I1122 10:21:56.600515 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-69pj5/must-gather-shff9"] Nov 22 10:21:57 crc kubenswrapper[4789]: I1122 10:21:57.377529 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-69pj5/must-gather-shff9" event={"ID":"3d58dd27-067a-4510-b219-340217ca2ecb","Type":"ContainerStarted","Data":"e4aa0bc86b635e1fc4f956deb62b7418c0835871e8ef4ac0f29c12fc4781790d"} Nov 22 10:22:02 crc kubenswrapper[4789]: I1122 10:22:02.925513 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-nv7kv" Nov 22 10:22:02 crc kubenswrapper[4789]: I1122 10:22:02.998699 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-nv7kv" Nov 22 10:22:03 crc kubenswrapper[4789]: I1122 10:22:03.181606 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-nv7kv"] Nov 22 10:22:04 crc kubenswrapper[4789]: I1122 10:22:04.452080 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-nv7kv" podUID="7e9fbb55-57e6-4fb3-8edd-8d91a82f5b65" containerName="registry-server" containerID="cri-o://751ce5a094ebba478a0dc7f88d2920aec21faa3f9d29f84bfacbc17b6a7e1cab" gracePeriod=2 Nov 22 10:22:04 crc kubenswrapper[4789]: I1122 10:22:04.940875 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nv7kv" Nov 22 10:22:05 crc kubenswrapper[4789]: I1122 10:22:05.077918 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e9fbb55-57e6-4fb3-8edd-8d91a82f5b65-catalog-content\") pod \"7e9fbb55-57e6-4fb3-8edd-8d91a82f5b65\" (UID: \"7e9fbb55-57e6-4fb3-8edd-8d91a82f5b65\") " Nov 22 10:22:05 crc kubenswrapper[4789]: I1122 10:22:05.078149 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e9fbb55-57e6-4fb3-8edd-8d91a82f5b65-utilities\") pod \"7e9fbb55-57e6-4fb3-8edd-8d91a82f5b65\" (UID: \"7e9fbb55-57e6-4fb3-8edd-8d91a82f5b65\") " Nov 22 10:22:05 crc kubenswrapper[4789]: I1122 10:22:05.078212 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b5zxd\" (UniqueName: \"kubernetes.io/projected/7e9fbb55-57e6-4fb3-8edd-8d91a82f5b65-kube-api-access-b5zxd\") pod \"7e9fbb55-57e6-4fb3-8edd-8d91a82f5b65\" (UID: \"7e9fbb55-57e6-4fb3-8edd-8d91a82f5b65\") " Nov 22 10:22:05 crc kubenswrapper[4789]: I1122 10:22:05.079458 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7e9fbb55-57e6-4fb3-8edd-8d91a82f5b65-utilities" (OuterVolumeSpecName: "utilities") pod "7e9fbb55-57e6-4fb3-8edd-8d91a82f5b65" (UID: "7e9fbb55-57e6-4fb3-8edd-8d91a82f5b65"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 10:22:05 crc kubenswrapper[4789]: I1122 10:22:05.086447 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e9fbb55-57e6-4fb3-8edd-8d91a82f5b65-kube-api-access-b5zxd" (OuterVolumeSpecName: "kube-api-access-b5zxd") pod "7e9fbb55-57e6-4fb3-8edd-8d91a82f5b65" (UID: "7e9fbb55-57e6-4fb3-8edd-8d91a82f5b65"). InnerVolumeSpecName "kube-api-access-b5zxd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 10:22:05 crc kubenswrapper[4789]: I1122 10:22:05.169021 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7e9fbb55-57e6-4fb3-8edd-8d91a82f5b65-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7e9fbb55-57e6-4fb3-8edd-8d91a82f5b65" (UID: "7e9fbb55-57e6-4fb3-8edd-8d91a82f5b65"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 10:22:05 crc kubenswrapper[4789]: I1122 10:22:05.181284 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e9fbb55-57e6-4fb3-8edd-8d91a82f5b65-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 10:22:05 crc kubenswrapper[4789]: I1122 10:22:05.181311 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b5zxd\" (UniqueName: \"kubernetes.io/projected/7e9fbb55-57e6-4fb3-8edd-8d91a82f5b65-kube-api-access-b5zxd\") on node \"crc\" DevicePath \"\"" Nov 22 10:22:05 crc kubenswrapper[4789]: I1122 10:22:05.181322 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e9fbb55-57e6-4fb3-8edd-8d91a82f5b65-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 10:22:05 crc kubenswrapper[4789]: I1122 10:22:05.468789 4789 generic.go:334] "Generic (PLEG): container finished" podID="7e9fbb55-57e6-4fb3-8edd-8d91a82f5b65" containerID="751ce5a094ebba478a0dc7f88d2920aec21faa3f9d29f84bfacbc17b6a7e1cab" exitCode=0 Nov 22 10:22:05 crc kubenswrapper[4789]: I1122 10:22:05.468868 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nv7kv" Nov 22 10:22:05 crc kubenswrapper[4789]: I1122 10:22:05.468900 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nv7kv" event={"ID":"7e9fbb55-57e6-4fb3-8edd-8d91a82f5b65","Type":"ContainerDied","Data":"751ce5a094ebba478a0dc7f88d2920aec21faa3f9d29f84bfacbc17b6a7e1cab"} Nov 22 10:22:05 crc kubenswrapper[4789]: I1122 10:22:05.468935 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nv7kv" event={"ID":"7e9fbb55-57e6-4fb3-8edd-8d91a82f5b65","Type":"ContainerDied","Data":"f80c2d544141798aca1408d05aec767cb91e5b6de9077b7b3e479db3405abde3"} Nov 22 10:22:05 crc kubenswrapper[4789]: I1122 10:22:05.468959 4789 scope.go:117] "RemoveContainer" containerID="751ce5a094ebba478a0dc7f88d2920aec21faa3f9d29f84bfacbc17b6a7e1cab" Nov 22 10:22:05 crc kubenswrapper[4789]: I1122 10:22:05.472668 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-69pj5/must-gather-shff9" event={"ID":"3d58dd27-067a-4510-b219-340217ca2ecb","Type":"ContainerStarted","Data":"b05af0f3351f28bf0a5ba02841067f047a6b70712943d2d47ab2398855d74fe5"} Nov 22 10:22:05 crc kubenswrapper[4789]: I1122 10:22:05.472736 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-69pj5/must-gather-shff9" event={"ID":"3d58dd27-067a-4510-b219-340217ca2ecb","Type":"ContainerStarted","Data":"d9f47947cfef03169cfbf6edde2ebc8c5aac347b5c132221e3678b4c50f81740"} Nov 22 10:22:05 crc kubenswrapper[4789]: I1122 10:22:05.498524 4789 scope.go:117] "RemoveContainer" containerID="763d3389dfd7436dbb6e4dc93df6025f4d45e05839ae79b84a46674a713e39cd" Nov 22 10:22:05 crc kubenswrapper[4789]: I1122 10:22:05.512893 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-69pj5/must-gather-shff9" podStartSLOduration=2.672196236 podStartE2EDuration="10.512862069s" podCreationTimestamp="2025-11-22 10:21:55 +0000 UTC" firstStartedPulling="2025-11-22 10:21:56.603615915 +0000 UTC m=+8730.838016198" lastFinishedPulling="2025-11-22 10:22:04.444281718 +0000 UTC m=+8738.678682031" observedRunningTime="2025-11-22 10:22:05.498941804 +0000 UTC m=+8739.733342127" watchObservedRunningTime="2025-11-22 10:22:05.512862069 +0000 UTC m=+8739.747262382" Nov 22 10:22:05 crc kubenswrapper[4789]: I1122 10:22:05.533534 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-nv7kv"] Nov 22 10:22:05 crc kubenswrapper[4789]: I1122 10:22:05.544028 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-nv7kv"] Nov 22 10:22:05 crc kubenswrapper[4789]: I1122 10:22:05.551304 4789 scope.go:117] "RemoveContainer" containerID="2a461ff89617e136fd19118efde1787cf7bc60f48f39d26428e0b99ebc8b81a4" Nov 22 10:22:05 crc kubenswrapper[4789]: I1122 10:22:05.597735 4789 scope.go:117] "RemoveContainer" containerID="751ce5a094ebba478a0dc7f88d2920aec21faa3f9d29f84bfacbc17b6a7e1cab" Nov 22 10:22:05 crc kubenswrapper[4789]: E1122 10:22:05.598207 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"751ce5a094ebba478a0dc7f88d2920aec21faa3f9d29f84bfacbc17b6a7e1cab\": container with ID starting with 751ce5a094ebba478a0dc7f88d2920aec21faa3f9d29f84bfacbc17b6a7e1cab not found: ID does not exist" containerID="751ce5a094ebba478a0dc7f88d2920aec21faa3f9d29f84bfacbc17b6a7e1cab" Nov 22 10:22:05 crc kubenswrapper[4789]: I1122 10:22:05.598261 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"751ce5a094ebba478a0dc7f88d2920aec21faa3f9d29f84bfacbc17b6a7e1cab"} err="failed to get container status \"751ce5a094ebba478a0dc7f88d2920aec21faa3f9d29f84bfacbc17b6a7e1cab\": rpc error: code = NotFound desc = could not find container \"751ce5a094ebba478a0dc7f88d2920aec21faa3f9d29f84bfacbc17b6a7e1cab\": container with ID starting with 751ce5a094ebba478a0dc7f88d2920aec21faa3f9d29f84bfacbc17b6a7e1cab not found: ID does not exist" Nov 22 10:22:05 crc kubenswrapper[4789]: I1122 10:22:05.598305 4789 scope.go:117] "RemoveContainer" containerID="763d3389dfd7436dbb6e4dc93df6025f4d45e05839ae79b84a46674a713e39cd" Nov 22 10:22:05 crc kubenswrapper[4789]: E1122 10:22:05.598699 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"763d3389dfd7436dbb6e4dc93df6025f4d45e05839ae79b84a46674a713e39cd\": container with ID starting with 763d3389dfd7436dbb6e4dc93df6025f4d45e05839ae79b84a46674a713e39cd not found: ID does not exist" containerID="763d3389dfd7436dbb6e4dc93df6025f4d45e05839ae79b84a46674a713e39cd" Nov 22 10:22:05 crc kubenswrapper[4789]: I1122 10:22:05.598723 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"763d3389dfd7436dbb6e4dc93df6025f4d45e05839ae79b84a46674a713e39cd"} err="failed to get container status \"763d3389dfd7436dbb6e4dc93df6025f4d45e05839ae79b84a46674a713e39cd\": rpc error: code = NotFound desc = could not find container \"763d3389dfd7436dbb6e4dc93df6025f4d45e05839ae79b84a46674a713e39cd\": container with ID starting with 763d3389dfd7436dbb6e4dc93df6025f4d45e05839ae79b84a46674a713e39cd not found: ID does not exist" Nov 22 10:22:05 crc kubenswrapper[4789]: I1122 10:22:05.598740 4789 scope.go:117] "RemoveContainer" containerID="2a461ff89617e136fd19118efde1787cf7bc60f48f39d26428e0b99ebc8b81a4" Nov 22 10:22:05 crc kubenswrapper[4789]: E1122 10:22:05.599115 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2a461ff89617e136fd19118efde1787cf7bc60f48f39d26428e0b99ebc8b81a4\": container with ID starting with 2a461ff89617e136fd19118efde1787cf7bc60f48f39d26428e0b99ebc8b81a4 not found: ID does not exist" containerID="2a461ff89617e136fd19118efde1787cf7bc60f48f39d26428e0b99ebc8b81a4" Nov 22 10:22:05 crc kubenswrapper[4789]: I1122 10:22:05.599155 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a461ff89617e136fd19118efde1787cf7bc60f48f39d26428e0b99ebc8b81a4"} err="failed to get container status \"2a461ff89617e136fd19118efde1787cf7bc60f48f39d26428e0b99ebc8b81a4\": rpc error: code = NotFound desc = could not find container \"2a461ff89617e136fd19118efde1787cf7bc60f48f39d26428e0b99ebc8b81a4\": container with ID starting with 2a461ff89617e136fd19118efde1787cf7bc60f48f39d26428e0b99ebc8b81a4 not found: ID does not exist" Nov 22 10:22:05 crc kubenswrapper[4789]: I1122 10:22:05.987556 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7e9fbb55-57e6-4fb3-8edd-8d91a82f5b65" path="/var/lib/kubelet/pods/7e9fbb55-57e6-4fb3-8edd-8d91a82f5b65/volumes" Nov 22 10:22:08 crc kubenswrapper[4789]: E1122 10:22:08.597594 4789 upgradeaware.go:441] Error proxying data from backend to client: writeto tcp 38.129.56.56:33418->38.129.56.56:37377: read tcp 38.129.56.56:33418->38.129.56.56:37377: read: connection reset by peer Nov 22 10:22:08 crc kubenswrapper[4789]: I1122 10:22:08.964845 4789 scope.go:117] "RemoveContainer" containerID="9e4b9858fc97b3d62c7338b8fc7e6fb3cc0a7ad892259a591c7f692014f9eac6" Nov 22 10:22:08 crc kubenswrapper[4789]: E1122 10:22:08.965196 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 10:22:09 crc kubenswrapper[4789]: E1122 10:22:09.136326 4789 upgradeaware.go:441] Error proxying data from backend to client: writeto tcp 38.129.56.56:33538->38.129.56.56:37377: read tcp 38.129.56.56:33538->38.129.56.56:37377: read: connection reset by peer Nov 22 10:22:09 crc kubenswrapper[4789]: I1122 10:22:09.932872 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-69pj5/crc-debug-dntkd"] Nov 22 10:22:09 crc kubenswrapper[4789]: E1122 10:22:09.933672 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e9fbb55-57e6-4fb3-8edd-8d91a82f5b65" containerName="extract-utilities" Nov 22 10:22:09 crc kubenswrapper[4789]: I1122 10:22:09.933684 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e9fbb55-57e6-4fb3-8edd-8d91a82f5b65" containerName="extract-utilities" Nov 22 10:22:09 crc kubenswrapper[4789]: E1122 10:22:09.933698 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e9fbb55-57e6-4fb3-8edd-8d91a82f5b65" containerName="extract-content" Nov 22 10:22:09 crc kubenswrapper[4789]: I1122 10:22:09.933704 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e9fbb55-57e6-4fb3-8edd-8d91a82f5b65" containerName="extract-content" Nov 22 10:22:09 crc kubenswrapper[4789]: E1122 10:22:09.933725 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e9fbb55-57e6-4fb3-8edd-8d91a82f5b65" containerName="registry-server" Nov 22 10:22:09 crc kubenswrapper[4789]: I1122 10:22:09.933731 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e9fbb55-57e6-4fb3-8edd-8d91a82f5b65" containerName="registry-server" Nov 22 10:22:09 crc kubenswrapper[4789]: I1122 10:22:09.933931 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e9fbb55-57e6-4fb3-8edd-8d91a82f5b65" containerName="registry-server" Nov 22 10:22:09 crc kubenswrapper[4789]: I1122 10:22:09.934641 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-69pj5/crc-debug-dntkd" Nov 22 10:22:10 crc kubenswrapper[4789]: I1122 10:22:10.083800 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ffc96261-8b2f-49a1-a156-4866d65941bd-host\") pod \"crc-debug-dntkd\" (UID: \"ffc96261-8b2f-49a1-a156-4866d65941bd\") " pod="openshift-must-gather-69pj5/crc-debug-dntkd" Nov 22 10:22:10 crc kubenswrapper[4789]: I1122 10:22:10.084009 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-96jlt\" (UniqueName: \"kubernetes.io/projected/ffc96261-8b2f-49a1-a156-4866d65941bd-kube-api-access-96jlt\") pod \"crc-debug-dntkd\" (UID: \"ffc96261-8b2f-49a1-a156-4866d65941bd\") " pod="openshift-must-gather-69pj5/crc-debug-dntkd" Nov 22 10:22:10 crc kubenswrapper[4789]: I1122 10:22:10.185546 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ffc96261-8b2f-49a1-a156-4866d65941bd-host\") pod \"crc-debug-dntkd\" (UID: \"ffc96261-8b2f-49a1-a156-4866d65941bd\") " pod="openshift-must-gather-69pj5/crc-debug-dntkd" Nov 22 10:22:10 crc kubenswrapper[4789]: I1122 10:22:10.185729 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-96jlt\" (UniqueName: \"kubernetes.io/projected/ffc96261-8b2f-49a1-a156-4866d65941bd-kube-api-access-96jlt\") pod \"crc-debug-dntkd\" (UID: \"ffc96261-8b2f-49a1-a156-4866d65941bd\") " pod="openshift-must-gather-69pj5/crc-debug-dntkd" Nov 22 10:22:10 crc kubenswrapper[4789]: I1122 10:22:10.186277 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ffc96261-8b2f-49a1-a156-4866d65941bd-host\") pod \"crc-debug-dntkd\" (UID: \"ffc96261-8b2f-49a1-a156-4866d65941bd\") " pod="openshift-must-gather-69pj5/crc-debug-dntkd" Nov 22 10:22:10 crc kubenswrapper[4789]: I1122 10:22:10.206387 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-96jlt\" (UniqueName: \"kubernetes.io/projected/ffc96261-8b2f-49a1-a156-4866d65941bd-kube-api-access-96jlt\") pod \"crc-debug-dntkd\" (UID: \"ffc96261-8b2f-49a1-a156-4866d65941bd\") " pod="openshift-must-gather-69pj5/crc-debug-dntkd" Nov 22 10:22:10 crc kubenswrapper[4789]: I1122 10:22:10.260831 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-69pj5/crc-debug-dntkd" Nov 22 10:22:10 crc kubenswrapper[4789]: W1122 10:22:10.294170 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podffc96261_8b2f_49a1_a156_4866d65941bd.slice/crio-cac44c4370daa3469d4ac27fe22ae32a93f41babe803345c752b575c59674d83 WatchSource:0}: Error finding container cac44c4370daa3469d4ac27fe22ae32a93f41babe803345c752b575c59674d83: Status 404 returned error can't find the container with id cac44c4370daa3469d4ac27fe22ae32a93f41babe803345c752b575c59674d83 Nov 22 10:22:10 crc kubenswrapper[4789]: I1122 10:22:10.537555 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-69pj5/crc-debug-dntkd" event={"ID":"ffc96261-8b2f-49a1-a156-4866d65941bd","Type":"ContainerStarted","Data":"cac44c4370daa3469d4ac27fe22ae32a93f41babe803345c752b575c59674d83"} Nov 22 10:22:19 crc kubenswrapper[4789]: I1122 10:22:19.965930 4789 scope.go:117] "RemoveContainer" containerID="9e4b9858fc97b3d62c7338b8fc7e6fb3cc0a7ad892259a591c7f692014f9eac6" Nov 22 10:22:19 crc kubenswrapper[4789]: E1122 10:22:19.966866 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 10:22:21 crc kubenswrapper[4789]: I1122 10:22:21.676463 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-69pj5/crc-debug-dntkd" event={"ID":"ffc96261-8b2f-49a1-a156-4866d65941bd","Type":"ContainerStarted","Data":"383dd4ebe97bdf99ffc013e8160ff119f885ae64f982a01a9bac9082cc9a2efb"} Nov 22 10:22:21 crc kubenswrapper[4789]: I1122 10:22:21.713962 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-69pj5/crc-debug-dntkd" podStartSLOduration=2.4219750380000002 podStartE2EDuration="12.713933979s" podCreationTimestamp="2025-11-22 10:22:09 +0000 UTC" firstStartedPulling="2025-11-22 10:22:10.297979072 +0000 UTC m=+8744.532379345" lastFinishedPulling="2025-11-22 10:22:20.589938013 +0000 UTC m=+8754.824338286" observedRunningTime="2025-11-22 10:22:21.696743416 +0000 UTC m=+8755.931143709" watchObservedRunningTime="2025-11-22 10:22:21.713933979 +0000 UTC m=+8755.948334282" Nov 22 10:22:26 crc kubenswrapper[4789]: E1122 10:22:26.966338 4789 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Nov 22 10:22:31 crc kubenswrapper[4789]: I1122 10:22:31.965659 4789 scope.go:117] "RemoveContainer" containerID="9e4b9858fc97b3d62c7338b8fc7e6fb3cc0a7ad892259a591c7f692014f9eac6" Nov 22 10:22:31 crc kubenswrapper[4789]: E1122 10:22:31.966349 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 10:22:44 crc kubenswrapper[4789]: I1122 10:22:44.966546 4789 scope.go:117] "RemoveContainer" containerID="9e4b9858fc97b3d62c7338b8fc7e6fb3cc0a7ad892259a591c7f692014f9eac6" Nov 22 10:22:44 crc kubenswrapper[4789]: E1122 10:22:44.967960 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 10:22:58 crc kubenswrapper[4789]: I1122 10:22:58.965509 4789 scope.go:117] "RemoveContainer" containerID="9e4b9858fc97b3d62c7338b8fc7e6fb3cc0a7ad892259a591c7f692014f9eac6" Nov 22 10:22:58 crc kubenswrapper[4789]: E1122 10:22:58.966970 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 10:23:07 crc kubenswrapper[4789]: I1122 10:23:07.173029 4789 generic.go:334] "Generic (PLEG): container finished" podID="ffc96261-8b2f-49a1-a156-4866d65941bd" containerID="383dd4ebe97bdf99ffc013e8160ff119f885ae64f982a01a9bac9082cc9a2efb" exitCode=0 Nov 22 10:23:07 crc kubenswrapper[4789]: I1122 10:23:07.173122 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-69pj5/crc-debug-dntkd" event={"ID":"ffc96261-8b2f-49a1-a156-4866d65941bd","Type":"ContainerDied","Data":"383dd4ebe97bdf99ffc013e8160ff119f885ae64f982a01a9bac9082cc9a2efb"} Nov 22 10:23:08 crc kubenswrapper[4789]: I1122 10:23:08.306091 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-69pj5/crc-debug-dntkd" Nov 22 10:23:08 crc kubenswrapper[4789]: I1122 10:23:08.345311 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-69pj5/crc-debug-dntkd"] Nov 22 10:23:08 crc kubenswrapper[4789]: I1122 10:23:08.349860 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ffc96261-8b2f-49a1-a156-4866d65941bd-host\") pod \"ffc96261-8b2f-49a1-a156-4866d65941bd\" (UID: \"ffc96261-8b2f-49a1-a156-4866d65941bd\") " Nov 22 10:23:08 crc kubenswrapper[4789]: I1122 10:23:08.349929 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-96jlt\" (UniqueName: \"kubernetes.io/projected/ffc96261-8b2f-49a1-a156-4866d65941bd-kube-api-access-96jlt\") pod \"ffc96261-8b2f-49a1-a156-4866d65941bd\" (UID: \"ffc96261-8b2f-49a1-a156-4866d65941bd\") " Nov 22 10:23:08 crc kubenswrapper[4789]: I1122 10:23:08.350006 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ffc96261-8b2f-49a1-a156-4866d65941bd-host" (OuterVolumeSpecName: "host") pod "ffc96261-8b2f-49a1-a156-4866d65941bd" (UID: "ffc96261-8b2f-49a1-a156-4866d65941bd"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 10:23:08 crc kubenswrapper[4789]: I1122 10:23:08.350397 4789 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ffc96261-8b2f-49a1-a156-4866d65941bd-host\") on node \"crc\" DevicePath \"\"" Nov 22 10:23:08 crc kubenswrapper[4789]: I1122 10:23:08.356119 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-69pj5/crc-debug-dntkd"] Nov 22 10:23:08 crc kubenswrapper[4789]: I1122 10:23:08.356919 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ffc96261-8b2f-49a1-a156-4866d65941bd-kube-api-access-96jlt" (OuterVolumeSpecName: "kube-api-access-96jlt") pod "ffc96261-8b2f-49a1-a156-4866d65941bd" (UID: "ffc96261-8b2f-49a1-a156-4866d65941bd"). InnerVolumeSpecName "kube-api-access-96jlt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 10:23:08 crc kubenswrapper[4789]: I1122 10:23:08.453038 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-96jlt\" (UniqueName: \"kubernetes.io/projected/ffc96261-8b2f-49a1-a156-4866d65941bd-kube-api-access-96jlt\") on node \"crc\" DevicePath \"\"" Nov 22 10:23:09 crc kubenswrapper[4789]: I1122 10:23:09.209160 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cac44c4370daa3469d4ac27fe22ae32a93f41babe803345c752b575c59674d83" Nov 22 10:23:09 crc kubenswrapper[4789]: I1122 10:23:09.209294 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-69pj5/crc-debug-dntkd" Nov 22 10:23:09 crc kubenswrapper[4789]: I1122 10:23:09.646554 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-69pj5/crc-debug-95c4b"] Nov 22 10:23:09 crc kubenswrapper[4789]: E1122 10:23:09.647348 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ffc96261-8b2f-49a1-a156-4866d65941bd" containerName="container-00" Nov 22 10:23:09 crc kubenswrapper[4789]: I1122 10:23:09.647363 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="ffc96261-8b2f-49a1-a156-4866d65941bd" containerName="container-00" Nov 22 10:23:09 crc kubenswrapper[4789]: I1122 10:23:09.647676 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="ffc96261-8b2f-49a1-a156-4866d65941bd" containerName="container-00" Nov 22 10:23:09 crc kubenswrapper[4789]: I1122 10:23:09.648451 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-69pj5/crc-debug-95c4b" Nov 22 10:23:09 crc kubenswrapper[4789]: I1122 10:23:09.676115 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/aa93065a-cc74-4caf-9d47-f724b6d4dc27-host\") pod \"crc-debug-95c4b\" (UID: \"aa93065a-cc74-4caf-9d47-f724b6d4dc27\") " pod="openshift-must-gather-69pj5/crc-debug-95c4b" Nov 22 10:23:09 crc kubenswrapper[4789]: I1122 10:23:09.676168 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qlqpf\" (UniqueName: \"kubernetes.io/projected/aa93065a-cc74-4caf-9d47-f724b6d4dc27-kube-api-access-qlqpf\") pod \"crc-debug-95c4b\" (UID: \"aa93065a-cc74-4caf-9d47-f724b6d4dc27\") " pod="openshift-must-gather-69pj5/crc-debug-95c4b" Nov 22 10:23:09 crc kubenswrapper[4789]: I1122 10:23:09.777460 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/aa93065a-cc74-4caf-9d47-f724b6d4dc27-host\") pod \"crc-debug-95c4b\" (UID: \"aa93065a-cc74-4caf-9d47-f724b6d4dc27\") " pod="openshift-must-gather-69pj5/crc-debug-95c4b" Nov 22 10:23:09 crc kubenswrapper[4789]: I1122 10:23:09.777555 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qlqpf\" (UniqueName: \"kubernetes.io/projected/aa93065a-cc74-4caf-9d47-f724b6d4dc27-kube-api-access-qlqpf\") pod \"crc-debug-95c4b\" (UID: \"aa93065a-cc74-4caf-9d47-f724b6d4dc27\") " pod="openshift-must-gather-69pj5/crc-debug-95c4b" Nov 22 10:23:09 crc kubenswrapper[4789]: I1122 10:23:09.777624 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/aa93065a-cc74-4caf-9d47-f724b6d4dc27-host\") pod \"crc-debug-95c4b\" (UID: \"aa93065a-cc74-4caf-9d47-f724b6d4dc27\") " pod="openshift-must-gather-69pj5/crc-debug-95c4b" Nov 22 10:23:09 crc kubenswrapper[4789]: I1122 10:23:09.801851 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qlqpf\" (UniqueName: \"kubernetes.io/projected/aa93065a-cc74-4caf-9d47-f724b6d4dc27-kube-api-access-qlqpf\") pod \"crc-debug-95c4b\" (UID: \"aa93065a-cc74-4caf-9d47-f724b6d4dc27\") " pod="openshift-must-gather-69pj5/crc-debug-95c4b" Nov 22 10:23:09 crc kubenswrapper[4789]: I1122 10:23:09.964115 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-69pj5/crc-debug-95c4b" Nov 22 10:23:09 crc kubenswrapper[4789]: I1122 10:23:09.988124 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ffc96261-8b2f-49a1-a156-4866d65941bd" path="/var/lib/kubelet/pods/ffc96261-8b2f-49a1-a156-4866d65941bd/volumes" Nov 22 10:23:10 crc kubenswrapper[4789]: I1122 10:23:10.222401 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-69pj5/crc-debug-95c4b" event={"ID":"aa93065a-cc74-4caf-9d47-f724b6d4dc27","Type":"ContainerStarted","Data":"0f8a31d71968d5a1066981eb7b45fff72f1faa11b65d3802f65daec03872ce06"} Nov 22 10:23:11 crc kubenswrapper[4789]: I1122 10:23:11.233111 4789 generic.go:334] "Generic (PLEG): container finished" podID="aa93065a-cc74-4caf-9d47-f724b6d4dc27" containerID="b755428c6bd9a67fdfed8ed1829d794d0d37705f81439dac9f79c368932d7194" exitCode=0 Nov 22 10:23:11 crc kubenswrapper[4789]: I1122 10:23:11.233195 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-69pj5/crc-debug-95c4b" event={"ID":"aa93065a-cc74-4caf-9d47-f724b6d4dc27","Type":"ContainerDied","Data":"b755428c6bd9a67fdfed8ed1829d794d0d37705f81439dac9f79c368932d7194"} Nov 22 10:23:12 crc kubenswrapper[4789]: I1122 10:23:12.354672 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-69pj5/crc-debug-95c4b" Nov 22 10:23:12 crc kubenswrapper[4789]: I1122 10:23:12.448710 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qlqpf\" (UniqueName: \"kubernetes.io/projected/aa93065a-cc74-4caf-9d47-f724b6d4dc27-kube-api-access-qlqpf\") pod \"aa93065a-cc74-4caf-9d47-f724b6d4dc27\" (UID: \"aa93065a-cc74-4caf-9d47-f724b6d4dc27\") " Nov 22 10:23:12 crc kubenswrapper[4789]: I1122 10:23:12.448793 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/aa93065a-cc74-4caf-9d47-f724b6d4dc27-host\") pod \"aa93065a-cc74-4caf-9d47-f724b6d4dc27\" (UID: \"aa93065a-cc74-4caf-9d47-f724b6d4dc27\") " Nov 22 10:23:12 crc kubenswrapper[4789]: I1122 10:23:12.448844 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aa93065a-cc74-4caf-9d47-f724b6d4dc27-host" (OuterVolumeSpecName: "host") pod "aa93065a-cc74-4caf-9d47-f724b6d4dc27" (UID: "aa93065a-cc74-4caf-9d47-f724b6d4dc27"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 10:23:12 crc kubenswrapper[4789]: I1122 10:23:12.449325 4789 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/aa93065a-cc74-4caf-9d47-f724b6d4dc27-host\") on node \"crc\" DevicePath \"\"" Nov 22 10:23:12 crc kubenswrapper[4789]: I1122 10:23:12.453774 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa93065a-cc74-4caf-9d47-f724b6d4dc27-kube-api-access-qlqpf" (OuterVolumeSpecName: "kube-api-access-qlqpf") pod "aa93065a-cc74-4caf-9d47-f724b6d4dc27" (UID: "aa93065a-cc74-4caf-9d47-f724b6d4dc27"). InnerVolumeSpecName "kube-api-access-qlqpf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 10:23:12 crc kubenswrapper[4789]: I1122 10:23:12.550568 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qlqpf\" (UniqueName: \"kubernetes.io/projected/aa93065a-cc74-4caf-9d47-f724b6d4dc27-kube-api-access-qlqpf\") on node \"crc\" DevicePath \"\"" Nov 22 10:23:12 crc kubenswrapper[4789]: I1122 10:23:12.965170 4789 scope.go:117] "RemoveContainer" containerID="9e4b9858fc97b3d62c7338b8fc7e6fb3cc0a7ad892259a591c7f692014f9eac6" Nov 22 10:23:13 crc kubenswrapper[4789]: I1122 10:23:13.251025 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-69pj5/crc-debug-95c4b" Nov 22 10:23:13 crc kubenswrapper[4789]: I1122 10:23:13.251028 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-69pj5/crc-debug-95c4b" event={"ID":"aa93065a-cc74-4caf-9d47-f724b6d4dc27","Type":"ContainerDied","Data":"0f8a31d71968d5a1066981eb7b45fff72f1faa11b65d3802f65daec03872ce06"} Nov 22 10:23:13 crc kubenswrapper[4789]: I1122 10:23:13.251799 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0f8a31d71968d5a1066981eb7b45fff72f1faa11b65d3802f65daec03872ce06" Nov 22 10:23:13 crc kubenswrapper[4789]: I1122 10:23:13.255151 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" event={"ID":"58aa8071-7ecc-4692-8789-2db04ec70647","Type":"ContainerStarted","Data":"757306d657b53911c979b1a984e697106c927039cb3f24d4ba4a2a51a772a120"} Nov 22 10:23:13 crc kubenswrapper[4789]: I1122 10:23:13.886291 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-69pj5/crc-debug-95c4b"] Nov 22 10:23:13 crc kubenswrapper[4789]: I1122 10:23:13.894205 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-69pj5/crc-debug-95c4b"] Nov 22 10:23:13 crc kubenswrapper[4789]: I1122 10:23:13.978376 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aa93065a-cc74-4caf-9d47-f724b6d4dc27" path="/var/lib/kubelet/pods/aa93065a-cc74-4caf-9d47-f724b6d4dc27/volumes" Nov 22 10:23:15 crc kubenswrapper[4789]: I1122 10:23:15.142215 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-69pj5/crc-debug-bkns2"] Nov 22 10:23:15 crc kubenswrapper[4789]: E1122 10:23:15.143599 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa93065a-cc74-4caf-9d47-f724b6d4dc27" containerName="container-00" Nov 22 10:23:15 crc kubenswrapper[4789]: I1122 10:23:15.143656 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa93065a-cc74-4caf-9d47-f724b6d4dc27" containerName="container-00" Nov 22 10:23:15 crc kubenswrapper[4789]: I1122 10:23:15.144061 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa93065a-cc74-4caf-9d47-f724b6d4dc27" containerName="container-00" Nov 22 10:23:15 crc kubenswrapper[4789]: I1122 10:23:15.145441 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-69pj5/crc-debug-bkns2" Nov 22 10:23:15 crc kubenswrapper[4789]: I1122 10:23:15.201221 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jvvqb\" (UniqueName: \"kubernetes.io/projected/92b833a1-a710-46f4-a0ae-4b7e5eeefeb3-kube-api-access-jvvqb\") pod \"crc-debug-bkns2\" (UID: \"92b833a1-a710-46f4-a0ae-4b7e5eeefeb3\") " pod="openshift-must-gather-69pj5/crc-debug-bkns2" Nov 22 10:23:15 crc kubenswrapper[4789]: I1122 10:23:15.201316 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/92b833a1-a710-46f4-a0ae-4b7e5eeefeb3-host\") pod \"crc-debug-bkns2\" (UID: \"92b833a1-a710-46f4-a0ae-4b7e5eeefeb3\") " pod="openshift-must-gather-69pj5/crc-debug-bkns2" Nov 22 10:23:15 crc kubenswrapper[4789]: I1122 10:23:15.304744 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jvvqb\" (UniqueName: \"kubernetes.io/projected/92b833a1-a710-46f4-a0ae-4b7e5eeefeb3-kube-api-access-jvvqb\") pod \"crc-debug-bkns2\" (UID: \"92b833a1-a710-46f4-a0ae-4b7e5eeefeb3\") " pod="openshift-must-gather-69pj5/crc-debug-bkns2" Nov 22 10:23:15 crc kubenswrapper[4789]: I1122 10:23:15.304921 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/92b833a1-a710-46f4-a0ae-4b7e5eeefeb3-host\") pod \"crc-debug-bkns2\" (UID: \"92b833a1-a710-46f4-a0ae-4b7e5eeefeb3\") " pod="openshift-must-gather-69pj5/crc-debug-bkns2" Nov 22 10:23:15 crc kubenswrapper[4789]: I1122 10:23:15.305242 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/92b833a1-a710-46f4-a0ae-4b7e5eeefeb3-host\") pod \"crc-debug-bkns2\" (UID: \"92b833a1-a710-46f4-a0ae-4b7e5eeefeb3\") " pod="openshift-must-gather-69pj5/crc-debug-bkns2" Nov 22 10:23:15 crc kubenswrapper[4789]: I1122 10:23:15.330838 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jvvqb\" (UniqueName: \"kubernetes.io/projected/92b833a1-a710-46f4-a0ae-4b7e5eeefeb3-kube-api-access-jvvqb\") pod \"crc-debug-bkns2\" (UID: \"92b833a1-a710-46f4-a0ae-4b7e5eeefeb3\") " pod="openshift-must-gather-69pj5/crc-debug-bkns2" Nov 22 10:23:15 crc kubenswrapper[4789]: I1122 10:23:15.478948 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-69pj5/crc-debug-bkns2" Nov 22 10:23:15 crc kubenswrapper[4789]: W1122 10:23:15.533867 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod92b833a1_a710_46f4_a0ae_4b7e5eeefeb3.slice/crio-aaec4448b0dbd56638f1470caa1a5ad8c65fdc4dd8c9b0182a7e81d39a71a2ed WatchSource:0}: Error finding container aaec4448b0dbd56638f1470caa1a5ad8c65fdc4dd8c9b0182a7e81d39a71a2ed: Status 404 returned error can't find the container with id aaec4448b0dbd56638f1470caa1a5ad8c65fdc4dd8c9b0182a7e81d39a71a2ed Nov 22 10:23:16 crc kubenswrapper[4789]: I1122 10:23:16.288424 4789 generic.go:334] "Generic (PLEG): container finished" podID="92b833a1-a710-46f4-a0ae-4b7e5eeefeb3" containerID="f2bbe69031df1849eea0a458dccdb00eabc3a23637ca2bdfd3966f6227b3fd92" exitCode=0 Nov 22 10:23:16 crc kubenswrapper[4789]: I1122 10:23:16.288509 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-69pj5/crc-debug-bkns2" event={"ID":"92b833a1-a710-46f4-a0ae-4b7e5eeefeb3","Type":"ContainerDied","Data":"f2bbe69031df1849eea0a458dccdb00eabc3a23637ca2bdfd3966f6227b3fd92"} Nov 22 10:23:16 crc kubenswrapper[4789]: I1122 10:23:16.289286 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-69pj5/crc-debug-bkns2" event={"ID":"92b833a1-a710-46f4-a0ae-4b7e5eeefeb3","Type":"ContainerStarted","Data":"aaec4448b0dbd56638f1470caa1a5ad8c65fdc4dd8c9b0182a7e81d39a71a2ed"} Nov 22 10:23:16 crc kubenswrapper[4789]: I1122 10:23:16.362402 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-69pj5/crc-debug-bkns2"] Nov 22 10:23:16 crc kubenswrapper[4789]: I1122 10:23:16.380942 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-69pj5/crc-debug-bkns2"] Nov 22 10:23:17 crc kubenswrapper[4789]: I1122 10:23:17.424614 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-69pj5/crc-debug-bkns2" Nov 22 10:23:17 crc kubenswrapper[4789]: I1122 10:23:17.453066 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jvvqb\" (UniqueName: \"kubernetes.io/projected/92b833a1-a710-46f4-a0ae-4b7e5eeefeb3-kube-api-access-jvvqb\") pod \"92b833a1-a710-46f4-a0ae-4b7e5eeefeb3\" (UID: \"92b833a1-a710-46f4-a0ae-4b7e5eeefeb3\") " Nov 22 10:23:17 crc kubenswrapper[4789]: I1122 10:23:17.453267 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/92b833a1-a710-46f4-a0ae-4b7e5eeefeb3-host\") pod \"92b833a1-a710-46f4-a0ae-4b7e5eeefeb3\" (UID: \"92b833a1-a710-46f4-a0ae-4b7e5eeefeb3\") " Nov 22 10:23:17 crc kubenswrapper[4789]: I1122 10:23:17.453382 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/92b833a1-a710-46f4-a0ae-4b7e5eeefeb3-host" (OuterVolumeSpecName: "host") pod "92b833a1-a710-46f4-a0ae-4b7e5eeefeb3" (UID: "92b833a1-a710-46f4-a0ae-4b7e5eeefeb3"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 10:23:17 crc kubenswrapper[4789]: I1122 10:23:17.453920 4789 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/92b833a1-a710-46f4-a0ae-4b7e5eeefeb3-host\") on node \"crc\" DevicePath \"\"" Nov 22 10:23:17 crc kubenswrapper[4789]: I1122 10:23:17.461151 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92b833a1-a710-46f4-a0ae-4b7e5eeefeb3-kube-api-access-jvvqb" (OuterVolumeSpecName: "kube-api-access-jvvqb") pod "92b833a1-a710-46f4-a0ae-4b7e5eeefeb3" (UID: "92b833a1-a710-46f4-a0ae-4b7e5eeefeb3"). InnerVolumeSpecName "kube-api-access-jvvqb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 10:23:17 crc kubenswrapper[4789]: I1122 10:23:17.561626 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jvvqb\" (UniqueName: \"kubernetes.io/projected/92b833a1-a710-46f4-a0ae-4b7e5eeefeb3-kube-api-access-jvvqb\") on node \"crc\" DevicePath \"\"" Nov 22 10:23:17 crc kubenswrapper[4789]: I1122 10:23:17.983359 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="92b833a1-a710-46f4-a0ae-4b7e5eeefeb3" path="/var/lib/kubelet/pods/92b833a1-a710-46f4-a0ae-4b7e5eeefeb3/volumes" Nov 22 10:23:18 crc kubenswrapper[4789]: I1122 10:23:18.313548 4789 scope.go:117] "RemoveContainer" containerID="f2bbe69031df1849eea0a458dccdb00eabc3a23637ca2bdfd3966f6227b3fd92" Nov 22 10:23:18 crc kubenswrapper[4789]: I1122 10:23:18.313603 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-69pj5/crc-debug-bkns2" Nov 22 10:23:54 crc kubenswrapper[4789]: E1122 10:23:54.965405 4789 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Nov 22 10:24:02 crc kubenswrapper[4789]: I1122 10:24:02.409939 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ansibletest-ansibletest_0e385574-644b-4047-a329-a392a52c663f/ansibletest-ansibletest/0.log" Nov 22 10:24:02 crc kubenswrapper[4789]: I1122 10:24:02.524306 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-54cbbcb976-cxb6c_ee6a65b2-109d-44d8-a4e0-62ab28d392e4/barbican-api/0.log" Nov 22 10:24:02 crc kubenswrapper[4789]: I1122 10:24:02.573478 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-54cbbcb976-cxb6c_ee6a65b2-109d-44d8-a4e0-62ab28d392e4/barbican-api-log/0.log" Nov 22 10:24:02 crc kubenswrapper[4789]: I1122 10:24:02.741076 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-7ccbbd9f98-jd5vz_affd3f11-c0d4-41a2-888b-f4ae0c7e3d41/barbican-keystone-listener/0.log" Nov 22 10:24:02 crc kubenswrapper[4789]: I1122 10:24:02.880417 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-75d6db87b5-kbchw_4be1163c-8900-4db9-a041-826526ae8751/barbican-worker/0.log" Nov 22 10:24:02 crc kubenswrapper[4789]: I1122 10:24:02.938210 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-75d6db87b5-kbchw_4be1163c-8900-4db9-a041-826526ae8751/barbican-worker-log/0.log" Nov 22 10:24:03 crc kubenswrapper[4789]: I1122 10:24:03.121941 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-bxq52_61e96af2-fd7c-4c68-959b-db900528ab2e/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Nov 22 10:24:03 crc kubenswrapper[4789]: I1122 10:24:03.191709 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-7ccbbd9f98-jd5vz_affd3f11-c0d4-41a2-888b-f4ae0c7e3d41/barbican-keystone-listener-log/0.log" Nov 22 10:24:03 crc kubenswrapper[4789]: I1122 10:24:03.302920 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_deee1bbe-0f19-4f57-bebc-de5b01ccac7f/ceilometer-central-agent/1.log" Nov 22 10:24:03 crc kubenswrapper[4789]: I1122 10:24:03.375318 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_deee1bbe-0f19-4f57-bebc-de5b01ccac7f/ceilometer-central-agent/0.log" Nov 22 10:24:03 crc kubenswrapper[4789]: I1122 10:24:03.382382 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_deee1bbe-0f19-4f57-bebc-de5b01ccac7f/ceilometer-notification-agent/0.log" Nov 22 10:24:03 crc kubenswrapper[4789]: I1122 10:24:03.490178 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_deee1bbe-0f19-4f57-bebc-de5b01ccac7f/sg-core/0.log" Nov 22 10:24:03 crc kubenswrapper[4789]: I1122 10:24:03.507701 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_deee1bbe-0f19-4f57-bebc-de5b01ccac7f/proxy-httpd/0.log" Nov 22 10:24:03 crc kubenswrapper[4789]: I1122 10:24:03.601829 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-client-edpm-deployment-openstack-edpm-ipam-k4r2f_f8d83872-bd0a-4fad-ae6d-3bfd1c175caa/ceph-client-edpm-deployment-openstack-edpm-ipam/0.log" Nov 22 10:24:03 crc kubenswrapper[4789]: I1122 10:24:03.733406 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-mdh5v_a05e1e58-64e0-4b6c-bab9-b4ad7a8eea2b/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam/0.log" Nov 22 10:24:03 crc kubenswrapper[4789]: I1122 10:24:03.992154 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_27920801-df7f-48f4-8830-6b190dea7234/cinder-api-log/0.log" Nov 22 10:24:04 crc kubenswrapper[4789]: I1122 10:24:04.061415 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_27920801-df7f-48f4-8830-6b190dea7234/cinder-api/0.log" Nov 22 10:24:04 crc kubenswrapper[4789]: I1122 10:24:04.184996 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_62901ed9-929f-443e-ae1b-dc75cb2e1514/cinder-backup/0.log" Nov 22 10:24:04 crc kubenswrapper[4789]: I1122 10:24:04.321301 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_62901ed9-929f-443e-ae1b-dc75cb2e1514/probe/0.log" Nov 22 10:24:04 crc kubenswrapper[4789]: I1122 10:24:04.326619 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_b24f838e-1881-4921-8f51-8bb32afc4177/cinder-scheduler/0.log" Nov 22 10:24:04 crc kubenswrapper[4789]: I1122 10:24:04.927341 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_2b1ad6b9-9f67-4533-bafd-5933ade7f938/probe/0.log" Nov 22 10:24:04 crc kubenswrapper[4789]: I1122 10:24:04.976371 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_2b1ad6b9-9f67-4533-bafd-5933ade7f938/cinder-volume/0.log" Nov 22 10:24:04 crc kubenswrapper[4789]: I1122 10:24:04.982905 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_b24f838e-1881-4921-8f51-8bb32afc4177/probe/0.log" Nov 22 10:24:05 crc kubenswrapper[4789]: I1122 10:24:05.136165 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-lwqsb_981cbfb1-edd3-449f-865e-b958220cf470/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Nov 22 10:24:05 crc kubenswrapper[4789]: I1122 10:24:05.201653 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-vt7vg_2f51ca85-18d1-429b-99ce-2785256d7bb8/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 22 10:24:05 crc kubenswrapper[4789]: I1122 10:24:05.331789 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6885d49d55-m8c6x_5e1d2ec8-1172-4bcc-a5e7-5cbbea574845/init/0.log" Nov 22 10:24:05 crc kubenswrapper[4789]: I1122 10:24:05.508981 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6885d49d55-m8c6x_5e1d2ec8-1172-4bcc-a5e7-5cbbea574845/init/0.log" Nov 22 10:24:05 crc kubenswrapper[4789]: I1122 10:24:05.667478 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_9b280529-6f38-4c33-8f60-e2aad8fbb2a1/glance-httpd/0.log" Nov 22 10:24:05 crc kubenswrapper[4789]: I1122 10:24:05.690944 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_9b280529-6f38-4c33-8f60-e2aad8fbb2a1/glance-log/0.log" Nov 22 10:24:05 crc kubenswrapper[4789]: I1122 10:24:05.745892 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6885d49d55-m8c6x_5e1d2ec8-1172-4bcc-a5e7-5cbbea574845/dnsmasq-dns/0.log" Nov 22 10:24:05 crc kubenswrapper[4789]: I1122 10:24:05.875487 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_8660ff9f-23fb-4be2-93f4-330ffdab1566/glance-log/0.log" Nov 22 10:24:05 crc kubenswrapper[4789]: I1122 10:24:05.889580 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_8660ff9f-23fb-4be2-93f4-330ffdab1566/glance-httpd/0.log" Nov 22 10:24:06 crc kubenswrapper[4789]: I1122 10:24:06.076539 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-7b778f8c8-6rl4f_d9fb66f7-ee5b-453a-a69b-56fb9337d898/horizon/0.log" Nov 22 10:24:06 crc kubenswrapper[4789]: I1122 10:24:06.095515 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizontest-tests-horizontest_e0ed751e-8e37-4e1f-9aa6-7fb4d9b31341/horizontest-tests-horizontest/0.log" Nov 22 10:24:06 crc kubenswrapper[4789]: I1122 10:24:06.338669 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-nvsts_07bdec58-1883-4ddb-9aa0-b6eca2498ed4/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Nov 22 10:24:06 crc kubenswrapper[4789]: I1122 10:24:06.359763 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-rrdbw_3798e707-7719-49a1-b644-ce53243a5131/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 22 10:24:06 crc kubenswrapper[4789]: I1122 10:24:06.598630 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29396701-s2qrc_716edb71-e43b-4217-86a7-adab440be7f5/keystone-cron/0.log" Nov 22 10:24:06 crc kubenswrapper[4789]: I1122 10:24:06.837540 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29396761-w9qjq_21453b21-488b-4068-8ba4-1fa71816009d/keystone-cron/0.log" Nov 22 10:24:06 crc kubenswrapper[4789]: I1122 10:24:06.910819 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_a9945dfe-b305-48fc-9a1e-f5559b19a8b3/kube-state-metrics/0.log" Nov 22 10:24:07 crc kubenswrapper[4789]: I1122 10:24:07.104067 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-vbx9r_fd548b7c-689e-4221-ba0f-5c57f8b12927/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Nov 22 10:24:07 crc kubenswrapper[4789]: I1122 10:24:07.345712 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_e21ea1ba-4437-4d76-be37-0b285909d631/manila-api-log/0.log" Nov 22 10:24:07 crc kubenswrapper[4789]: I1122 10:24:07.506183 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_e21ea1ba-4437-4d76-be37-0b285909d631/manila-api/0.log" Nov 22 10:24:07 crc kubenswrapper[4789]: I1122 10:24:07.613245 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_8a211302-5dc3-4ba1-959e-882c398a7f6c/probe/0.log" Nov 22 10:24:07 crc kubenswrapper[4789]: I1122 10:24:07.637870 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_8a211302-5dc3-4ba1-959e-882c398a7f6c/manila-scheduler/0.log" Nov 22 10:24:07 crc kubenswrapper[4789]: I1122 10:24:07.704264 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-7b778f8c8-6rl4f_d9fb66f7-ee5b-453a-a69b-56fb9337d898/horizon-log/0.log" Nov 22 10:24:07 crc kubenswrapper[4789]: I1122 10:24:07.877773 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_f4f5702a-2376-40d4-9079-945c44d5466f/probe/0.log" Nov 22 10:24:07 crc kubenswrapper[4789]: I1122 10:24:07.943291 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_f4f5702a-2376-40d4-9079-945c44d5466f/manila-share/0.log" Nov 22 10:24:08 crc kubenswrapper[4789]: I1122 10:24:08.534360 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-qn45p_65952d06-ccf5-425f-bb4a-ad127a6844df/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Nov 22 10:24:09 crc kubenswrapper[4789]: I1122 10:24:09.242155 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-5fc55b98d9-vqxcz_a5636b40-2e27-4073-8f3a-1d429a885936/neutron-httpd/0.log" Nov 22 10:24:10 crc kubenswrapper[4789]: I1122 10:24:10.067396 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-5c64d54575-bxqxv_60140344-7346-4115-adcb-45e3f7f366a0/keystone-api/0.log" Nov 22 10:24:10 crc kubenswrapper[4789]: I1122 10:24:10.106079 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-5fc55b98d9-vqxcz_a5636b40-2e27-4073-8f3a-1d429a885936/neutron-api/0.log" Nov 22 10:24:10 crc kubenswrapper[4789]: I1122 10:24:10.841929 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_298b1252-2a3e-4f1a-a5af-5e573964ddd4/nova-cell1-conductor-conductor/0.log" Nov 22 10:24:10 crc kubenswrapper[4789]: I1122 10:24:10.885411 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_94983cbb-ec60-487e-8feb-42196210f1b6/nova-cell0-conductor-conductor/0.log" Nov 22 10:24:11 crc kubenswrapper[4789]: I1122 10:24:11.303817 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_99fff521-c886-48d2-8858-1ad7fb4a33f2/nova-cell1-novncproxy-novncproxy/0.log" Nov 22 10:24:11 crc kubenswrapper[4789]: I1122 10:24:11.583208 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b47tx_2c231e91-12f1-47e4-8947-1bf03bcae704/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam/0.log" Nov 22 10:24:11 crc kubenswrapper[4789]: I1122 10:24:11.888644 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_588fc366-e2d8-40bc-b09e-ebcf032e440f/nova-metadata-log/0.log" Nov 22 10:24:13 crc kubenswrapper[4789]: I1122 10:24:13.012295 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_a1cea96e-2c65-4231-a4b4-bf6d197c5981/nova-scheduler-scheduler/0.log" Nov 22 10:24:13 crc kubenswrapper[4789]: I1122 10:24:13.298273 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_9585ddf1-5799-43c3-a747-860ef1bf8380/nova-api-log/0.log" Nov 22 10:24:13 crc kubenswrapper[4789]: I1122 10:24:13.425406 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_7778e90a-45cf-4592-ab3d-68c75258e77f/mysql-bootstrap/0.log" Nov 22 10:24:13 crc kubenswrapper[4789]: I1122 10:24:13.649255 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_7778e90a-45cf-4592-ab3d-68c75258e77f/mysql-bootstrap/0.log" Nov 22 10:24:13 crc kubenswrapper[4789]: I1122 10:24:13.693969 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_7778e90a-45cf-4592-ab3d-68c75258e77f/galera/0.log" Nov 22 10:24:13 crc kubenswrapper[4789]: I1122 10:24:13.895045 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_51737c30-e744-4266-a235-1ddd1c115c4e/mysql-bootstrap/0.log" Nov 22 10:24:14 crc kubenswrapper[4789]: I1122 10:24:14.076608 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_51737c30-e744-4266-a235-1ddd1c115c4e/mysql-bootstrap/0.log" Nov 22 10:24:14 crc kubenswrapper[4789]: I1122 10:24:14.142830 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_51737c30-e744-4266-a235-1ddd1c115c4e/galera/0.log" Nov 22 10:24:14 crc kubenswrapper[4789]: I1122 10:24:14.322673 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_13cfa613-fa3b-4a2f-a228-0fbf526e489e/openstackclient/0.log" Nov 22 10:24:14 crc kubenswrapper[4789]: I1122 10:24:14.415194 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_9585ddf1-5799-43c3-a747-860ef1bf8380/nova-api-api/0.log" Nov 22 10:24:14 crc kubenswrapper[4789]: I1122 10:24:14.598231 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-cl8m4_e6f04e32-3dca-4d6d-8c8d-19ea15f46875/ovn-controller/0.log" Nov 22 10:24:14 crc kubenswrapper[4789]: I1122 10:24:14.883886 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-c6fhr_0e0f4cec-d122-4b44-8344-1cd6e515ef19/openstack-network-exporter/0.log" Nov 22 10:24:15 crc kubenswrapper[4789]: I1122 10:24:15.034943 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-gjjwv_ea7bbfd1-bdb6-41b2-87f5-2fa2bbe0c804/ovsdb-server-init/0.log" Nov 22 10:24:15 crc kubenswrapper[4789]: I1122 10:24:15.131332 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-kzmrw"] Nov 22 10:24:15 crc kubenswrapper[4789]: E1122 10:24:15.131713 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92b833a1-a710-46f4-a0ae-4b7e5eeefeb3" containerName="container-00" Nov 22 10:24:15 crc kubenswrapper[4789]: I1122 10:24:15.131725 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="92b833a1-a710-46f4-a0ae-4b7e5eeefeb3" containerName="container-00" Nov 22 10:24:15 crc kubenswrapper[4789]: I1122 10:24:15.131937 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="92b833a1-a710-46f4-a0ae-4b7e5eeefeb3" containerName="container-00" Nov 22 10:24:15 crc kubenswrapper[4789]: I1122 10:24:15.133153 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kzmrw" Nov 22 10:24:15 crc kubenswrapper[4789]: I1122 10:24:15.189891 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kzmrw"] Nov 22 10:24:15 crc kubenswrapper[4789]: I1122 10:24:15.271947 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-gjjwv_ea7bbfd1-bdb6-41b2-87f5-2fa2bbe0c804/ovsdb-server-init/0.log" Nov 22 10:24:15 crc kubenswrapper[4789]: I1122 10:24:15.272679 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8dce9e8b-6c82-44b6-98fe-72555fd207fe-catalog-content\") pod \"community-operators-kzmrw\" (UID: \"8dce9e8b-6c82-44b6-98fe-72555fd207fe\") " pod="openshift-marketplace/community-operators-kzmrw" Nov 22 10:24:15 crc kubenswrapper[4789]: I1122 10:24:15.272775 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m7gdk\" (UniqueName: \"kubernetes.io/projected/8dce9e8b-6c82-44b6-98fe-72555fd207fe-kube-api-access-m7gdk\") pod \"community-operators-kzmrw\" (UID: \"8dce9e8b-6c82-44b6-98fe-72555fd207fe\") " pod="openshift-marketplace/community-operators-kzmrw" Nov 22 10:24:15 crc kubenswrapper[4789]: I1122 10:24:15.272974 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8dce9e8b-6c82-44b6-98fe-72555fd207fe-utilities\") pod \"community-operators-kzmrw\" (UID: \"8dce9e8b-6c82-44b6-98fe-72555fd207fe\") " pod="openshift-marketplace/community-operators-kzmrw" Nov 22 10:24:15 crc kubenswrapper[4789]: I1122 10:24:15.371790 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-gjjwv_ea7bbfd1-bdb6-41b2-87f5-2fa2bbe0c804/ovs-vswitchd/0.log" Nov 22 10:24:15 crc kubenswrapper[4789]: I1122 10:24:15.378383 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8dce9e8b-6c82-44b6-98fe-72555fd207fe-utilities\") pod \"community-operators-kzmrw\" (UID: \"8dce9e8b-6c82-44b6-98fe-72555fd207fe\") " pod="openshift-marketplace/community-operators-kzmrw" Nov 22 10:24:15 crc kubenswrapper[4789]: I1122 10:24:15.378431 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8dce9e8b-6c82-44b6-98fe-72555fd207fe-catalog-content\") pod \"community-operators-kzmrw\" (UID: \"8dce9e8b-6c82-44b6-98fe-72555fd207fe\") " pod="openshift-marketplace/community-operators-kzmrw" Nov 22 10:24:15 crc kubenswrapper[4789]: I1122 10:24:15.378475 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m7gdk\" (UniqueName: \"kubernetes.io/projected/8dce9e8b-6c82-44b6-98fe-72555fd207fe-kube-api-access-m7gdk\") pod \"community-operators-kzmrw\" (UID: \"8dce9e8b-6c82-44b6-98fe-72555fd207fe\") " pod="openshift-marketplace/community-operators-kzmrw" Nov 22 10:24:15 crc kubenswrapper[4789]: I1122 10:24:15.379271 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8dce9e8b-6c82-44b6-98fe-72555fd207fe-utilities\") pod \"community-operators-kzmrw\" (UID: \"8dce9e8b-6c82-44b6-98fe-72555fd207fe\") " pod="openshift-marketplace/community-operators-kzmrw" Nov 22 10:24:15 crc kubenswrapper[4789]: I1122 10:24:15.379485 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8dce9e8b-6c82-44b6-98fe-72555fd207fe-catalog-content\") pod \"community-operators-kzmrw\" (UID: \"8dce9e8b-6c82-44b6-98fe-72555fd207fe\") " pod="openshift-marketplace/community-operators-kzmrw" Nov 22 10:24:15 crc kubenswrapper[4789]: I1122 10:24:15.402909 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m7gdk\" (UniqueName: \"kubernetes.io/projected/8dce9e8b-6c82-44b6-98fe-72555fd207fe-kube-api-access-m7gdk\") pod \"community-operators-kzmrw\" (UID: \"8dce9e8b-6c82-44b6-98fe-72555fd207fe\") " pod="openshift-marketplace/community-operators-kzmrw" Nov 22 10:24:15 crc kubenswrapper[4789]: I1122 10:24:15.435493 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-gjjwv_ea7bbfd1-bdb6-41b2-87f5-2fa2bbe0c804/ovsdb-server/0.log" Nov 22 10:24:15 crc kubenswrapper[4789]: I1122 10:24:15.480164 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kzmrw" Nov 22 10:24:15 crc kubenswrapper[4789]: I1122 10:24:15.851386 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-tqgkq_afd5b574-89f9-4b13-a3ee-b5e4c39095ff/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Nov 22 10:24:15 crc kubenswrapper[4789]: I1122 10:24:15.923002 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_ec49bfda-737c-49c9-8bc3-cfceb08e5f88/openstack-network-exporter/0.log" Nov 22 10:24:16 crc kubenswrapper[4789]: I1122 10:24:16.031689 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_ec49bfda-737c-49c9-8bc3-cfceb08e5f88/ovn-northd/0.log" Nov 22 10:24:16 crc kubenswrapper[4789]: I1122 10:24:16.065488 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kzmrw"] Nov 22 10:24:16 crc kubenswrapper[4789]: I1122 10:24:16.129189 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_b479a8a1-0fba-4a72-b830-2a68e2bd9be4/openstack-network-exporter/0.log" Nov 22 10:24:16 crc kubenswrapper[4789]: I1122 10:24:16.219850 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_b479a8a1-0fba-4a72-b830-2a68e2bd9be4/ovsdbserver-nb/0.log" Nov 22 10:24:16 crc kubenswrapper[4789]: I1122 10:24:16.303788 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_588fc366-e2d8-40bc-b09e-ebcf032e440f/nova-metadata-metadata/0.log" Nov 22 10:24:16 crc kubenswrapper[4789]: I1122 10:24:16.396800 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_c2d44daa-85ba-473a-a29f-2f2587bfd2fa/openstack-network-exporter/0.log" Nov 22 10:24:16 crc kubenswrapper[4789]: I1122 10:24:16.439953 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_c2d44daa-85ba-473a-a29f-2f2587bfd2fa/ovsdbserver-sb/0.log" Nov 22 10:24:16 crc kubenswrapper[4789]: I1122 10:24:16.748152 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_7ab63a4d-2e35-4b93-803b-13a38a43b8d6/setup-container/0.log" Nov 22 10:24:16 crc kubenswrapper[4789]: I1122 10:24:16.898363 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_7ab63a4d-2e35-4b93-803b-13a38a43b8d6/setup-container/0.log" Nov 22 10:24:16 crc kubenswrapper[4789]: I1122 10:24:16.926949 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_7ab63a4d-2e35-4b93-803b-13a38a43b8d6/rabbitmq/0.log" Nov 22 10:24:16 crc kubenswrapper[4789]: I1122 10:24:16.949990 4789 generic.go:334] "Generic (PLEG): container finished" podID="8dce9e8b-6c82-44b6-98fe-72555fd207fe" containerID="1e7f483acd37076489e62984fc6dbfd11663f3747c2d3dad2b7b0815098aee14" exitCode=0 Nov 22 10:24:16 crc kubenswrapper[4789]: I1122 10:24:16.950036 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kzmrw" event={"ID":"8dce9e8b-6c82-44b6-98fe-72555fd207fe","Type":"ContainerDied","Data":"1e7f483acd37076489e62984fc6dbfd11663f3747c2d3dad2b7b0815098aee14"} Nov 22 10:24:16 crc kubenswrapper[4789]: I1122 10:24:16.950061 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kzmrw" event={"ID":"8dce9e8b-6c82-44b6-98fe-72555fd207fe","Type":"ContainerStarted","Data":"557e0bcf17d6ea5b728842bb927cc67824c91d4422c84553007c72b37faa954c"} Nov 22 10:24:16 crc kubenswrapper[4789]: I1122 10:24:16.952134 4789 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 22 10:24:17 crc kubenswrapper[4789]: I1122 10:24:17.063014 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-7846c8b766-kj5xm_87e6a5f6-fbad-4cdf-bc22-1b6be733812a/placement-api/0.log" Nov 22 10:24:17 crc kubenswrapper[4789]: I1122 10:24:17.113368 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_1f834b47-b9a1-4813-bcac-3b5161eceeac/setup-container/0.log" Nov 22 10:24:17 crc kubenswrapper[4789]: I1122 10:24:17.208679 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-7846c8b766-kj5xm_87e6a5f6-fbad-4cdf-bc22-1b6be733812a/placement-log/0.log" Nov 22 10:24:17 crc kubenswrapper[4789]: I1122 10:24:17.374587 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_1f834b47-b9a1-4813-bcac-3b5161eceeac/rabbitmq/0.log" Nov 22 10:24:17 crc kubenswrapper[4789]: I1122 10:24:17.419775 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-pbfnm_d82458d2-439e-479d-8315-32c793ed9401/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 22 10:24:17 crc kubenswrapper[4789]: I1122 10:24:17.424495 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_1f834b47-b9a1-4813-bcac-3b5161eceeac/setup-container/0.log" Nov 22 10:24:17 crc kubenswrapper[4789]: I1122 10:24:17.617263 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-qmwns_6f03e6f2-254d-4e82-bdda-04a73ae3b36a/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Nov 22 10:24:17 crc kubenswrapper[4789]: I1122 10:24:17.728971 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-bmfgh_b8ccf801-a404-4a95-bda9-5a5569d45a51/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 22 10:24:17 crc kubenswrapper[4789]: I1122 10:24:17.877343 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-h22cd_3901b861-f71d-4272-ba5c-2fc28c9e7972/ssh-known-hosts-edpm-deployment/0.log" Nov 22 10:24:17 crc kubenswrapper[4789]: I1122 10:24:17.998461 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kzmrw" event={"ID":"8dce9e8b-6c82-44b6-98fe-72555fd207fe","Type":"ContainerStarted","Data":"53c3208a697e10684aa3dc5c2652540b4c50fbfc141817dce0891a39c3d3bdcb"} Nov 22 10:24:18 crc kubenswrapper[4789]: I1122 10:24:18.318674 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest-s00-full_18bbf966-a509-4faa-a8de-28045e5c01b4/tempest-tests-tempest-tests-runner/0.log" Nov 22 10:24:18 crc kubenswrapper[4789]: I1122 10:24:18.392264 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest-s01-single-test_21bbc026-9e56-43bc-a3bf-0ec807812dba/tempest-tests-tempest-tests-runner/0.log" Nov 22 10:24:18 crc kubenswrapper[4789]: I1122 10:24:18.581809 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-horizontest-horizontest-tests-horizontest_7d0c8264-b357-4277-ba5e-924871a1d088/test-operator-logs-container/0.log" Nov 22 10:24:18 crc kubenswrapper[4789]: I1122 10:24:18.584390 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-ansibletest-ansibletest-ansibletest_7108f947-1001-4cb3-992f-7e6edf860db4/test-operator-logs-container/0.log" Nov 22 10:24:18 crc kubenswrapper[4789]: I1122 10:24:18.794159 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_b7a4abc5-a120-4e83-ae40-efd239ae25fe/test-operator-logs-container/0.log" Nov 22 10:24:18 crc kubenswrapper[4789]: I1122 10:24:18.836096 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tobiko-tobiko-tests-tobiko_01b8c525-b595-4f88-825a-0eda2c77c604/test-operator-logs-container/0.log" Nov 22 10:24:19 crc kubenswrapper[4789]: I1122 10:24:19.009984 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tobiko-tests-tobiko-s00-podified-functional_68695f3e-75c7-4752-a896-dbec939c0dd6/tobiko-tests-tobiko/0.log" Nov 22 10:24:19 crc kubenswrapper[4789]: I1122 10:24:19.120455 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tobiko-tests-tobiko-s01-sanity_d3fd6f1c-a6a5-406f-9eae-b6627046b275/tobiko-tests-tobiko/0.log" Nov 22 10:24:19 crc kubenswrapper[4789]: I1122 10:24:19.201500 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-n4w68_b8b49b82-f67f-4416-86f1-474272d6fdee/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Nov 22 10:24:20 crc kubenswrapper[4789]: I1122 10:24:20.003413 4789 generic.go:334] "Generic (PLEG): container finished" podID="8dce9e8b-6c82-44b6-98fe-72555fd207fe" containerID="53c3208a697e10684aa3dc5c2652540b4c50fbfc141817dce0891a39c3d3bdcb" exitCode=0 Nov 22 10:24:20 crc kubenswrapper[4789]: I1122 10:24:20.003466 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kzmrw" event={"ID":"8dce9e8b-6c82-44b6-98fe-72555fd207fe","Type":"ContainerDied","Data":"53c3208a697e10684aa3dc5c2652540b4c50fbfc141817dce0891a39c3d3bdcb"} Nov 22 10:24:21 crc kubenswrapper[4789]: I1122 10:24:21.054498 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kzmrw" event={"ID":"8dce9e8b-6c82-44b6-98fe-72555fd207fe","Type":"ContainerStarted","Data":"3ebea3b10892a2014fb49f7f2fa8503463b293f87457d3c795d65009de0cef6d"} Nov 22 10:24:21 crc kubenswrapper[4789]: I1122 10:24:21.098200 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-kzmrw" podStartSLOduration=2.66410183 podStartE2EDuration="6.098166339s" podCreationTimestamp="2025-11-22 10:24:15 +0000 UTC" firstStartedPulling="2025-11-22 10:24:16.951927931 +0000 UTC m=+8871.186328204" lastFinishedPulling="2025-11-22 10:24:20.38599244 +0000 UTC m=+8874.620392713" observedRunningTime="2025-11-22 10:24:21.07887139 +0000 UTC m=+8875.313271663" watchObservedRunningTime="2025-11-22 10:24:21.098166339 +0000 UTC m=+8875.332566612" Nov 22 10:24:25 crc kubenswrapper[4789]: I1122 10:24:25.480825 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-kzmrw" Nov 22 10:24:25 crc kubenswrapper[4789]: I1122 10:24:25.481309 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-kzmrw" Nov 22 10:24:25 crc kubenswrapper[4789]: I1122 10:24:25.533724 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-kzmrw" Nov 22 10:24:26 crc kubenswrapper[4789]: I1122 10:24:26.137465 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-kzmrw" Nov 22 10:24:26 crc kubenswrapper[4789]: I1122 10:24:26.194204 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-kzmrw"] Nov 22 10:24:27 crc kubenswrapper[4789]: I1122 10:24:27.068061 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_bd7daba2-62fe-48dc-8e83-a3764c57f9ba/memcached/0.log" Nov 22 10:24:28 crc kubenswrapper[4789]: I1122 10:24:28.105390 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-kzmrw" podUID="8dce9e8b-6c82-44b6-98fe-72555fd207fe" containerName="registry-server" containerID="cri-o://3ebea3b10892a2014fb49f7f2fa8503463b293f87457d3c795d65009de0cef6d" gracePeriod=2 Nov 22 10:24:28 crc kubenswrapper[4789]: I1122 10:24:28.584651 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kzmrw" Nov 22 10:24:28 crc kubenswrapper[4789]: I1122 10:24:28.617830 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m7gdk\" (UniqueName: \"kubernetes.io/projected/8dce9e8b-6c82-44b6-98fe-72555fd207fe-kube-api-access-m7gdk\") pod \"8dce9e8b-6c82-44b6-98fe-72555fd207fe\" (UID: \"8dce9e8b-6c82-44b6-98fe-72555fd207fe\") " Nov 22 10:24:28 crc kubenswrapper[4789]: I1122 10:24:28.617980 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8dce9e8b-6c82-44b6-98fe-72555fd207fe-utilities\") pod \"8dce9e8b-6c82-44b6-98fe-72555fd207fe\" (UID: \"8dce9e8b-6c82-44b6-98fe-72555fd207fe\") " Nov 22 10:24:28 crc kubenswrapper[4789]: I1122 10:24:28.617998 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8dce9e8b-6c82-44b6-98fe-72555fd207fe-catalog-content\") pod \"8dce9e8b-6c82-44b6-98fe-72555fd207fe\" (UID: \"8dce9e8b-6c82-44b6-98fe-72555fd207fe\") " Nov 22 10:24:28 crc kubenswrapper[4789]: I1122 10:24:28.620789 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8dce9e8b-6c82-44b6-98fe-72555fd207fe-utilities" (OuterVolumeSpecName: "utilities") pod "8dce9e8b-6c82-44b6-98fe-72555fd207fe" (UID: "8dce9e8b-6c82-44b6-98fe-72555fd207fe"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 10:24:28 crc kubenswrapper[4789]: I1122 10:24:28.623798 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8dce9e8b-6c82-44b6-98fe-72555fd207fe-kube-api-access-m7gdk" (OuterVolumeSpecName: "kube-api-access-m7gdk") pod "8dce9e8b-6c82-44b6-98fe-72555fd207fe" (UID: "8dce9e8b-6c82-44b6-98fe-72555fd207fe"). InnerVolumeSpecName "kube-api-access-m7gdk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 10:24:28 crc kubenswrapper[4789]: I1122 10:24:28.679606 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8dce9e8b-6c82-44b6-98fe-72555fd207fe-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8dce9e8b-6c82-44b6-98fe-72555fd207fe" (UID: "8dce9e8b-6c82-44b6-98fe-72555fd207fe"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 10:24:28 crc kubenswrapper[4789]: I1122 10:24:28.720356 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m7gdk\" (UniqueName: \"kubernetes.io/projected/8dce9e8b-6c82-44b6-98fe-72555fd207fe-kube-api-access-m7gdk\") on node \"crc\" DevicePath \"\"" Nov 22 10:24:28 crc kubenswrapper[4789]: I1122 10:24:28.720387 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8dce9e8b-6c82-44b6-98fe-72555fd207fe-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 10:24:28 crc kubenswrapper[4789]: I1122 10:24:28.720397 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8dce9e8b-6c82-44b6-98fe-72555fd207fe-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 10:24:29 crc kubenswrapper[4789]: I1122 10:24:29.114546 4789 generic.go:334] "Generic (PLEG): container finished" podID="8dce9e8b-6c82-44b6-98fe-72555fd207fe" containerID="3ebea3b10892a2014fb49f7f2fa8503463b293f87457d3c795d65009de0cef6d" exitCode=0 Nov 22 10:24:29 crc kubenswrapper[4789]: I1122 10:24:29.114596 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kzmrw" event={"ID":"8dce9e8b-6c82-44b6-98fe-72555fd207fe","Type":"ContainerDied","Data":"3ebea3b10892a2014fb49f7f2fa8503463b293f87457d3c795d65009de0cef6d"} Nov 22 10:24:29 crc kubenswrapper[4789]: I1122 10:24:29.114633 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kzmrw" event={"ID":"8dce9e8b-6c82-44b6-98fe-72555fd207fe","Type":"ContainerDied","Data":"557e0bcf17d6ea5b728842bb927cc67824c91d4422c84553007c72b37faa954c"} Nov 22 10:24:29 crc kubenswrapper[4789]: I1122 10:24:29.114639 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kzmrw" Nov 22 10:24:29 crc kubenswrapper[4789]: I1122 10:24:29.114654 4789 scope.go:117] "RemoveContainer" containerID="3ebea3b10892a2014fb49f7f2fa8503463b293f87457d3c795d65009de0cef6d" Nov 22 10:24:29 crc kubenswrapper[4789]: I1122 10:24:29.135374 4789 scope.go:117] "RemoveContainer" containerID="53c3208a697e10684aa3dc5c2652540b4c50fbfc141817dce0891a39c3d3bdcb" Nov 22 10:24:29 crc kubenswrapper[4789]: I1122 10:24:29.158820 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-kzmrw"] Nov 22 10:24:29 crc kubenswrapper[4789]: I1122 10:24:29.173647 4789 scope.go:117] "RemoveContainer" containerID="1e7f483acd37076489e62984fc6dbfd11663f3747c2d3dad2b7b0815098aee14" Nov 22 10:24:29 crc kubenswrapper[4789]: I1122 10:24:29.176923 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-kzmrw"] Nov 22 10:24:29 crc kubenswrapper[4789]: I1122 10:24:29.224338 4789 scope.go:117] "RemoveContainer" containerID="3ebea3b10892a2014fb49f7f2fa8503463b293f87457d3c795d65009de0cef6d" Nov 22 10:24:29 crc kubenswrapper[4789]: E1122 10:24:29.228378 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3ebea3b10892a2014fb49f7f2fa8503463b293f87457d3c795d65009de0cef6d\": container with ID starting with 3ebea3b10892a2014fb49f7f2fa8503463b293f87457d3c795d65009de0cef6d not found: ID does not exist" containerID="3ebea3b10892a2014fb49f7f2fa8503463b293f87457d3c795d65009de0cef6d" Nov 22 10:24:29 crc kubenswrapper[4789]: I1122 10:24:29.228441 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3ebea3b10892a2014fb49f7f2fa8503463b293f87457d3c795d65009de0cef6d"} err="failed to get container status \"3ebea3b10892a2014fb49f7f2fa8503463b293f87457d3c795d65009de0cef6d\": rpc error: code = NotFound desc = could not find container \"3ebea3b10892a2014fb49f7f2fa8503463b293f87457d3c795d65009de0cef6d\": container with ID starting with 3ebea3b10892a2014fb49f7f2fa8503463b293f87457d3c795d65009de0cef6d not found: ID does not exist" Nov 22 10:24:29 crc kubenswrapper[4789]: I1122 10:24:29.228466 4789 scope.go:117] "RemoveContainer" containerID="53c3208a697e10684aa3dc5c2652540b4c50fbfc141817dce0891a39c3d3bdcb" Nov 22 10:24:29 crc kubenswrapper[4789]: E1122 10:24:29.228863 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"53c3208a697e10684aa3dc5c2652540b4c50fbfc141817dce0891a39c3d3bdcb\": container with ID starting with 53c3208a697e10684aa3dc5c2652540b4c50fbfc141817dce0891a39c3d3bdcb not found: ID does not exist" containerID="53c3208a697e10684aa3dc5c2652540b4c50fbfc141817dce0891a39c3d3bdcb" Nov 22 10:24:29 crc kubenswrapper[4789]: I1122 10:24:29.228901 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53c3208a697e10684aa3dc5c2652540b4c50fbfc141817dce0891a39c3d3bdcb"} err="failed to get container status \"53c3208a697e10684aa3dc5c2652540b4c50fbfc141817dce0891a39c3d3bdcb\": rpc error: code = NotFound desc = could not find container \"53c3208a697e10684aa3dc5c2652540b4c50fbfc141817dce0891a39c3d3bdcb\": container with ID starting with 53c3208a697e10684aa3dc5c2652540b4c50fbfc141817dce0891a39c3d3bdcb not found: ID does not exist" Nov 22 10:24:29 crc kubenswrapper[4789]: I1122 10:24:29.228929 4789 scope.go:117] "RemoveContainer" containerID="1e7f483acd37076489e62984fc6dbfd11663f3747c2d3dad2b7b0815098aee14" Nov 22 10:24:29 crc kubenswrapper[4789]: E1122 10:24:29.230684 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1e7f483acd37076489e62984fc6dbfd11663f3747c2d3dad2b7b0815098aee14\": container with ID starting with 1e7f483acd37076489e62984fc6dbfd11663f3747c2d3dad2b7b0815098aee14 not found: ID does not exist" containerID="1e7f483acd37076489e62984fc6dbfd11663f3747c2d3dad2b7b0815098aee14" Nov 22 10:24:29 crc kubenswrapper[4789]: I1122 10:24:29.230715 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e7f483acd37076489e62984fc6dbfd11663f3747c2d3dad2b7b0815098aee14"} err="failed to get container status \"1e7f483acd37076489e62984fc6dbfd11663f3747c2d3dad2b7b0815098aee14\": rpc error: code = NotFound desc = could not find container \"1e7f483acd37076489e62984fc6dbfd11663f3747c2d3dad2b7b0815098aee14\": container with ID starting with 1e7f483acd37076489e62984fc6dbfd11663f3747c2d3dad2b7b0815098aee14 not found: ID does not exist" Nov 22 10:24:29 crc kubenswrapper[4789]: I1122 10:24:29.983011 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8dce9e8b-6c82-44b6-98fe-72555fd207fe" path="/var/lib/kubelet/pods/8dce9e8b-6c82-44b6-98fe-72555fd207fe/volumes" Nov 22 10:24:43 crc kubenswrapper[4789]: I1122 10:24:43.275314 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287wk9d5_aeb88caf-4a0d-450a-9646-36471be5fd49/util/0.log" Nov 22 10:24:43 crc kubenswrapper[4789]: I1122 10:24:43.462001 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287wk9d5_aeb88caf-4a0d-450a-9646-36471be5fd49/pull/0.log" Nov 22 10:24:43 crc kubenswrapper[4789]: I1122 10:24:43.482055 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287wk9d5_aeb88caf-4a0d-450a-9646-36471be5fd49/pull/0.log" Nov 22 10:24:43 crc kubenswrapper[4789]: I1122 10:24:43.494499 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287wk9d5_aeb88caf-4a0d-450a-9646-36471be5fd49/util/0.log" Nov 22 10:24:43 crc kubenswrapper[4789]: I1122 10:24:43.621677 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287wk9d5_aeb88caf-4a0d-450a-9646-36471be5fd49/util/0.log" Nov 22 10:24:43 crc kubenswrapper[4789]: I1122 10:24:43.644830 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287wk9d5_aeb88caf-4a0d-450a-9646-36471be5fd49/pull/0.log" Nov 22 10:24:43 crc kubenswrapper[4789]: I1122 10:24:43.661960 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287wk9d5_aeb88caf-4a0d-450a-9646-36471be5fd49/extract/0.log" Nov 22 10:24:43 crc kubenswrapper[4789]: I1122 10:24:43.786550 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7768f8c84f-hwbvw_93c2cbf6-757a-4de1-9f83-115787d74f31/kube-rbac-proxy/0.log" Nov 22 10:24:43 crc kubenswrapper[4789]: I1122 10:24:43.898744 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7768f8c84f-hwbvw_93c2cbf6-757a-4de1-9f83-115787d74f31/manager/0.log" Nov 22 10:24:43 crc kubenswrapper[4789]: I1122 10:24:43.946992 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6d8fd67bf7-2klns_55d6892d-7d96-4518-8d36-58048cf96c52/kube-rbac-proxy/0.log" Nov 22 10:24:44 crc kubenswrapper[4789]: I1122 10:24:44.033337 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6d8fd67bf7-2klns_55d6892d-7d96-4518-8d36-58048cf96c52/manager/0.log" Nov 22 10:24:44 crc kubenswrapper[4789]: I1122 10:24:44.077510 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-56dfb6b67f-p4vmr_77bd1957-6cd6-418b-9b2c-ac879bbca15e/kube-rbac-proxy/0.log" Nov 22 10:24:44 crc kubenswrapper[4789]: I1122 10:24:44.100114 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-56dfb6b67f-p4vmr_77bd1957-6cd6-418b-9b2c-ac879bbca15e/manager/0.log" Nov 22 10:24:44 crc kubenswrapper[4789]: I1122 10:24:44.259789 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-8667fbf6f6-k5kms_5c6cfa2e-ba14-48da-a43d-6ea46a90bad1/kube-rbac-proxy/0.log" Nov 22 10:24:44 crc kubenswrapper[4789]: I1122 10:24:44.344490 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-8667fbf6f6-k5kms_5c6cfa2e-ba14-48da-a43d-6ea46a90bad1/manager/0.log" Nov 22 10:24:44 crc kubenswrapper[4789]: I1122 10:24:44.405987 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-bf4c6585d-6shgh_0cf958d5-222e-4598-a1a7-fe887644cd24/kube-rbac-proxy/0.log" Nov 22 10:24:44 crc kubenswrapper[4789]: I1122 10:24:44.437796 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-bf4c6585d-6shgh_0cf958d5-222e-4598-a1a7-fe887644cd24/manager/0.log" Nov 22 10:24:44 crc kubenswrapper[4789]: I1122 10:24:44.511855 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-5d86b44686-nk4zk_9e20512a-a64f-4f47-a7ef-b91f24390aef/kube-rbac-proxy/0.log" Nov 22 10:24:44 crc kubenswrapper[4789]: I1122 10:24:44.561301 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-5d86b44686-nk4zk_9e20512a-a64f-4f47-a7ef-b91f24390aef/manager/0.log" Nov 22 10:24:44 crc kubenswrapper[4789]: I1122 10:24:44.679631 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-769d9c7585-f7bd8_84fafb83-aaff-4197-90d1-578c74da9299/kube-rbac-proxy/0.log" Nov 22 10:24:44 crc kubenswrapper[4789]: I1122 10:24:44.817313 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-5c75d7c94b-66wlz_29d2daa3-90b3-4dea-9612-edc4cc0dcc30/kube-rbac-proxy/0.log" Nov 22 10:24:44 crc kubenswrapper[4789]: I1122 10:24:44.824539 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-769d9c7585-f7bd8_84fafb83-aaff-4197-90d1-578c74da9299/manager/0.log" Nov 22 10:24:44 crc kubenswrapper[4789]: I1122 10:24:44.867715 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-5c75d7c94b-66wlz_29d2daa3-90b3-4dea-9612-edc4cc0dcc30/manager/0.log" Nov 22 10:24:44 crc kubenswrapper[4789]: I1122 10:24:44.976647 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7879fb76fd-zn5xp_490be79a-ce0c-4a3e-a819-7ef482fe453d/kube-rbac-proxy/0.log" Nov 22 10:24:45 crc kubenswrapper[4789]: I1122 10:24:45.030359 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7879fb76fd-zn5xp_490be79a-ce0c-4a3e-a819-7ef482fe453d/manager/0.log" Nov 22 10:24:45 crc kubenswrapper[4789]: I1122 10:24:45.124565 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7bb88cb858-9g8lk_50571e3f-4e49-47b6-aa7f-2c714e8ef88f/kube-rbac-proxy/0.log" Nov 22 10:24:45 crc kubenswrapper[4789]: I1122 10:24:45.233288 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7bb88cb858-9g8lk_50571e3f-4e49-47b6-aa7f-2c714e8ef88f/manager/0.log" Nov 22 10:24:45 crc kubenswrapper[4789]: I1122 10:24:45.249089 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-6f8c5b86cb-rrcjv_c4962710-54a6-4538-90ce-be3e27062015/kube-rbac-proxy/0.log" Nov 22 10:24:45 crc kubenswrapper[4789]: I1122 10:24:45.334028 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-6f8c5b86cb-rrcjv_c4962710-54a6-4538-90ce-be3e27062015/manager/0.log" Nov 22 10:24:45 crc kubenswrapper[4789]: I1122 10:24:45.399668 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-66b7d6f598-hn78m_c4a79bb9-1d05-4c14-a668-6f63f521c98d/kube-rbac-proxy/0.log" Nov 22 10:24:45 crc kubenswrapper[4789]: I1122 10:24:45.487898 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-66b7d6f598-hn78m_c4a79bb9-1d05-4c14-a668-6f63f521c98d/manager/0.log" Nov 22 10:24:45 crc kubenswrapper[4789]: I1122 10:24:45.598494 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-86d796d84d-4lzpn_2ec2f4f5-6ca9-4dae-ae2e-655ef780f3dd/kube-rbac-proxy/0.log" Nov 22 10:24:45 crc kubenswrapper[4789]: I1122 10:24:45.655199 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-86d796d84d-4lzpn_2ec2f4f5-6ca9-4dae-ae2e-655ef780f3dd/manager/0.log" Nov 22 10:24:45 crc kubenswrapper[4789]: I1122 10:24:45.749826 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-6fdc856c5d-mccpp_bc429681-3e2a-4318-870d-ef62aef7d55b/kube-rbac-proxy/0.log" Nov 22 10:24:45 crc kubenswrapper[4789]: I1122 10:24:45.806358 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-6fdc856c5d-mccpp_bc429681-3e2a-4318-870d-ef62aef7d55b/manager/0.log" Nov 22 10:24:45 crc kubenswrapper[4789]: I1122 10:24:45.869302 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-79d88dcd44kf66v_d42b84d7-0575-401f-b556-fdaa161588c9/kube-rbac-proxy/0.log" Nov 22 10:24:45 crc kubenswrapper[4789]: I1122 10:24:45.961270 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-79d88dcd44kf66v_d42b84d7-0575-401f-b556-fdaa161588c9/manager/0.log" Nov 22 10:24:46 crc kubenswrapper[4789]: I1122 10:24:46.007081 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-6cb9dc54f8-fzmzt_03652378-7433-48c7-a601-aea6d67f1a14/kube-rbac-proxy/0.log" Nov 22 10:24:46 crc kubenswrapper[4789]: I1122 10:24:46.210559 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-55c966b9c6-psmvq_5729bc29-a8c9-4e1c-93fe-4f1532823cb9/kube-rbac-proxy/0.log" Nov 22 10:24:46 crc kubenswrapper[4789]: I1122 10:24:46.303501 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-55c966b9c6-psmvq_5729bc29-a8c9-4e1c-93fe-4f1532823cb9/operator/0.log" Nov 22 10:24:46 crc kubenswrapper[4789]: I1122 10:24:46.647809 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-9z5kz_20a926be-0d41-41ef-bb6c-70e2d19ec270/registry-server/0.log" Nov 22 10:24:46 crc kubenswrapper[4789]: I1122 10:24:46.723224 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-5bdf4f7f7f-29chq_4cbc0d4d-5b2e-4678-84e9-7eb69b3742e0/kube-rbac-proxy/0.log" Nov 22 10:24:46 crc kubenswrapper[4789]: I1122 10:24:46.785454 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-5bdf4f7f7f-29chq_4cbc0d4d-5b2e-4678-84e9-7eb69b3742e0/manager/0.log" Nov 22 10:24:46 crc kubenswrapper[4789]: I1122 10:24:46.917183 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-6dc664666c-p26lg_a41e302b-afc7-45bc-8694-29e5ea87edc8/manager/0.log" Nov 22 10:24:46 crc kubenswrapper[4789]: I1122 10:24:46.945005 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-6dc664666c-p26lg_a41e302b-afc7-45bc-8694-29e5ea87edc8/kube-rbac-proxy/0.log" Nov 22 10:24:47 crc kubenswrapper[4789]: I1122 10:24:47.116918 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-77vzm_c94020a6-11c7-4ad5-90bb-8dd40e08ffd2/operator/0.log" Nov 22 10:24:47 crc kubenswrapper[4789]: I1122 10:24:47.163414 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-799cb6ffd6-h272x_0478f782-f1c5-412e-9964-749ee7f83c58/kube-rbac-proxy/0.log" Nov 22 10:24:47 crc kubenswrapper[4789]: I1122 10:24:47.200628 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-6cb9dc54f8-fzmzt_03652378-7433-48c7-a601-aea6d67f1a14/manager/0.log" Nov 22 10:24:47 crc kubenswrapper[4789]: I1122 10:24:47.237626 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-799cb6ffd6-h272x_0478f782-f1c5-412e-9964-749ee7f83c58/manager/0.log" Nov 22 10:24:47 crc kubenswrapper[4789]: I1122 10:24:47.343123 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-7798859c74-j9vct_e79f2bb5-95ac-4fbc-a065-da8067f25cb4/kube-rbac-proxy/0.log" Nov 22 10:24:47 crc kubenswrapper[4789]: I1122 10:24:47.416767 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-7798859c74-j9vct_e79f2bb5-95ac-4fbc-a065-da8067f25cb4/manager/0.log" Nov 22 10:24:47 crc kubenswrapper[4789]: I1122 10:24:47.419058 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-7b6cddbf79-2d282_f3873dc8-add1-47e4-9fee-e1023782e6da/kube-rbac-proxy/0.log" Nov 22 10:24:47 crc kubenswrapper[4789]: I1122 10:24:47.499120 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-7b6cddbf79-2d282_f3873dc8-add1-47e4-9fee-e1023782e6da/manager/1.log" Nov 22 10:24:47 crc kubenswrapper[4789]: I1122 10:24:47.581554 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-7b6cddbf79-2d282_f3873dc8-add1-47e4-9fee-e1023782e6da/manager/0.log" Nov 22 10:24:47 crc kubenswrapper[4789]: I1122 10:24:47.604450 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-7cd4fb6f79-kcq2g_ca7faed1-038d-4ec8-800d-35848427a921/kube-rbac-proxy/0.log" Nov 22 10:24:47 crc kubenswrapper[4789]: I1122 10:24:47.647790 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-7cd4fb6f79-kcq2g_ca7faed1-038d-4ec8-800d-35848427a921/manager/0.log" Nov 22 10:25:03 crc kubenswrapper[4789]: E1122 10:25:03.966177 4789 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Nov 22 10:25:04 crc kubenswrapper[4789]: I1122 10:25:04.062121 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-fr4zx_10112213-cdd9-4bab-a7d5-bc7a3fab5a48/control-plane-machine-set-operator/0.log" Nov 22 10:25:04 crc kubenswrapper[4789]: I1122 10:25:04.186334 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-spnzw_d086acbe-d982-4880-9538-04a184cb4148/kube-rbac-proxy/0.log" Nov 22 10:25:04 crc kubenswrapper[4789]: I1122 10:25:04.203620 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-spnzw_d086acbe-d982-4880-9538-04a184cb4148/machine-api-operator/0.log" Nov 22 10:25:17 crc kubenswrapper[4789]: I1122 10:25:17.496033 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-d6qv9_6dd4b23f-fd7b-4893-850b-ae8d269b4006/cert-manager-controller/0.log" Nov 22 10:25:17 crc kubenswrapper[4789]: I1122 10:25:17.652233 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-jn69f_d8d7bb44-57f4-4daf-a00f-13e0209ecf92/cert-manager-cainjector/0.log" Nov 22 10:25:17 crc kubenswrapper[4789]: I1122 10:25:17.689532 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-txzv2_ca83ef6b-884d-44a3-9570-8088d240e4e8/cert-manager-webhook/0.log" Nov 22 10:25:30 crc kubenswrapper[4789]: I1122 10:25:30.628917 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-5874bd7bc5-pfnkv_c88315b2-ec94-4272-91f1-3127418ae55c/nmstate-console-plugin/0.log" Nov 22 10:25:30 crc kubenswrapper[4789]: I1122 10:25:30.795308 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-hccc7_d94d3cb7-9188-42a2-b911-4f7665349830/nmstate-handler/0.log" Nov 22 10:25:30 crc kubenswrapper[4789]: I1122 10:25:30.850817 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-5dcf9c57c5-t6j77_15f327f3-5cc9-4ce4-b291-efdffea58145/kube-rbac-proxy/0.log" Nov 22 10:25:30 crc kubenswrapper[4789]: I1122 10:25:30.863701 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-5dcf9c57c5-t6j77_15f327f3-5cc9-4ce4-b291-efdffea58145/nmstate-metrics/0.log" Nov 22 10:25:31 crc kubenswrapper[4789]: I1122 10:25:31.015104 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-557fdffb88-cfdtx_42d193be-d22b-4731-baed-4773d41c095d/nmstate-operator/0.log" Nov 22 10:25:31 crc kubenswrapper[4789]: I1122 10:25:31.063479 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6b89b748d8-59p65_4dc70724-5816-4152-8d2f-763faeddd54f/nmstate-webhook/0.log" Nov 22 10:25:35 crc kubenswrapper[4789]: I1122 10:25:35.372469 4789 patch_prober.go:28] interesting pod/machine-config-daemon-pxvrm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 10:25:35 crc kubenswrapper[4789]: I1122 10:25:35.373103 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 10:25:46 crc kubenswrapper[4789]: I1122 10:25:46.438524 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6c7b4b5f48-wm2pz_1f4fdc0e-ffe3-4a34-a9b4-0b4c70fccefa/kube-rbac-proxy/0.log" Nov 22 10:25:46 crc kubenswrapper[4789]: I1122 10:25:46.549890 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6c7b4b5f48-wm2pz_1f4fdc0e-ffe3-4a34-a9b4-0b4c70fccefa/controller/0.log" Nov 22 10:25:46 crc kubenswrapper[4789]: I1122 10:25:46.652041 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qwjlg_4b633875-b147-4ac7-979c-07a76bedb3fb/cp-frr-files/0.log" Nov 22 10:25:46 crc kubenswrapper[4789]: I1122 10:25:46.764620 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qwjlg_4b633875-b147-4ac7-979c-07a76bedb3fb/cp-frr-files/0.log" Nov 22 10:25:46 crc kubenswrapper[4789]: I1122 10:25:46.820270 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qwjlg_4b633875-b147-4ac7-979c-07a76bedb3fb/cp-reloader/0.log" Nov 22 10:25:46 crc kubenswrapper[4789]: I1122 10:25:46.844050 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qwjlg_4b633875-b147-4ac7-979c-07a76bedb3fb/cp-metrics/0.log" Nov 22 10:25:46 crc kubenswrapper[4789]: I1122 10:25:46.854816 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qwjlg_4b633875-b147-4ac7-979c-07a76bedb3fb/cp-reloader/0.log" Nov 22 10:25:47 crc kubenswrapper[4789]: I1122 10:25:47.004995 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qwjlg_4b633875-b147-4ac7-979c-07a76bedb3fb/cp-frr-files/0.log" Nov 22 10:25:47 crc kubenswrapper[4789]: I1122 10:25:47.013780 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qwjlg_4b633875-b147-4ac7-979c-07a76bedb3fb/cp-reloader/0.log" Nov 22 10:25:47 crc kubenswrapper[4789]: I1122 10:25:47.057196 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qwjlg_4b633875-b147-4ac7-979c-07a76bedb3fb/cp-metrics/0.log" Nov 22 10:25:47 crc kubenswrapper[4789]: I1122 10:25:47.073516 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qwjlg_4b633875-b147-4ac7-979c-07a76bedb3fb/cp-metrics/0.log" Nov 22 10:25:47 crc kubenswrapper[4789]: I1122 10:25:47.226390 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qwjlg_4b633875-b147-4ac7-979c-07a76bedb3fb/cp-reloader/0.log" Nov 22 10:25:47 crc kubenswrapper[4789]: I1122 10:25:47.249238 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qwjlg_4b633875-b147-4ac7-979c-07a76bedb3fb/cp-frr-files/0.log" Nov 22 10:25:47 crc kubenswrapper[4789]: I1122 10:25:47.263151 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qwjlg_4b633875-b147-4ac7-979c-07a76bedb3fb/controller/0.log" Nov 22 10:25:47 crc kubenswrapper[4789]: I1122 10:25:47.273427 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qwjlg_4b633875-b147-4ac7-979c-07a76bedb3fb/cp-metrics/0.log" Nov 22 10:25:47 crc kubenswrapper[4789]: I1122 10:25:47.393057 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qwjlg_4b633875-b147-4ac7-979c-07a76bedb3fb/frr-metrics/0.log" Nov 22 10:25:47 crc kubenswrapper[4789]: I1122 10:25:47.442585 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qwjlg_4b633875-b147-4ac7-979c-07a76bedb3fb/kube-rbac-proxy/0.log" Nov 22 10:25:47 crc kubenswrapper[4789]: I1122 10:25:47.459803 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qwjlg_4b633875-b147-4ac7-979c-07a76bedb3fb/kube-rbac-proxy-frr/0.log" Nov 22 10:25:47 crc kubenswrapper[4789]: I1122 10:25:47.621575 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qwjlg_4b633875-b147-4ac7-979c-07a76bedb3fb/reloader/0.log" Nov 22 10:25:47 crc kubenswrapper[4789]: I1122 10:25:47.692469 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-6998585d5-4shz4_eda37095-1c8e-49ab-b085-20c8cf688be0/frr-k8s-webhook-server/0.log" Nov 22 10:25:47 crc kubenswrapper[4789]: I1122 10:25:47.843676 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-686d48fbd8-6jwbz_701af49e-3c52-45f0-8d1e-c32b8c606b36/manager/0.log" Nov 22 10:25:48 crc kubenswrapper[4789]: I1122 10:25:48.009329 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-6f6d9fb676-r76lj_39bf0884-9c38-4839-9f2c-0a7502e961f8/webhook-server/0.log" Nov 22 10:25:48 crc kubenswrapper[4789]: I1122 10:25:48.193031 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-6tqgm_6a534620-1ee2-419c-ae0d-cdf2495f1ea0/kube-rbac-proxy/0.log" Nov 22 10:25:48 crc kubenswrapper[4789]: I1122 10:25:48.812572 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-6tqgm_6a534620-1ee2-419c-ae0d-cdf2495f1ea0/speaker/0.log" Nov 22 10:25:49 crc kubenswrapper[4789]: I1122 10:25:49.492200 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qwjlg_4b633875-b147-4ac7-979c-07a76bedb3fb/frr/0.log" Nov 22 10:26:05 crc kubenswrapper[4789]: I1122 10:26:05.172273 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ewkwwn_a4bfa2cc-45a0-4c18-9a13-19fd8cc06420/util/0.log" Nov 22 10:26:05 crc kubenswrapper[4789]: I1122 10:26:05.351479 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ewkwwn_a4bfa2cc-45a0-4c18-9a13-19fd8cc06420/util/0.log" Nov 22 10:26:05 crc kubenswrapper[4789]: I1122 10:26:05.353473 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ewkwwn_a4bfa2cc-45a0-4c18-9a13-19fd8cc06420/pull/0.log" Nov 22 10:26:05 crc kubenswrapper[4789]: I1122 10:26:05.372623 4789 patch_prober.go:28] interesting pod/machine-config-daemon-pxvrm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 10:26:05 crc kubenswrapper[4789]: I1122 10:26:05.372679 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 10:26:05 crc kubenswrapper[4789]: I1122 10:26:05.405955 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ewkwwn_a4bfa2cc-45a0-4c18-9a13-19fd8cc06420/pull/0.log" Nov 22 10:26:05 crc kubenswrapper[4789]: I1122 10:26:05.554167 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ewkwwn_a4bfa2cc-45a0-4c18-9a13-19fd8cc06420/util/0.log" Nov 22 10:26:05 crc kubenswrapper[4789]: I1122 10:26:05.586656 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ewkwwn_a4bfa2cc-45a0-4c18-9a13-19fd8cc06420/extract/0.log" Nov 22 10:26:05 crc kubenswrapper[4789]: I1122 10:26:05.587027 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ewkwwn_a4bfa2cc-45a0-4c18-9a13-19fd8cc06420/pull/0.log" Nov 22 10:26:05 crc kubenswrapper[4789]: I1122 10:26:05.745527 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-442k7_a74f8c2f-6d6f-4529-ae51-d72146b30b10/extract-utilities/0.log" Nov 22 10:26:05 crc kubenswrapper[4789]: I1122 10:26:05.897457 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-442k7_a74f8c2f-6d6f-4529-ae51-d72146b30b10/extract-utilities/0.log" Nov 22 10:26:05 crc kubenswrapper[4789]: I1122 10:26:05.903355 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-442k7_a74f8c2f-6d6f-4529-ae51-d72146b30b10/extract-content/0.log" Nov 22 10:26:05 crc kubenswrapper[4789]: I1122 10:26:05.929004 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-442k7_a74f8c2f-6d6f-4529-ae51-d72146b30b10/extract-content/0.log" Nov 22 10:26:06 crc kubenswrapper[4789]: I1122 10:26:06.088042 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-442k7_a74f8c2f-6d6f-4529-ae51-d72146b30b10/extract-utilities/0.log" Nov 22 10:26:06 crc kubenswrapper[4789]: I1122 10:26:06.101005 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-442k7_a74f8c2f-6d6f-4529-ae51-d72146b30b10/extract-content/0.log" Nov 22 10:26:06 crc kubenswrapper[4789]: I1122 10:26:06.361241 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rpzdx_60fde5c9-9cd4-4966-8183-01e3c24970b0/extract-utilities/0.log" Nov 22 10:26:06 crc kubenswrapper[4789]: I1122 10:26:06.576671 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rpzdx_60fde5c9-9cd4-4966-8183-01e3c24970b0/extract-utilities/0.log" Nov 22 10:26:06 crc kubenswrapper[4789]: I1122 10:26:06.629192 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rpzdx_60fde5c9-9cd4-4966-8183-01e3c24970b0/extract-content/0.log" Nov 22 10:26:06 crc kubenswrapper[4789]: I1122 10:26:06.651621 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rpzdx_60fde5c9-9cd4-4966-8183-01e3c24970b0/extract-content/0.log" Nov 22 10:26:06 crc kubenswrapper[4789]: I1122 10:26:06.812186 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rpzdx_60fde5c9-9cd4-4966-8183-01e3c24970b0/extract-utilities/0.log" Nov 22 10:26:06 crc kubenswrapper[4789]: I1122 10:26:06.860966 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rpzdx_60fde5c9-9cd4-4966-8183-01e3c24970b0/extract-content/0.log" Nov 22 10:26:07 crc kubenswrapper[4789]: I1122 10:26:07.099805 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6vdrbs_8f9ce070-1816-46d5-a16f-b3b8d1a1f8f3/util/0.log" Nov 22 10:26:07 crc kubenswrapper[4789]: I1122 10:26:07.146971 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-442k7_a74f8c2f-6d6f-4529-ae51-d72146b30b10/registry-server/0.log" Nov 22 10:26:07 crc kubenswrapper[4789]: I1122 10:26:07.342844 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6vdrbs_8f9ce070-1816-46d5-a16f-b3b8d1a1f8f3/pull/0.log" Nov 22 10:26:07 crc kubenswrapper[4789]: I1122 10:26:07.361603 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6vdrbs_8f9ce070-1816-46d5-a16f-b3b8d1a1f8f3/pull/0.log" Nov 22 10:26:07 crc kubenswrapper[4789]: I1122 10:26:07.371101 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6vdrbs_8f9ce070-1816-46d5-a16f-b3b8d1a1f8f3/util/0.log" Nov 22 10:26:07 crc kubenswrapper[4789]: I1122 10:26:07.582261 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6vdrbs_8f9ce070-1816-46d5-a16f-b3b8d1a1f8f3/util/0.log" Nov 22 10:26:07 crc kubenswrapper[4789]: I1122 10:26:07.627453 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6vdrbs_8f9ce070-1816-46d5-a16f-b3b8d1a1f8f3/extract/0.log" Nov 22 10:26:07 crc kubenswrapper[4789]: I1122 10:26:07.635569 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6vdrbs_8f9ce070-1816-46d5-a16f-b3b8d1a1f8f3/pull/0.log" Nov 22 10:26:07 crc kubenswrapper[4789]: I1122 10:26:07.849982 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-2tqfp_279eae2d-e07b-4a6e-984f-7348f2a7641a/marketplace-operator/0.log" Nov 22 10:26:07 crc kubenswrapper[4789]: I1122 10:26:07.916715 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-b96v6_c6da9b7c-c7b1-4a6c-afac-525ffaa3b204/extract-utilities/0.log" Nov 22 10:26:08 crc kubenswrapper[4789]: I1122 10:26:08.105574 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rpzdx_60fde5c9-9cd4-4966-8183-01e3c24970b0/registry-server/0.log" Nov 22 10:26:08 crc kubenswrapper[4789]: I1122 10:26:08.148451 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-b96v6_c6da9b7c-c7b1-4a6c-afac-525ffaa3b204/extract-utilities/0.log" Nov 22 10:26:08 crc kubenswrapper[4789]: I1122 10:26:08.151922 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-b96v6_c6da9b7c-c7b1-4a6c-afac-525ffaa3b204/extract-content/0.log" Nov 22 10:26:08 crc kubenswrapper[4789]: I1122 10:26:08.209799 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-b96v6_c6da9b7c-c7b1-4a6c-afac-525ffaa3b204/extract-content/0.log" Nov 22 10:26:08 crc kubenswrapper[4789]: I1122 10:26:08.338338 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-b96v6_c6da9b7c-c7b1-4a6c-afac-525ffaa3b204/extract-content/0.log" Nov 22 10:26:08 crc kubenswrapper[4789]: I1122 10:26:08.391548 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-b96v6_c6da9b7c-c7b1-4a6c-afac-525ffaa3b204/extract-utilities/0.log" Nov 22 10:26:08 crc kubenswrapper[4789]: I1122 10:26:08.407907 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-z9cvt_a253f042-14c5-4345-8072-55ab86201242/extract-utilities/0.log" Nov 22 10:26:08 crc kubenswrapper[4789]: I1122 10:26:08.635849 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-z9cvt_a253f042-14c5-4345-8072-55ab86201242/extract-utilities/0.log" Nov 22 10:26:08 crc kubenswrapper[4789]: I1122 10:26:08.651820 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-b96v6_c6da9b7c-c7b1-4a6c-afac-525ffaa3b204/registry-server/0.log" Nov 22 10:26:08 crc kubenswrapper[4789]: I1122 10:26:08.685909 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-z9cvt_a253f042-14c5-4345-8072-55ab86201242/extract-content/0.log" Nov 22 10:26:08 crc kubenswrapper[4789]: I1122 10:26:08.687949 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-z9cvt_a253f042-14c5-4345-8072-55ab86201242/extract-content/0.log" Nov 22 10:26:08 crc kubenswrapper[4789]: I1122 10:26:08.871361 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-z9cvt_a253f042-14c5-4345-8072-55ab86201242/extract-content/0.log" Nov 22 10:26:08 crc kubenswrapper[4789]: I1122 10:26:08.874460 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-z9cvt_a253f042-14c5-4345-8072-55ab86201242/extract-utilities/0.log" Nov 22 10:26:09 crc kubenswrapper[4789]: I1122 10:26:09.870859 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-z9cvt_a253f042-14c5-4345-8072-55ab86201242/registry-server/0.log" Nov 22 10:26:32 crc kubenswrapper[4789]: E1122 10:26:32.965735 4789 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Nov 22 10:26:35 crc kubenswrapper[4789]: I1122 10:26:35.371984 4789 patch_prober.go:28] interesting pod/machine-config-daemon-pxvrm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 10:26:35 crc kubenswrapper[4789]: I1122 10:26:35.372610 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 10:26:35 crc kubenswrapper[4789]: I1122 10:26:35.372666 4789 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" Nov 22 10:26:35 crc kubenswrapper[4789]: I1122 10:26:35.373601 4789 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"757306d657b53911c979b1a984e697106c927039cb3f24d4ba4a2a51a772a120"} pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 22 10:26:35 crc kubenswrapper[4789]: I1122 10:26:35.373671 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" containerID="cri-o://757306d657b53911c979b1a984e697106c927039cb3f24d4ba4a2a51a772a120" gracePeriod=600 Nov 22 10:26:35 crc kubenswrapper[4789]: I1122 10:26:35.586097 4789 generic.go:334] "Generic (PLEG): container finished" podID="58aa8071-7ecc-4692-8789-2db04ec70647" containerID="757306d657b53911c979b1a984e697106c927039cb3f24d4ba4a2a51a772a120" exitCode=0 Nov 22 10:26:35 crc kubenswrapper[4789]: I1122 10:26:35.586149 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" event={"ID":"58aa8071-7ecc-4692-8789-2db04ec70647","Type":"ContainerDied","Data":"757306d657b53911c979b1a984e697106c927039cb3f24d4ba4a2a51a772a120"} Nov 22 10:26:35 crc kubenswrapper[4789]: I1122 10:26:35.586268 4789 scope.go:117] "RemoveContainer" containerID="9e4b9858fc97b3d62c7338b8fc7e6fb3cc0a7ad892259a591c7f692014f9eac6" Nov 22 10:26:36 crc kubenswrapper[4789]: I1122 10:26:36.596857 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" event={"ID":"58aa8071-7ecc-4692-8789-2db04ec70647","Type":"ContainerStarted","Data":"3590e488179d593a3c8981249b3f907f5a30f296b7d9ba57504c20a09c6eb5bc"} Nov 22 10:26:42 crc kubenswrapper[4789]: E1122 10:26:42.534667 4789 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.129.56.56:55310->38.129.56.56:37377: write tcp 38.129.56.56:55310->38.129.56.56:37377: write: broken pipe Nov 22 10:27:48 crc kubenswrapper[4789]: E1122 10:27:48.965392 4789 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Nov 22 10:28:20 crc kubenswrapper[4789]: I1122 10:28:20.752697 4789 generic.go:334] "Generic (PLEG): container finished" podID="3d58dd27-067a-4510-b219-340217ca2ecb" containerID="d9f47947cfef03169cfbf6edde2ebc8c5aac347b5c132221e3678b4c50f81740" exitCode=0 Nov 22 10:28:20 crc kubenswrapper[4789]: I1122 10:28:20.752834 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-69pj5/must-gather-shff9" event={"ID":"3d58dd27-067a-4510-b219-340217ca2ecb","Type":"ContainerDied","Data":"d9f47947cfef03169cfbf6edde2ebc8c5aac347b5c132221e3678b4c50f81740"} Nov 22 10:28:20 crc kubenswrapper[4789]: I1122 10:28:20.754529 4789 scope.go:117] "RemoveContainer" containerID="d9f47947cfef03169cfbf6edde2ebc8c5aac347b5c132221e3678b4c50f81740" Nov 22 10:28:20 crc kubenswrapper[4789]: I1122 10:28:20.898309 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-69pj5_must-gather-shff9_3d58dd27-067a-4510-b219-340217ca2ecb/gather/0.log" Nov 22 10:28:26 crc kubenswrapper[4789]: I1122 10:28:26.718855 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-q68j2"] Nov 22 10:28:26 crc kubenswrapper[4789]: E1122 10:28:26.720193 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8dce9e8b-6c82-44b6-98fe-72555fd207fe" containerName="extract-content" Nov 22 10:28:26 crc kubenswrapper[4789]: I1122 10:28:26.720214 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="8dce9e8b-6c82-44b6-98fe-72555fd207fe" containerName="extract-content" Nov 22 10:28:26 crc kubenswrapper[4789]: E1122 10:28:26.720239 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8dce9e8b-6c82-44b6-98fe-72555fd207fe" containerName="registry-server" Nov 22 10:28:26 crc kubenswrapper[4789]: I1122 10:28:26.720247 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="8dce9e8b-6c82-44b6-98fe-72555fd207fe" containerName="registry-server" Nov 22 10:28:26 crc kubenswrapper[4789]: E1122 10:28:26.720277 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8dce9e8b-6c82-44b6-98fe-72555fd207fe" containerName="extract-utilities" Nov 22 10:28:26 crc kubenswrapper[4789]: I1122 10:28:26.720286 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="8dce9e8b-6c82-44b6-98fe-72555fd207fe" containerName="extract-utilities" Nov 22 10:28:26 crc kubenswrapper[4789]: I1122 10:28:26.720547 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="8dce9e8b-6c82-44b6-98fe-72555fd207fe" containerName="registry-server" Nov 22 10:28:26 crc kubenswrapper[4789]: I1122 10:28:26.722486 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q68j2" Nov 22 10:28:26 crc kubenswrapper[4789]: I1122 10:28:26.736055 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-q68j2"] Nov 22 10:28:26 crc kubenswrapper[4789]: I1122 10:28:26.767980 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fq6vm\" (UniqueName: \"kubernetes.io/projected/10f0a532-e096-4c93-a487-e6d5c85863aa-kube-api-access-fq6vm\") pod \"redhat-marketplace-q68j2\" (UID: \"10f0a532-e096-4c93-a487-e6d5c85863aa\") " pod="openshift-marketplace/redhat-marketplace-q68j2" Nov 22 10:28:26 crc kubenswrapper[4789]: I1122 10:28:26.768022 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10f0a532-e096-4c93-a487-e6d5c85863aa-catalog-content\") pod \"redhat-marketplace-q68j2\" (UID: \"10f0a532-e096-4c93-a487-e6d5c85863aa\") " pod="openshift-marketplace/redhat-marketplace-q68j2" Nov 22 10:28:26 crc kubenswrapper[4789]: I1122 10:28:26.768049 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10f0a532-e096-4c93-a487-e6d5c85863aa-utilities\") pod \"redhat-marketplace-q68j2\" (UID: \"10f0a532-e096-4c93-a487-e6d5c85863aa\") " pod="openshift-marketplace/redhat-marketplace-q68j2" Nov 22 10:28:26 crc kubenswrapper[4789]: I1122 10:28:26.869917 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fq6vm\" (UniqueName: \"kubernetes.io/projected/10f0a532-e096-4c93-a487-e6d5c85863aa-kube-api-access-fq6vm\") pod \"redhat-marketplace-q68j2\" (UID: \"10f0a532-e096-4c93-a487-e6d5c85863aa\") " pod="openshift-marketplace/redhat-marketplace-q68j2" Nov 22 10:28:26 crc kubenswrapper[4789]: I1122 10:28:26.870210 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10f0a532-e096-4c93-a487-e6d5c85863aa-catalog-content\") pod \"redhat-marketplace-q68j2\" (UID: \"10f0a532-e096-4c93-a487-e6d5c85863aa\") " pod="openshift-marketplace/redhat-marketplace-q68j2" Nov 22 10:28:26 crc kubenswrapper[4789]: I1122 10:28:26.870398 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10f0a532-e096-4c93-a487-e6d5c85863aa-utilities\") pod \"redhat-marketplace-q68j2\" (UID: \"10f0a532-e096-4c93-a487-e6d5c85863aa\") " pod="openshift-marketplace/redhat-marketplace-q68j2" Nov 22 10:28:26 crc kubenswrapper[4789]: I1122 10:28:26.870638 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10f0a532-e096-4c93-a487-e6d5c85863aa-catalog-content\") pod \"redhat-marketplace-q68j2\" (UID: \"10f0a532-e096-4c93-a487-e6d5c85863aa\") " pod="openshift-marketplace/redhat-marketplace-q68j2" Nov 22 10:28:26 crc kubenswrapper[4789]: I1122 10:28:26.870826 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10f0a532-e096-4c93-a487-e6d5c85863aa-utilities\") pod \"redhat-marketplace-q68j2\" (UID: \"10f0a532-e096-4c93-a487-e6d5c85863aa\") " pod="openshift-marketplace/redhat-marketplace-q68j2" Nov 22 10:28:26 crc kubenswrapper[4789]: I1122 10:28:26.901260 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fq6vm\" (UniqueName: \"kubernetes.io/projected/10f0a532-e096-4c93-a487-e6d5c85863aa-kube-api-access-fq6vm\") pod \"redhat-marketplace-q68j2\" (UID: \"10f0a532-e096-4c93-a487-e6d5c85863aa\") " pod="openshift-marketplace/redhat-marketplace-q68j2" Nov 22 10:28:27 crc kubenswrapper[4789]: I1122 10:28:27.066704 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q68j2" Nov 22 10:28:27 crc kubenswrapper[4789]: I1122 10:28:27.516713 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-q68j2"] Nov 22 10:28:27 crc kubenswrapper[4789]: I1122 10:28:27.825822 4789 generic.go:334] "Generic (PLEG): container finished" podID="10f0a532-e096-4c93-a487-e6d5c85863aa" containerID="1cd5199fa64ea1684c5053c7fedcdceb55a65d3edd26ab8d3434c132df218c85" exitCode=0 Nov 22 10:28:27 crc kubenswrapper[4789]: I1122 10:28:27.825866 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q68j2" event={"ID":"10f0a532-e096-4c93-a487-e6d5c85863aa","Type":"ContainerDied","Data":"1cd5199fa64ea1684c5053c7fedcdceb55a65d3edd26ab8d3434c132df218c85"} Nov 22 10:28:27 crc kubenswrapper[4789]: I1122 10:28:27.825894 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q68j2" event={"ID":"10f0a532-e096-4c93-a487-e6d5c85863aa","Type":"ContainerStarted","Data":"7d00568ef1fff087eabe8f67c9729da4d596448455cfb9ce52ca0bb941c5cda9"} Nov 22 10:28:29 crc kubenswrapper[4789]: I1122 10:28:29.773303 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-69pj5/must-gather-shff9"] Nov 22 10:28:29 crc kubenswrapper[4789]: I1122 10:28:29.774377 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-69pj5/must-gather-shff9" podUID="3d58dd27-067a-4510-b219-340217ca2ecb" containerName="copy" containerID="cri-o://b05af0f3351f28bf0a5ba02841067f047a6b70712943d2d47ab2398855d74fe5" gracePeriod=2 Nov 22 10:28:29 crc kubenswrapper[4789]: I1122 10:28:29.784369 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-69pj5/must-gather-shff9"] Nov 22 10:28:29 crc kubenswrapper[4789]: I1122 10:28:29.845123 4789 generic.go:334] "Generic (PLEG): container finished" podID="10f0a532-e096-4c93-a487-e6d5c85863aa" containerID="b50124f5d4ef5d9ead0133040334d609739ff1a8dfc13dcac26a646da2aa694f" exitCode=0 Nov 22 10:28:29 crc kubenswrapper[4789]: I1122 10:28:29.845186 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q68j2" event={"ID":"10f0a532-e096-4c93-a487-e6d5c85863aa","Type":"ContainerDied","Data":"b50124f5d4ef5d9ead0133040334d609739ff1a8dfc13dcac26a646da2aa694f"} Nov 22 10:28:30 crc kubenswrapper[4789]: I1122 10:28:30.230237 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-69pj5_must-gather-shff9_3d58dd27-067a-4510-b219-340217ca2ecb/copy/0.log" Nov 22 10:28:30 crc kubenswrapper[4789]: I1122 10:28:30.230982 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-69pj5/must-gather-shff9" Nov 22 10:28:30 crc kubenswrapper[4789]: I1122 10:28:30.333776 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wt9r5\" (UniqueName: \"kubernetes.io/projected/3d58dd27-067a-4510-b219-340217ca2ecb-kube-api-access-wt9r5\") pod \"3d58dd27-067a-4510-b219-340217ca2ecb\" (UID: \"3d58dd27-067a-4510-b219-340217ca2ecb\") " Nov 22 10:28:30 crc kubenswrapper[4789]: I1122 10:28:30.333934 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/3d58dd27-067a-4510-b219-340217ca2ecb-must-gather-output\") pod \"3d58dd27-067a-4510-b219-340217ca2ecb\" (UID: \"3d58dd27-067a-4510-b219-340217ca2ecb\") " Nov 22 10:28:30 crc kubenswrapper[4789]: I1122 10:28:30.340140 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d58dd27-067a-4510-b219-340217ca2ecb-kube-api-access-wt9r5" (OuterVolumeSpecName: "kube-api-access-wt9r5") pod "3d58dd27-067a-4510-b219-340217ca2ecb" (UID: "3d58dd27-067a-4510-b219-340217ca2ecb"). InnerVolumeSpecName "kube-api-access-wt9r5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 10:28:30 crc kubenswrapper[4789]: I1122 10:28:30.437827 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wt9r5\" (UniqueName: \"kubernetes.io/projected/3d58dd27-067a-4510-b219-340217ca2ecb-kube-api-access-wt9r5\") on node \"crc\" DevicePath \"\"" Nov 22 10:28:30 crc kubenswrapper[4789]: I1122 10:28:30.520692 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3d58dd27-067a-4510-b219-340217ca2ecb-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "3d58dd27-067a-4510-b219-340217ca2ecb" (UID: "3d58dd27-067a-4510-b219-340217ca2ecb"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 10:28:30 crc kubenswrapper[4789]: I1122 10:28:30.539711 4789 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/3d58dd27-067a-4510-b219-340217ca2ecb-must-gather-output\") on node \"crc\" DevicePath \"\"" Nov 22 10:28:30 crc kubenswrapper[4789]: I1122 10:28:30.854145 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-69pj5_must-gather-shff9_3d58dd27-067a-4510-b219-340217ca2ecb/copy/0.log" Nov 22 10:28:30 crc kubenswrapper[4789]: I1122 10:28:30.854842 4789 generic.go:334] "Generic (PLEG): container finished" podID="3d58dd27-067a-4510-b219-340217ca2ecb" containerID="b05af0f3351f28bf0a5ba02841067f047a6b70712943d2d47ab2398855d74fe5" exitCode=143 Nov 22 10:28:30 crc kubenswrapper[4789]: I1122 10:28:30.854920 4789 scope.go:117] "RemoveContainer" containerID="b05af0f3351f28bf0a5ba02841067f047a6b70712943d2d47ab2398855d74fe5" Nov 22 10:28:30 crc kubenswrapper[4789]: I1122 10:28:30.854945 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-69pj5/must-gather-shff9" Nov 22 10:28:30 crc kubenswrapper[4789]: I1122 10:28:30.858072 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q68j2" event={"ID":"10f0a532-e096-4c93-a487-e6d5c85863aa","Type":"ContainerStarted","Data":"1f47a7165e0e921f39d1b7a09634065752cc24d193a670a3a183596bc0dfc55f"} Nov 22 10:28:30 crc kubenswrapper[4789]: I1122 10:28:30.877927 4789 scope.go:117] "RemoveContainer" containerID="d9f47947cfef03169cfbf6edde2ebc8c5aac347b5c132221e3678b4c50f81740" Nov 22 10:28:30 crc kubenswrapper[4789]: I1122 10:28:30.890507 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-q68j2" podStartSLOduration=2.212540078 podStartE2EDuration="4.890482233s" podCreationTimestamp="2025-11-22 10:28:26 +0000 UTC" firstStartedPulling="2025-11-22 10:28:27.827686168 +0000 UTC m=+9122.062086451" lastFinishedPulling="2025-11-22 10:28:30.505628333 +0000 UTC m=+9124.740028606" observedRunningTime="2025-11-22 10:28:30.873132786 +0000 UTC m=+9125.107533059" watchObservedRunningTime="2025-11-22 10:28:30.890482233 +0000 UTC m=+9125.124882506" Nov 22 10:28:30 crc kubenswrapper[4789]: I1122 10:28:30.960634 4789 scope.go:117] "RemoveContainer" containerID="b05af0f3351f28bf0a5ba02841067f047a6b70712943d2d47ab2398855d74fe5" Nov 22 10:28:30 crc kubenswrapper[4789]: E1122 10:28:30.961727 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b05af0f3351f28bf0a5ba02841067f047a6b70712943d2d47ab2398855d74fe5\": container with ID starting with b05af0f3351f28bf0a5ba02841067f047a6b70712943d2d47ab2398855d74fe5 not found: ID does not exist" containerID="b05af0f3351f28bf0a5ba02841067f047a6b70712943d2d47ab2398855d74fe5" Nov 22 10:28:30 crc kubenswrapper[4789]: I1122 10:28:30.961783 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b05af0f3351f28bf0a5ba02841067f047a6b70712943d2d47ab2398855d74fe5"} err="failed to get container status \"b05af0f3351f28bf0a5ba02841067f047a6b70712943d2d47ab2398855d74fe5\": rpc error: code = NotFound desc = could not find container \"b05af0f3351f28bf0a5ba02841067f047a6b70712943d2d47ab2398855d74fe5\": container with ID starting with b05af0f3351f28bf0a5ba02841067f047a6b70712943d2d47ab2398855d74fe5 not found: ID does not exist" Nov 22 10:28:30 crc kubenswrapper[4789]: I1122 10:28:30.961809 4789 scope.go:117] "RemoveContainer" containerID="d9f47947cfef03169cfbf6edde2ebc8c5aac347b5c132221e3678b4c50f81740" Nov 22 10:28:30 crc kubenswrapper[4789]: E1122 10:28:30.962051 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d9f47947cfef03169cfbf6edde2ebc8c5aac347b5c132221e3678b4c50f81740\": container with ID starting with d9f47947cfef03169cfbf6edde2ebc8c5aac347b5c132221e3678b4c50f81740 not found: ID does not exist" containerID="d9f47947cfef03169cfbf6edde2ebc8c5aac347b5c132221e3678b4c50f81740" Nov 22 10:28:30 crc kubenswrapper[4789]: I1122 10:28:30.962105 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9f47947cfef03169cfbf6edde2ebc8c5aac347b5c132221e3678b4c50f81740"} err="failed to get container status \"d9f47947cfef03169cfbf6edde2ebc8c5aac347b5c132221e3678b4c50f81740\": rpc error: code = NotFound desc = could not find container \"d9f47947cfef03169cfbf6edde2ebc8c5aac347b5c132221e3678b4c50f81740\": container with ID starting with d9f47947cfef03169cfbf6edde2ebc8c5aac347b5c132221e3678b4c50f81740 not found: ID does not exist" Nov 22 10:28:31 crc kubenswrapper[4789]: I1122 10:28:31.625067 4789 scope.go:117] "RemoveContainer" containerID="383dd4ebe97bdf99ffc013e8160ff119f885ae64f982a01a9bac9082cc9a2efb" Nov 22 10:28:31 crc kubenswrapper[4789]: I1122 10:28:31.978827 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3d58dd27-067a-4510-b219-340217ca2ecb" path="/var/lib/kubelet/pods/3d58dd27-067a-4510-b219-340217ca2ecb/volumes" Nov 22 10:28:35 crc kubenswrapper[4789]: I1122 10:28:35.372445 4789 patch_prober.go:28] interesting pod/machine-config-daemon-pxvrm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 10:28:35 crc kubenswrapper[4789]: I1122 10:28:35.373020 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 10:28:37 crc kubenswrapper[4789]: I1122 10:28:37.067075 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-q68j2" Nov 22 10:28:37 crc kubenswrapper[4789]: I1122 10:28:37.068117 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-q68j2" Nov 22 10:28:37 crc kubenswrapper[4789]: I1122 10:28:37.150303 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-q68j2" Nov 22 10:28:38 crc kubenswrapper[4789]: I1122 10:28:38.238255 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-q68j2" Nov 22 10:28:38 crc kubenswrapper[4789]: I1122 10:28:38.302995 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-q68j2"] Nov 22 10:28:39 crc kubenswrapper[4789]: I1122 10:28:39.953627 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-q68j2" podUID="10f0a532-e096-4c93-a487-e6d5c85863aa" containerName="registry-server" containerID="cri-o://1f47a7165e0e921f39d1b7a09634065752cc24d193a670a3a183596bc0dfc55f" gracePeriod=2 Nov 22 10:28:40 crc kubenswrapper[4789]: I1122 10:28:40.780966 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q68j2" Nov 22 10:28:40 crc kubenswrapper[4789]: I1122 10:28:40.950927 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10f0a532-e096-4c93-a487-e6d5c85863aa-utilities\") pod \"10f0a532-e096-4c93-a487-e6d5c85863aa\" (UID: \"10f0a532-e096-4c93-a487-e6d5c85863aa\") " Nov 22 10:28:40 crc kubenswrapper[4789]: I1122 10:28:40.951082 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10f0a532-e096-4c93-a487-e6d5c85863aa-catalog-content\") pod \"10f0a532-e096-4c93-a487-e6d5c85863aa\" (UID: \"10f0a532-e096-4c93-a487-e6d5c85863aa\") " Nov 22 10:28:40 crc kubenswrapper[4789]: I1122 10:28:40.951177 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fq6vm\" (UniqueName: \"kubernetes.io/projected/10f0a532-e096-4c93-a487-e6d5c85863aa-kube-api-access-fq6vm\") pod \"10f0a532-e096-4c93-a487-e6d5c85863aa\" (UID: \"10f0a532-e096-4c93-a487-e6d5c85863aa\") " Nov 22 10:28:40 crc kubenswrapper[4789]: I1122 10:28:40.953268 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/10f0a532-e096-4c93-a487-e6d5c85863aa-utilities" (OuterVolumeSpecName: "utilities") pod "10f0a532-e096-4c93-a487-e6d5c85863aa" (UID: "10f0a532-e096-4c93-a487-e6d5c85863aa"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 10:28:40 crc kubenswrapper[4789]: I1122 10:28:40.963359 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/10f0a532-e096-4c93-a487-e6d5c85863aa-kube-api-access-fq6vm" (OuterVolumeSpecName: "kube-api-access-fq6vm") pod "10f0a532-e096-4c93-a487-e6d5c85863aa" (UID: "10f0a532-e096-4c93-a487-e6d5c85863aa"). InnerVolumeSpecName "kube-api-access-fq6vm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 10:28:40 crc kubenswrapper[4789]: I1122 10:28:40.968412 4789 generic.go:334] "Generic (PLEG): container finished" podID="10f0a532-e096-4c93-a487-e6d5c85863aa" containerID="1f47a7165e0e921f39d1b7a09634065752cc24d193a670a3a183596bc0dfc55f" exitCode=0 Nov 22 10:28:40 crc kubenswrapper[4789]: I1122 10:28:40.968457 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q68j2" event={"ID":"10f0a532-e096-4c93-a487-e6d5c85863aa","Type":"ContainerDied","Data":"1f47a7165e0e921f39d1b7a09634065752cc24d193a670a3a183596bc0dfc55f"} Nov 22 10:28:40 crc kubenswrapper[4789]: I1122 10:28:40.968486 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q68j2" event={"ID":"10f0a532-e096-4c93-a487-e6d5c85863aa","Type":"ContainerDied","Data":"7d00568ef1fff087eabe8f67c9729da4d596448455cfb9ce52ca0bb941c5cda9"} Nov 22 10:28:40 crc kubenswrapper[4789]: I1122 10:28:40.968506 4789 scope.go:117] "RemoveContainer" containerID="1f47a7165e0e921f39d1b7a09634065752cc24d193a670a3a183596bc0dfc55f" Nov 22 10:28:40 crc kubenswrapper[4789]: I1122 10:28:40.968659 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q68j2" Nov 22 10:28:41 crc kubenswrapper[4789]: I1122 10:28:41.005660 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/10f0a532-e096-4c93-a487-e6d5c85863aa-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "10f0a532-e096-4c93-a487-e6d5c85863aa" (UID: "10f0a532-e096-4c93-a487-e6d5c85863aa"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 10:28:41 crc kubenswrapper[4789]: I1122 10:28:41.037257 4789 scope.go:117] "RemoveContainer" containerID="b50124f5d4ef5d9ead0133040334d609739ff1a8dfc13dcac26a646da2aa694f" Nov 22 10:28:41 crc kubenswrapper[4789]: I1122 10:28:41.068469 4789 scope.go:117] "RemoveContainer" containerID="1cd5199fa64ea1684c5053c7fedcdceb55a65d3edd26ab8d3434c132df218c85" Nov 22 10:28:41 crc kubenswrapper[4789]: I1122 10:28:41.070896 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10f0a532-e096-4c93-a487-e6d5c85863aa-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 10:28:41 crc kubenswrapper[4789]: I1122 10:28:41.070942 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10f0a532-e096-4c93-a487-e6d5c85863aa-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 10:28:41 crc kubenswrapper[4789]: I1122 10:28:41.070959 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fq6vm\" (UniqueName: \"kubernetes.io/projected/10f0a532-e096-4c93-a487-e6d5c85863aa-kube-api-access-fq6vm\") on node \"crc\" DevicePath \"\"" Nov 22 10:28:41 crc kubenswrapper[4789]: I1122 10:28:41.128467 4789 scope.go:117] "RemoveContainer" containerID="1f47a7165e0e921f39d1b7a09634065752cc24d193a670a3a183596bc0dfc55f" Nov 22 10:28:41 crc kubenswrapper[4789]: E1122 10:28:41.129308 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1f47a7165e0e921f39d1b7a09634065752cc24d193a670a3a183596bc0dfc55f\": container with ID starting with 1f47a7165e0e921f39d1b7a09634065752cc24d193a670a3a183596bc0dfc55f not found: ID does not exist" containerID="1f47a7165e0e921f39d1b7a09634065752cc24d193a670a3a183596bc0dfc55f" Nov 22 10:28:41 crc kubenswrapper[4789]: I1122 10:28:41.129346 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f47a7165e0e921f39d1b7a09634065752cc24d193a670a3a183596bc0dfc55f"} err="failed to get container status \"1f47a7165e0e921f39d1b7a09634065752cc24d193a670a3a183596bc0dfc55f\": rpc error: code = NotFound desc = could not find container \"1f47a7165e0e921f39d1b7a09634065752cc24d193a670a3a183596bc0dfc55f\": container with ID starting with 1f47a7165e0e921f39d1b7a09634065752cc24d193a670a3a183596bc0dfc55f not found: ID does not exist" Nov 22 10:28:41 crc kubenswrapper[4789]: I1122 10:28:41.129371 4789 scope.go:117] "RemoveContainer" containerID="b50124f5d4ef5d9ead0133040334d609739ff1a8dfc13dcac26a646da2aa694f" Nov 22 10:28:41 crc kubenswrapper[4789]: E1122 10:28:41.129731 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b50124f5d4ef5d9ead0133040334d609739ff1a8dfc13dcac26a646da2aa694f\": container with ID starting with b50124f5d4ef5d9ead0133040334d609739ff1a8dfc13dcac26a646da2aa694f not found: ID does not exist" containerID="b50124f5d4ef5d9ead0133040334d609739ff1a8dfc13dcac26a646da2aa694f" Nov 22 10:28:41 crc kubenswrapper[4789]: I1122 10:28:41.129784 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b50124f5d4ef5d9ead0133040334d609739ff1a8dfc13dcac26a646da2aa694f"} err="failed to get container status \"b50124f5d4ef5d9ead0133040334d609739ff1a8dfc13dcac26a646da2aa694f\": rpc error: code = NotFound desc = could not find container \"b50124f5d4ef5d9ead0133040334d609739ff1a8dfc13dcac26a646da2aa694f\": container with ID starting with b50124f5d4ef5d9ead0133040334d609739ff1a8dfc13dcac26a646da2aa694f not found: ID does not exist" Nov 22 10:28:41 crc kubenswrapper[4789]: I1122 10:28:41.129814 4789 scope.go:117] "RemoveContainer" containerID="1cd5199fa64ea1684c5053c7fedcdceb55a65d3edd26ab8d3434c132df218c85" Nov 22 10:28:41 crc kubenswrapper[4789]: E1122 10:28:41.130053 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1cd5199fa64ea1684c5053c7fedcdceb55a65d3edd26ab8d3434c132df218c85\": container with ID starting with 1cd5199fa64ea1684c5053c7fedcdceb55a65d3edd26ab8d3434c132df218c85 not found: ID does not exist" containerID="1cd5199fa64ea1684c5053c7fedcdceb55a65d3edd26ab8d3434c132df218c85" Nov 22 10:28:41 crc kubenswrapper[4789]: I1122 10:28:41.130080 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1cd5199fa64ea1684c5053c7fedcdceb55a65d3edd26ab8d3434c132df218c85"} err="failed to get container status \"1cd5199fa64ea1684c5053c7fedcdceb55a65d3edd26ab8d3434c132df218c85\": rpc error: code = NotFound desc = could not find container \"1cd5199fa64ea1684c5053c7fedcdceb55a65d3edd26ab8d3434c132df218c85\": container with ID starting with 1cd5199fa64ea1684c5053c7fedcdceb55a65d3edd26ab8d3434c132df218c85 not found: ID does not exist" Nov 22 10:28:41 crc kubenswrapper[4789]: I1122 10:28:41.328833 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-q68j2"] Nov 22 10:28:41 crc kubenswrapper[4789]: I1122 10:28:41.346718 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-q68j2"] Nov 22 10:28:41 crc kubenswrapper[4789]: I1122 10:28:41.979137 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="10f0a532-e096-4c93-a487-e6d5c85863aa" path="/var/lib/kubelet/pods/10f0a532-e096-4c93-a487-e6d5c85863aa/volumes" Nov 22 10:28:52 crc kubenswrapper[4789]: E1122 10:28:52.966478 4789 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Nov 22 10:29:05 crc kubenswrapper[4789]: I1122 10:29:05.372942 4789 patch_prober.go:28] interesting pod/machine-config-daemon-pxvrm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 10:29:05 crc kubenswrapper[4789]: I1122 10:29:05.373619 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 10:29:31 crc kubenswrapper[4789]: I1122 10:29:31.728995 4789 scope.go:117] "RemoveContainer" containerID="b755428c6bd9a67fdfed8ed1829d794d0d37705f81439dac9f79c368932d7194" Nov 22 10:29:35 crc kubenswrapper[4789]: I1122 10:29:35.371968 4789 patch_prober.go:28] interesting pod/machine-config-daemon-pxvrm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 10:29:35 crc kubenswrapper[4789]: I1122 10:29:35.373044 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 10:29:35 crc kubenswrapper[4789]: I1122 10:29:35.373114 4789 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" Nov 22 10:29:35 crc kubenswrapper[4789]: I1122 10:29:35.374354 4789 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3590e488179d593a3c8981249b3f907f5a30f296b7d9ba57504c20a09c6eb5bc"} pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 22 10:29:35 crc kubenswrapper[4789]: I1122 10:29:35.374501 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" containerName="machine-config-daemon" containerID="cri-o://3590e488179d593a3c8981249b3f907f5a30f296b7d9ba57504c20a09c6eb5bc" gracePeriod=600 Nov 22 10:29:35 crc kubenswrapper[4789]: E1122 10:29:35.506497 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 10:29:35 crc kubenswrapper[4789]: I1122 10:29:35.696626 4789 generic.go:334] "Generic (PLEG): container finished" podID="58aa8071-7ecc-4692-8789-2db04ec70647" containerID="3590e488179d593a3c8981249b3f907f5a30f296b7d9ba57504c20a09c6eb5bc" exitCode=0 Nov 22 10:29:35 crc kubenswrapper[4789]: I1122 10:29:35.696821 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" event={"ID":"58aa8071-7ecc-4692-8789-2db04ec70647","Type":"ContainerDied","Data":"3590e488179d593a3c8981249b3f907f5a30f296b7d9ba57504c20a09c6eb5bc"} Nov 22 10:29:35 crc kubenswrapper[4789]: I1122 10:29:35.697177 4789 scope.go:117] "RemoveContainer" containerID="757306d657b53911c979b1a984e697106c927039cb3f24d4ba4a2a51a772a120" Nov 22 10:29:35 crc kubenswrapper[4789]: I1122 10:29:35.698419 4789 scope.go:117] "RemoveContainer" containerID="3590e488179d593a3c8981249b3f907f5a30f296b7d9ba57504c20a09c6eb5bc" Nov 22 10:29:35 crc kubenswrapper[4789]: E1122 10:29:35.698923 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 10:29:46 crc kubenswrapper[4789]: I1122 10:29:46.965411 4789 scope.go:117] "RemoveContainer" containerID="3590e488179d593a3c8981249b3f907f5a30f296b7d9ba57504c20a09c6eb5bc" Nov 22 10:29:46 crc kubenswrapper[4789]: E1122 10:29:46.966596 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 10:30:00 crc kubenswrapper[4789]: I1122 10:30:00.185965 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396790-f6vwc"] Nov 22 10:30:00 crc kubenswrapper[4789]: E1122 10:30:00.187333 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d58dd27-067a-4510-b219-340217ca2ecb" containerName="copy" Nov 22 10:30:00 crc kubenswrapper[4789]: I1122 10:30:00.187358 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d58dd27-067a-4510-b219-340217ca2ecb" containerName="copy" Nov 22 10:30:00 crc kubenswrapper[4789]: E1122 10:30:00.187391 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10f0a532-e096-4c93-a487-e6d5c85863aa" containerName="extract-content" Nov 22 10:30:00 crc kubenswrapper[4789]: I1122 10:30:00.187404 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="10f0a532-e096-4c93-a487-e6d5c85863aa" containerName="extract-content" Nov 22 10:30:00 crc kubenswrapper[4789]: E1122 10:30:00.187449 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10f0a532-e096-4c93-a487-e6d5c85863aa" containerName="registry-server" Nov 22 10:30:00 crc kubenswrapper[4789]: I1122 10:30:00.187463 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="10f0a532-e096-4c93-a487-e6d5c85863aa" containerName="registry-server" Nov 22 10:30:00 crc kubenswrapper[4789]: E1122 10:30:00.187485 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d58dd27-067a-4510-b219-340217ca2ecb" containerName="gather" Nov 22 10:30:00 crc kubenswrapper[4789]: I1122 10:30:00.187499 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d58dd27-067a-4510-b219-340217ca2ecb" containerName="gather" Nov 22 10:30:00 crc kubenswrapper[4789]: E1122 10:30:00.187522 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10f0a532-e096-4c93-a487-e6d5c85863aa" containerName="extract-utilities" Nov 22 10:30:00 crc kubenswrapper[4789]: I1122 10:30:00.187534 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="10f0a532-e096-4c93-a487-e6d5c85863aa" containerName="extract-utilities" Nov 22 10:30:00 crc kubenswrapper[4789]: I1122 10:30:00.187910 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d58dd27-067a-4510-b219-340217ca2ecb" containerName="gather" Nov 22 10:30:00 crc kubenswrapper[4789]: I1122 10:30:00.187960 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d58dd27-067a-4510-b219-340217ca2ecb" containerName="copy" Nov 22 10:30:00 crc kubenswrapper[4789]: I1122 10:30:00.187994 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="10f0a532-e096-4c93-a487-e6d5c85863aa" containerName="registry-server" Nov 22 10:30:00 crc kubenswrapper[4789]: I1122 10:30:00.189064 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396790-f6vwc" Nov 22 10:30:00 crc kubenswrapper[4789]: I1122 10:30:00.190914 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 22 10:30:00 crc kubenswrapper[4789]: I1122 10:30:00.192606 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 22 10:30:00 crc kubenswrapper[4789]: I1122 10:30:00.198995 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396790-f6vwc"] Nov 22 10:30:00 crc kubenswrapper[4789]: I1122 10:30:00.226551 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1cfa05ee-c012-4bf4-876c-2ac48f6e5974-config-volume\") pod \"collect-profiles-29396790-f6vwc\" (UID: \"1cfa05ee-c012-4bf4-876c-2ac48f6e5974\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396790-f6vwc" Nov 22 10:30:00 crc kubenswrapper[4789]: I1122 10:30:00.226679 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2xn7q\" (UniqueName: \"kubernetes.io/projected/1cfa05ee-c012-4bf4-876c-2ac48f6e5974-kube-api-access-2xn7q\") pod \"collect-profiles-29396790-f6vwc\" (UID: \"1cfa05ee-c012-4bf4-876c-2ac48f6e5974\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396790-f6vwc" Nov 22 10:30:00 crc kubenswrapper[4789]: I1122 10:30:00.226857 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1cfa05ee-c012-4bf4-876c-2ac48f6e5974-secret-volume\") pod \"collect-profiles-29396790-f6vwc\" (UID: \"1cfa05ee-c012-4bf4-876c-2ac48f6e5974\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396790-f6vwc" Nov 22 10:30:00 crc kubenswrapper[4789]: I1122 10:30:00.328677 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1cfa05ee-c012-4bf4-876c-2ac48f6e5974-secret-volume\") pod \"collect-profiles-29396790-f6vwc\" (UID: \"1cfa05ee-c012-4bf4-876c-2ac48f6e5974\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396790-f6vwc" Nov 22 10:30:00 crc kubenswrapper[4789]: I1122 10:30:00.328810 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1cfa05ee-c012-4bf4-876c-2ac48f6e5974-config-volume\") pod \"collect-profiles-29396790-f6vwc\" (UID: \"1cfa05ee-c012-4bf4-876c-2ac48f6e5974\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396790-f6vwc" Nov 22 10:30:00 crc kubenswrapper[4789]: I1122 10:30:00.328928 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2xn7q\" (UniqueName: \"kubernetes.io/projected/1cfa05ee-c012-4bf4-876c-2ac48f6e5974-kube-api-access-2xn7q\") pod \"collect-profiles-29396790-f6vwc\" (UID: \"1cfa05ee-c012-4bf4-876c-2ac48f6e5974\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396790-f6vwc" Nov 22 10:30:00 crc kubenswrapper[4789]: I1122 10:30:00.329812 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1cfa05ee-c012-4bf4-876c-2ac48f6e5974-config-volume\") pod \"collect-profiles-29396790-f6vwc\" (UID: \"1cfa05ee-c012-4bf4-876c-2ac48f6e5974\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396790-f6vwc" Nov 22 10:30:00 crc kubenswrapper[4789]: I1122 10:30:00.335399 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1cfa05ee-c012-4bf4-876c-2ac48f6e5974-secret-volume\") pod \"collect-profiles-29396790-f6vwc\" (UID: \"1cfa05ee-c012-4bf4-876c-2ac48f6e5974\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396790-f6vwc" Nov 22 10:30:00 crc kubenswrapper[4789]: I1122 10:30:00.346346 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2xn7q\" (UniqueName: \"kubernetes.io/projected/1cfa05ee-c012-4bf4-876c-2ac48f6e5974-kube-api-access-2xn7q\") pod \"collect-profiles-29396790-f6vwc\" (UID: \"1cfa05ee-c012-4bf4-876c-2ac48f6e5974\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29396790-f6vwc" Nov 22 10:30:00 crc kubenswrapper[4789]: I1122 10:30:00.511918 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396790-f6vwc" Nov 22 10:30:01 crc kubenswrapper[4789]: I1122 10:30:01.000238 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396790-f6vwc"] Nov 22 10:30:01 crc kubenswrapper[4789]: I1122 10:30:01.022799 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29396790-f6vwc" event={"ID":"1cfa05ee-c012-4bf4-876c-2ac48f6e5974","Type":"ContainerStarted","Data":"62f5b0b5ecc40bbf70e04212f8d99f17faf341699e7aa87b2c419fe156be2786"} Nov 22 10:30:01 crc kubenswrapper[4789]: I1122 10:30:01.965745 4789 scope.go:117] "RemoveContainer" containerID="3590e488179d593a3c8981249b3f907f5a30f296b7d9ba57504c20a09c6eb5bc" Nov 22 10:30:01 crc kubenswrapper[4789]: E1122 10:30:01.966510 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 10:30:02 crc kubenswrapper[4789]: I1122 10:30:02.042611 4789 generic.go:334] "Generic (PLEG): container finished" podID="1cfa05ee-c012-4bf4-876c-2ac48f6e5974" containerID="dc93b25bfa7aa74ca8d0716b86ede14d7657cf475d6a4871858371885fe750bf" exitCode=0 Nov 22 10:30:02 crc kubenswrapper[4789]: I1122 10:30:02.042677 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29396790-f6vwc" event={"ID":"1cfa05ee-c012-4bf4-876c-2ac48f6e5974","Type":"ContainerDied","Data":"dc93b25bfa7aa74ca8d0716b86ede14d7657cf475d6a4871858371885fe750bf"} Nov 22 10:30:03 crc kubenswrapper[4789]: I1122 10:30:03.496905 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396790-f6vwc" Nov 22 10:30:03 crc kubenswrapper[4789]: I1122 10:30:03.615233 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1cfa05ee-c012-4bf4-876c-2ac48f6e5974-config-volume\") pod \"1cfa05ee-c012-4bf4-876c-2ac48f6e5974\" (UID: \"1cfa05ee-c012-4bf4-876c-2ac48f6e5974\") " Nov 22 10:30:03 crc kubenswrapper[4789]: I1122 10:30:03.615354 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2xn7q\" (UniqueName: \"kubernetes.io/projected/1cfa05ee-c012-4bf4-876c-2ac48f6e5974-kube-api-access-2xn7q\") pod \"1cfa05ee-c012-4bf4-876c-2ac48f6e5974\" (UID: \"1cfa05ee-c012-4bf4-876c-2ac48f6e5974\") " Nov 22 10:30:03 crc kubenswrapper[4789]: I1122 10:30:03.615613 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1cfa05ee-c012-4bf4-876c-2ac48f6e5974-secret-volume\") pod \"1cfa05ee-c012-4bf4-876c-2ac48f6e5974\" (UID: \"1cfa05ee-c012-4bf4-876c-2ac48f6e5974\") " Nov 22 10:30:03 crc kubenswrapper[4789]: I1122 10:30:03.616261 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1cfa05ee-c012-4bf4-876c-2ac48f6e5974-config-volume" (OuterVolumeSpecName: "config-volume") pod "1cfa05ee-c012-4bf4-876c-2ac48f6e5974" (UID: "1cfa05ee-c012-4bf4-876c-2ac48f6e5974"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 10:30:03 crc kubenswrapper[4789]: I1122 10:30:03.621944 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1cfa05ee-c012-4bf4-876c-2ac48f6e5974-kube-api-access-2xn7q" (OuterVolumeSpecName: "kube-api-access-2xn7q") pod "1cfa05ee-c012-4bf4-876c-2ac48f6e5974" (UID: "1cfa05ee-c012-4bf4-876c-2ac48f6e5974"). InnerVolumeSpecName "kube-api-access-2xn7q". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 10:30:03 crc kubenswrapper[4789]: I1122 10:30:03.622665 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1cfa05ee-c012-4bf4-876c-2ac48f6e5974-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "1cfa05ee-c012-4bf4-876c-2ac48f6e5974" (UID: "1cfa05ee-c012-4bf4-876c-2ac48f6e5974"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 10:30:03 crc kubenswrapper[4789]: I1122 10:30:03.718078 4789 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1cfa05ee-c012-4bf4-876c-2ac48f6e5974-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 22 10:30:03 crc kubenswrapper[4789]: I1122 10:30:03.718123 4789 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1cfa05ee-c012-4bf4-876c-2ac48f6e5974-config-volume\") on node \"crc\" DevicePath \"\"" Nov 22 10:30:03 crc kubenswrapper[4789]: I1122 10:30:03.718137 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2xn7q\" (UniqueName: \"kubernetes.io/projected/1cfa05ee-c012-4bf4-876c-2ac48f6e5974-kube-api-access-2xn7q\") on node \"crc\" DevicePath \"\"" Nov 22 10:30:04 crc kubenswrapper[4789]: I1122 10:30:04.066481 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29396790-f6vwc" event={"ID":"1cfa05ee-c012-4bf4-876c-2ac48f6e5974","Type":"ContainerDied","Data":"62f5b0b5ecc40bbf70e04212f8d99f17faf341699e7aa87b2c419fe156be2786"} Nov 22 10:30:04 crc kubenswrapper[4789]: I1122 10:30:04.066539 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="62f5b0b5ecc40bbf70e04212f8d99f17faf341699e7aa87b2c419fe156be2786" Nov 22 10:30:04 crc kubenswrapper[4789]: I1122 10:30:04.066855 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29396790-f6vwc" Nov 22 10:30:04 crc kubenswrapper[4789]: I1122 10:30:04.596964 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396745-q52jz"] Nov 22 10:30:04 crc kubenswrapper[4789]: I1122 10:30:04.613456 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29396745-q52jz"] Nov 22 10:30:05 crc kubenswrapper[4789]: I1122 10:30:05.985135 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="adfe0251-8ae5-48bc-a0b2-23399ddaf4d8" path="/var/lib/kubelet/pods/adfe0251-8ae5-48bc-a0b2-23399ddaf4d8/volumes" Nov 22 10:30:13 crc kubenswrapper[4789]: I1122 10:30:13.965869 4789 scope.go:117] "RemoveContainer" containerID="3590e488179d593a3c8981249b3f907f5a30f296b7d9ba57504c20a09c6eb5bc" Nov 22 10:30:13 crc kubenswrapper[4789]: E1122 10:30:13.967171 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 10:30:21 crc kubenswrapper[4789]: E1122 10:30:21.969398 4789 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Nov 22 10:30:28 crc kubenswrapper[4789]: I1122 10:30:28.966424 4789 scope.go:117] "RemoveContainer" containerID="3590e488179d593a3c8981249b3f907f5a30f296b7d9ba57504c20a09c6eb5bc" Nov 22 10:30:28 crc kubenswrapper[4789]: E1122 10:30:28.968206 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 10:30:31 crc kubenswrapper[4789]: I1122 10:30:31.834823 4789 scope.go:117] "RemoveContainer" containerID="7a17bd5bc847af1fd9e52cbfc735adb9c04cdb9695e08dd9e2e0b0eee77a160f" Nov 22 10:30:41 crc kubenswrapper[4789]: I1122 10:30:41.970056 4789 scope.go:117] "RemoveContainer" containerID="3590e488179d593a3c8981249b3f907f5a30f296b7d9ba57504c20a09c6eb5bc" Nov 22 10:30:41 crc kubenswrapper[4789]: E1122 10:30:41.971533 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 10:30:55 crc kubenswrapper[4789]: I1122 10:30:55.965460 4789 scope.go:117] "RemoveContainer" containerID="3590e488179d593a3c8981249b3f907f5a30f296b7d9ba57504c20a09c6eb5bc" Nov 22 10:30:55 crc kubenswrapper[4789]: E1122 10:30:55.967855 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 10:31:07 crc kubenswrapper[4789]: I1122 10:31:07.970398 4789 scope.go:117] "RemoveContainer" containerID="3590e488179d593a3c8981249b3f907f5a30f296b7d9ba57504c20a09c6eb5bc" Nov 22 10:31:07 crc kubenswrapper[4789]: E1122 10:31:07.971154 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 10:31:20 crc kubenswrapper[4789]: I1122 10:31:20.965735 4789 scope.go:117] "RemoveContainer" containerID="3590e488179d593a3c8981249b3f907f5a30f296b7d9ba57504c20a09c6eb5bc" Nov 22 10:31:20 crc kubenswrapper[4789]: E1122 10:31:20.966784 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 10:31:31 crc kubenswrapper[4789]: I1122 10:31:31.965948 4789 scope.go:117] "RemoveContainer" containerID="3590e488179d593a3c8981249b3f907f5a30f296b7d9ba57504c20a09c6eb5bc" Nov 22 10:31:31 crc kubenswrapper[4789]: E1122 10:31:31.967016 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 10:31:42 crc kubenswrapper[4789]: I1122 10:31:42.988871 4789 scope.go:117] "RemoveContainer" containerID="3590e488179d593a3c8981249b3f907f5a30f296b7d9ba57504c20a09c6eb5bc" Nov 22 10:31:42 crc kubenswrapper[4789]: E1122 10:31:42.992818 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 10:31:51 crc kubenswrapper[4789]: E1122 10:31:51.966248 4789 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Nov 22 10:31:57 crc kubenswrapper[4789]: I1122 10:31:57.972565 4789 scope.go:117] "RemoveContainer" containerID="3590e488179d593a3c8981249b3f907f5a30f296b7d9ba57504c20a09c6eb5bc" Nov 22 10:31:57 crc kubenswrapper[4789]: E1122 10:31:57.973491 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 10:32:12 crc kubenswrapper[4789]: I1122 10:32:12.965993 4789 scope.go:117] "RemoveContainer" containerID="3590e488179d593a3c8981249b3f907f5a30f296b7d9ba57504c20a09c6eb5bc" Nov 22 10:32:12 crc kubenswrapper[4789]: E1122 10:32:12.967254 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 10:32:26 crc kubenswrapper[4789]: I1122 10:32:26.966454 4789 scope.go:117] "RemoveContainer" containerID="3590e488179d593a3c8981249b3f907f5a30f296b7d9ba57504c20a09c6eb5bc" Nov 22 10:32:26 crc kubenswrapper[4789]: E1122 10:32:26.967676 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 10:32:39 crc kubenswrapper[4789]: I1122 10:32:39.967953 4789 scope.go:117] "RemoveContainer" containerID="3590e488179d593a3c8981249b3f907f5a30f296b7d9ba57504c20a09c6eb5bc" Nov 22 10:32:39 crc kubenswrapper[4789]: E1122 10:32:39.969131 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 10:32:51 crc kubenswrapper[4789]: I1122 10:32:51.970229 4789 scope.go:117] "RemoveContainer" containerID="3590e488179d593a3c8981249b3f907f5a30f296b7d9ba57504c20a09c6eb5bc" Nov 22 10:32:51 crc kubenswrapper[4789]: E1122 10:32:51.970827 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 10:32:57 crc kubenswrapper[4789]: I1122 10:32:57.866107 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-jtjsx"] Nov 22 10:32:57 crc kubenswrapper[4789]: E1122 10:32:57.867112 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1cfa05ee-c012-4bf4-876c-2ac48f6e5974" containerName="collect-profiles" Nov 22 10:32:57 crc kubenswrapper[4789]: I1122 10:32:57.867134 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="1cfa05ee-c012-4bf4-876c-2ac48f6e5974" containerName="collect-profiles" Nov 22 10:32:57 crc kubenswrapper[4789]: I1122 10:32:57.867518 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="1cfa05ee-c012-4bf4-876c-2ac48f6e5974" containerName="collect-profiles" Nov 22 10:32:57 crc kubenswrapper[4789]: I1122 10:32:57.870087 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jtjsx" Nov 22 10:32:57 crc kubenswrapper[4789]: I1122 10:32:57.880700 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jtjsx"] Nov 22 10:32:57 crc kubenswrapper[4789]: I1122 10:32:57.953579 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c4b4396-9cd6-44ee-b736-744c281dccda-utilities\") pod \"redhat-operators-jtjsx\" (UID: \"7c4b4396-9cd6-44ee-b736-744c281dccda\") " pod="openshift-marketplace/redhat-operators-jtjsx" Nov 22 10:32:57 crc kubenswrapper[4789]: I1122 10:32:57.953672 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c4b4396-9cd6-44ee-b736-744c281dccda-catalog-content\") pod \"redhat-operators-jtjsx\" (UID: \"7c4b4396-9cd6-44ee-b736-744c281dccda\") " pod="openshift-marketplace/redhat-operators-jtjsx" Nov 22 10:32:57 crc kubenswrapper[4789]: I1122 10:32:57.953737 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tmqr9\" (UniqueName: \"kubernetes.io/projected/7c4b4396-9cd6-44ee-b736-744c281dccda-kube-api-access-tmqr9\") pod \"redhat-operators-jtjsx\" (UID: \"7c4b4396-9cd6-44ee-b736-744c281dccda\") " pod="openshift-marketplace/redhat-operators-jtjsx" Nov 22 10:32:58 crc kubenswrapper[4789]: I1122 10:32:58.055395 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c4b4396-9cd6-44ee-b736-744c281dccda-utilities\") pod \"redhat-operators-jtjsx\" (UID: \"7c4b4396-9cd6-44ee-b736-744c281dccda\") " pod="openshift-marketplace/redhat-operators-jtjsx" Nov 22 10:32:58 crc kubenswrapper[4789]: I1122 10:32:58.055457 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c4b4396-9cd6-44ee-b736-744c281dccda-catalog-content\") pod \"redhat-operators-jtjsx\" (UID: \"7c4b4396-9cd6-44ee-b736-744c281dccda\") " pod="openshift-marketplace/redhat-operators-jtjsx" Nov 22 10:32:58 crc kubenswrapper[4789]: I1122 10:32:58.055487 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tmqr9\" (UniqueName: \"kubernetes.io/projected/7c4b4396-9cd6-44ee-b736-744c281dccda-kube-api-access-tmqr9\") pod \"redhat-operators-jtjsx\" (UID: \"7c4b4396-9cd6-44ee-b736-744c281dccda\") " pod="openshift-marketplace/redhat-operators-jtjsx" Nov 22 10:32:58 crc kubenswrapper[4789]: I1122 10:32:58.055913 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c4b4396-9cd6-44ee-b736-744c281dccda-utilities\") pod \"redhat-operators-jtjsx\" (UID: \"7c4b4396-9cd6-44ee-b736-744c281dccda\") " pod="openshift-marketplace/redhat-operators-jtjsx" Nov 22 10:32:58 crc kubenswrapper[4789]: I1122 10:32:58.056130 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c4b4396-9cd6-44ee-b736-744c281dccda-catalog-content\") pod \"redhat-operators-jtjsx\" (UID: \"7c4b4396-9cd6-44ee-b736-744c281dccda\") " pod="openshift-marketplace/redhat-operators-jtjsx" Nov 22 10:32:58 crc kubenswrapper[4789]: I1122 10:32:58.075965 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tmqr9\" (UniqueName: \"kubernetes.io/projected/7c4b4396-9cd6-44ee-b736-744c281dccda-kube-api-access-tmqr9\") pod \"redhat-operators-jtjsx\" (UID: \"7c4b4396-9cd6-44ee-b736-744c281dccda\") " pod="openshift-marketplace/redhat-operators-jtjsx" Nov 22 10:32:58 crc kubenswrapper[4789]: I1122 10:32:58.192987 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jtjsx" Nov 22 10:32:58 crc kubenswrapper[4789]: I1122 10:32:58.691281 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jtjsx"] Nov 22 10:32:59 crc kubenswrapper[4789]: I1122 10:32:59.191422 4789 generic.go:334] "Generic (PLEG): container finished" podID="7c4b4396-9cd6-44ee-b736-744c281dccda" containerID="4c6d1f4a8480c373c9c1346e158c78b5f67b02d87bc337c7f40b1dbf31227302" exitCode=0 Nov 22 10:32:59 crc kubenswrapper[4789]: I1122 10:32:59.191479 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jtjsx" event={"ID":"7c4b4396-9cd6-44ee-b736-744c281dccda","Type":"ContainerDied","Data":"4c6d1f4a8480c373c9c1346e158c78b5f67b02d87bc337c7f40b1dbf31227302"} Nov 22 10:32:59 crc kubenswrapper[4789]: I1122 10:32:59.191506 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jtjsx" event={"ID":"7c4b4396-9cd6-44ee-b736-744c281dccda","Type":"ContainerStarted","Data":"a0fd79cf2401631af87aa8acea3feaf28f5b4b077c58bafb07ebcab9e70f3305"} Nov 22 10:32:59 crc kubenswrapper[4789]: I1122 10:32:59.205093 4789 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 22 10:33:00 crc kubenswrapper[4789]: I1122 10:33:00.209917 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jtjsx" event={"ID":"7c4b4396-9cd6-44ee-b736-744c281dccda","Type":"ContainerStarted","Data":"85011863f594c87e9eeb6fa11672c447349dd5af41c26144a9e863d647ec3736"} Nov 22 10:33:02 crc kubenswrapper[4789]: I1122 10:33:02.234396 4789 generic.go:334] "Generic (PLEG): container finished" podID="7c4b4396-9cd6-44ee-b736-744c281dccda" containerID="85011863f594c87e9eeb6fa11672c447349dd5af41c26144a9e863d647ec3736" exitCode=0 Nov 22 10:33:02 crc kubenswrapper[4789]: I1122 10:33:02.234517 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jtjsx" event={"ID":"7c4b4396-9cd6-44ee-b736-744c281dccda","Type":"ContainerDied","Data":"85011863f594c87e9eeb6fa11672c447349dd5af41c26144a9e863d647ec3736"} Nov 22 10:33:04 crc kubenswrapper[4789]: I1122 10:33:04.274976 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jtjsx" event={"ID":"7c4b4396-9cd6-44ee-b736-744c281dccda","Type":"ContainerStarted","Data":"2c6614aa1434dc93f02fed545089fd1c9825fcf85f662c5a9cbca1d1747a84d2"} Nov 22 10:33:04 crc kubenswrapper[4789]: I1122 10:33:04.333087 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-jtjsx" podStartSLOduration=3.363786974 podStartE2EDuration="7.333051073s" podCreationTimestamp="2025-11-22 10:32:57 +0000 UTC" firstStartedPulling="2025-11-22 10:32:59.204896018 +0000 UTC m=+9393.439296291" lastFinishedPulling="2025-11-22 10:33:03.174160107 +0000 UTC m=+9397.408560390" observedRunningTime="2025-11-22 10:33:04.305235183 +0000 UTC m=+9398.539635476" watchObservedRunningTime="2025-11-22 10:33:04.333051073 +0000 UTC m=+9398.567451376" Nov 22 10:33:04 crc kubenswrapper[4789]: I1122 10:33:04.965237 4789 scope.go:117] "RemoveContainer" containerID="3590e488179d593a3c8981249b3f907f5a30f296b7d9ba57504c20a09c6eb5bc" Nov 22 10:33:04 crc kubenswrapper[4789]: E1122 10:33:04.965642 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 10:33:08 crc kubenswrapper[4789]: I1122 10:33:08.193772 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-jtjsx" Nov 22 10:33:08 crc kubenswrapper[4789]: I1122 10:33:08.194261 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-jtjsx" Nov 22 10:33:09 crc kubenswrapper[4789]: I1122 10:33:09.258513 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-jtjsx" podUID="7c4b4396-9cd6-44ee-b736-744c281dccda" containerName="registry-server" probeResult="failure" output=< Nov 22 10:33:09 crc kubenswrapper[4789]: timeout: failed to connect service ":50051" within 1s Nov 22 10:33:09 crc kubenswrapper[4789]: > Nov 22 10:33:16 crc kubenswrapper[4789]: I1122 10:33:16.964982 4789 scope.go:117] "RemoveContainer" containerID="3590e488179d593a3c8981249b3f907f5a30f296b7d9ba57504c20a09c6eb5bc" Nov 22 10:33:16 crc kubenswrapper[4789]: E1122 10:33:16.965393 4789 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Nov 22 10:33:16 crc kubenswrapper[4789]: E1122 10:33:16.966036 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 10:33:18 crc kubenswrapper[4789]: I1122 10:33:18.261177 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-jtjsx" Nov 22 10:33:18 crc kubenswrapper[4789]: I1122 10:33:18.326414 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-jtjsx" Nov 22 10:33:18 crc kubenswrapper[4789]: I1122 10:33:18.498699 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jtjsx"] Nov 22 10:33:19 crc kubenswrapper[4789]: I1122 10:33:19.423437 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-jtjsx" podUID="7c4b4396-9cd6-44ee-b736-744c281dccda" containerName="registry-server" containerID="cri-o://2c6614aa1434dc93f02fed545089fd1c9825fcf85f662c5a9cbca1d1747a84d2" gracePeriod=2 Nov 22 10:33:19 crc kubenswrapper[4789]: I1122 10:33:19.893507 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jtjsx" Nov 22 10:33:19 crc kubenswrapper[4789]: I1122 10:33:19.935980 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tmqr9\" (UniqueName: \"kubernetes.io/projected/7c4b4396-9cd6-44ee-b736-744c281dccda-kube-api-access-tmqr9\") pod \"7c4b4396-9cd6-44ee-b736-744c281dccda\" (UID: \"7c4b4396-9cd6-44ee-b736-744c281dccda\") " Nov 22 10:33:19 crc kubenswrapper[4789]: I1122 10:33:19.936048 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c4b4396-9cd6-44ee-b736-744c281dccda-utilities\") pod \"7c4b4396-9cd6-44ee-b736-744c281dccda\" (UID: \"7c4b4396-9cd6-44ee-b736-744c281dccda\") " Nov 22 10:33:19 crc kubenswrapper[4789]: I1122 10:33:19.936246 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c4b4396-9cd6-44ee-b736-744c281dccda-catalog-content\") pod \"7c4b4396-9cd6-44ee-b736-744c281dccda\" (UID: \"7c4b4396-9cd6-44ee-b736-744c281dccda\") " Nov 22 10:33:19 crc kubenswrapper[4789]: I1122 10:33:19.937314 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7c4b4396-9cd6-44ee-b736-744c281dccda-utilities" (OuterVolumeSpecName: "utilities") pod "7c4b4396-9cd6-44ee-b736-744c281dccda" (UID: "7c4b4396-9cd6-44ee-b736-744c281dccda"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 10:33:19 crc kubenswrapper[4789]: I1122 10:33:19.949959 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c4b4396-9cd6-44ee-b736-744c281dccda-kube-api-access-tmqr9" (OuterVolumeSpecName: "kube-api-access-tmqr9") pod "7c4b4396-9cd6-44ee-b736-744c281dccda" (UID: "7c4b4396-9cd6-44ee-b736-744c281dccda"). InnerVolumeSpecName "kube-api-access-tmqr9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 10:33:20 crc kubenswrapper[4789]: I1122 10:33:20.034230 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7c4b4396-9cd6-44ee-b736-744c281dccda-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7c4b4396-9cd6-44ee-b736-744c281dccda" (UID: "7c4b4396-9cd6-44ee-b736-744c281dccda"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 10:33:20 crc kubenswrapper[4789]: I1122 10:33:20.038736 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tmqr9\" (UniqueName: \"kubernetes.io/projected/7c4b4396-9cd6-44ee-b736-744c281dccda-kube-api-access-tmqr9\") on node \"crc\" DevicePath \"\"" Nov 22 10:33:20 crc kubenswrapper[4789]: I1122 10:33:20.038779 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c4b4396-9cd6-44ee-b736-744c281dccda-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 10:33:20 crc kubenswrapper[4789]: I1122 10:33:20.038792 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c4b4396-9cd6-44ee-b736-744c281dccda-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 10:33:20 crc kubenswrapper[4789]: I1122 10:33:20.441904 4789 generic.go:334] "Generic (PLEG): container finished" podID="7c4b4396-9cd6-44ee-b736-744c281dccda" containerID="2c6614aa1434dc93f02fed545089fd1c9825fcf85f662c5a9cbca1d1747a84d2" exitCode=0 Nov 22 10:33:20 crc kubenswrapper[4789]: I1122 10:33:20.441949 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jtjsx" event={"ID":"7c4b4396-9cd6-44ee-b736-744c281dccda","Type":"ContainerDied","Data":"2c6614aa1434dc93f02fed545089fd1c9825fcf85f662c5a9cbca1d1747a84d2"} Nov 22 10:33:20 crc kubenswrapper[4789]: I1122 10:33:20.441977 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jtjsx" event={"ID":"7c4b4396-9cd6-44ee-b736-744c281dccda","Type":"ContainerDied","Data":"a0fd79cf2401631af87aa8acea3feaf28f5b4b077c58bafb07ebcab9e70f3305"} Nov 22 10:33:20 crc kubenswrapper[4789]: I1122 10:33:20.442017 4789 scope.go:117] "RemoveContainer" containerID="2c6614aa1434dc93f02fed545089fd1c9825fcf85f662c5a9cbca1d1747a84d2" Nov 22 10:33:20 crc kubenswrapper[4789]: I1122 10:33:20.442148 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jtjsx" Nov 22 10:33:20 crc kubenswrapper[4789]: I1122 10:33:20.480874 4789 scope.go:117] "RemoveContainer" containerID="85011863f594c87e9eeb6fa11672c447349dd5af41c26144a9e863d647ec3736" Nov 22 10:33:20 crc kubenswrapper[4789]: I1122 10:33:20.481077 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jtjsx"] Nov 22 10:33:20 crc kubenswrapper[4789]: I1122 10:33:20.493071 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-jtjsx"] Nov 22 10:33:20 crc kubenswrapper[4789]: I1122 10:33:20.506301 4789 scope.go:117] "RemoveContainer" containerID="4c6d1f4a8480c373c9c1346e158c78b5f67b02d87bc337c7f40b1dbf31227302" Nov 22 10:33:20 crc kubenswrapper[4789]: I1122 10:33:20.553604 4789 scope.go:117] "RemoveContainer" containerID="2c6614aa1434dc93f02fed545089fd1c9825fcf85f662c5a9cbca1d1747a84d2" Nov 22 10:33:20 crc kubenswrapper[4789]: E1122 10:33:20.554390 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2c6614aa1434dc93f02fed545089fd1c9825fcf85f662c5a9cbca1d1747a84d2\": container with ID starting with 2c6614aa1434dc93f02fed545089fd1c9825fcf85f662c5a9cbca1d1747a84d2 not found: ID does not exist" containerID="2c6614aa1434dc93f02fed545089fd1c9825fcf85f662c5a9cbca1d1747a84d2" Nov 22 10:33:20 crc kubenswrapper[4789]: I1122 10:33:20.554449 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2c6614aa1434dc93f02fed545089fd1c9825fcf85f662c5a9cbca1d1747a84d2"} err="failed to get container status \"2c6614aa1434dc93f02fed545089fd1c9825fcf85f662c5a9cbca1d1747a84d2\": rpc error: code = NotFound desc = could not find container \"2c6614aa1434dc93f02fed545089fd1c9825fcf85f662c5a9cbca1d1747a84d2\": container with ID starting with 2c6614aa1434dc93f02fed545089fd1c9825fcf85f662c5a9cbca1d1747a84d2 not found: ID does not exist" Nov 22 10:33:20 crc kubenswrapper[4789]: I1122 10:33:20.554486 4789 scope.go:117] "RemoveContainer" containerID="85011863f594c87e9eeb6fa11672c447349dd5af41c26144a9e863d647ec3736" Nov 22 10:33:20 crc kubenswrapper[4789]: E1122 10:33:20.556065 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"85011863f594c87e9eeb6fa11672c447349dd5af41c26144a9e863d647ec3736\": container with ID starting with 85011863f594c87e9eeb6fa11672c447349dd5af41c26144a9e863d647ec3736 not found: ID does not exist" containerID="85011863f594c87e9eeb6fa11672c447349dd5af41c26144a9e863d647ec3736" Nov 22 10:33:20 crc kubenswrapper[4789]: I1122 10:33:20.556099 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"85011863f594c87e9eeb6fa11672c447349dd5af41c26144a9e863d647ec3736"} err="failed to get container status \"85011863f594c87e9eeb6fa11672c447349dd5af41c26144a9e863d647ec3736\": rpc error: code = NotFound desc = could not find container \"85011863f594c87e9eeb6fa11672c447349dd5af41c26144a9e863d647ec3736\": container with ID starting with 85011863f594c87e9eeb6fa11672c447349dd5af41c26144a9e863d647ec3736 not found: ID does not exist" Nov 22 10:33:20 crc kubenswrapper[4789]: I1122 10:33:20.556158 4789 scope.go:117] "RemoveContainer" containerID="4c6d1f4a8480c373c9c1346e158c78b5f67b02d87bc337c7f40b1dbf31227302" Nov 22 10:33:20 crc kubenswrapper[4789]: E1122 10:33:20.556472 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4c6d1f4a8480c373c9c1346e158c78b5f67b02d87bc337c7f40b1dbf31227302\": container with ID starting with 4c6d1f4a8480c373c9c1346e158c78b5f67b02d87bc337c7f40b1dbf31227302 not found: ID does not exist" containerID="4c6d1f4a8480c373c9c1346e158c78b5f67b02d87bc337c7f40b1dbf31227302" Nov 22 10:33:20 crc kubenswrapper[4789]: I1122 10:33:20.556503 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4c6d1f4a8480c373c9c1346e158c78b5f67b02d87bc337c7f40b1dbf31227302"} err="failed to get container status \"4c6d1f4a8480c373c9c1346e158c78b5f67b02d87bc337c7f40b1dbf31227302\": rpc error: code = NotFound desc = could not find container \"4c6d1f4a8480c373c9c1346e158c78b5f67b02d87bc337c7f40b1dbf31227302\": container with ID starting with 4c6d1f4a8480c373c9c1346e158c78b5f67b02d87bc337c7f40b1dbf31227302 not found: ID does not exist" Nov 22 10:33:21 crc kubenswrapper[4789]: I1122 10:33:21.988852 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7c4b4396-9cd6-44ee-b736-744c281dccda" path="/var/lib/kubelet/pods/7c4b4396-9cd6-44ee-b736-744c281dccda/volumes" Nov 22 10:33:30 crc kubenswrapper[4789]: I1122 10:33:30.965967 4789 scope.go:117] "RemoveContainer" containerID="3590e488179d593a3c8981249b3f907f5a30f296b7d9ba57504c20a09c6eb5bc" Nov 22 10:33:30 crc kubenswrapper[4789]: E1122 10:33:30.966875 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 10:33:44 crc kubenswrapper[4789]: I1122 10:33:44.966087 4789 scope.go:117] "RemoveContainer" containerID="3590e488179d593a3c8981249b3f907f5a30f296b7d9ba57504c20a09c6eb5bc" Nov 22 10:33:44 crc kubenswrapper[4789]: E1122 10:33:44.966979 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 10:33:52 crc kubenswrapper[4789]: I1122 10:33:52.838934 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-496fm"] Nov 22 10:33:52 crc kubenswrapper[4789]: E1122 10:33:52.840423 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c4b4396-9cd6-44ee-b736-744c281dccda" containerName="extract-utilities" Nov 22 10:33:52 crc kubenswrapper[4789]: I1122 10:33:52.840455 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c4b4396-9cd6-44ee-b736-744c281dccda" containerName="extract-utilities" Nov 22 10:33:52 crc kubenswrapper[4789]: E1122 10:33:52.840489 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c4b4396-9cd6-44ee-b736-744c281dccda" containerName="extract-content" Nov 22 10:33:52 crc kubenswrapper[4789]: I1122 10:33:52.840509 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c4b4396-9cd6-44ee-b736-744c281dccda" containerName="extract-content" Nov 22 10:33:52 crc kubenswrapper[4789]: E1122 10:33:52.840554 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c4b4396-9cd6-44ee-b736-744c281dccda" containerName="registry-server" Nov 22 10:33:52 crc kubenswrapper[4789]: I1122 10:33:52.840574 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c4b4396-9cd6-44ee-b736-744c281dccda" containerName="registry-server" Nov 22 10:33:52 crc kubenswrapper[4789]: I1122 10:33:52.841043 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c4b4396-9cd6-44ee-b736-744c281dccda" containerName="registry-server" Nov 22 10:33:52 crc kubenswrapper[4789]: I1122 10:33:52.843708 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-496fm" Nov 22 10:33:52 crc kubenswrapper[4789]: I1122 10:33:52.855712 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-496fm"] Nov 22 10:33:52 crc kubenswrapper[4789]: I1122 10:33:52.904375 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hwrpz\" (UniqueName: \"kubernetes.io/projected/cf7918b8-b504-4596-a5b1-8fa1fba433b1-kube-api-access-hwrpz\") pod \"certified-operators-496fm\" (UID: \"cf7918b8-b504-4596-a5b1-8fa1fba433b1\") " pod="openshift-marketplace/certified-operators-496fm" Nov 22 10:33:52 crc kubenswrapper[4789]: I1122 10:33:52.904432 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cf7918b8-b504-4596-a5b1-8fa1fba433b1-utilities\") pod \"certified-operators-496fm\" (UID: \"cf7918b8-b504-4596-a5b1-8fa1fba433b1\") " pod="openshift-marketplace/certified-operators-496fm" Nov 22 10:33:52 crc kubenswrapper[4789]: I1122 10:33:52.904633 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cf7918b8-b504-4596-a5b1-8fa1fba433b1-catalog-content\") pod \"certified-operators-496fm\" (UID: \"cf7918b8-b504-4596-a5b1-8fa1fba433b1\") " pod="openshift-marketplace/certified-operators-496fm" Nov 22 10:33:53 crc kubenswrapper[4789]: I1122 10:33:53.007086 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hwrpz\" (UniqueName: \"kubernetes.io/projected/cf7918b8-b504-4596-a5b1-8fa1fba433b1-kube-api-access-hwrpz\") pod \"certified-operators-496fm\" (UID: \"cf7918b8-b504-4596-a5b1-8fa1fba433b1\") " pod="openshift-marketplace/certified-operators-496fm" Nov 22 10:33:53 crc kubenswrapper[4789]: I1122 10:33:53.007146 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cf7918b8-b504-4596-a5b1-8fa1fba433b1-utilities\") pod \"certified-operators-496fm\" (UID: \"cf7918b8-b504-4596-a5b1-8fa1fba433b1\") " pod="openshift-marketplace/certified-operators-496fm" Nov 22 10:33:53 crc kubenswrapper[4789]: I1122 10:33:53.007236 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cf7918b8-b504-4596-a5b1-8fa1fba433b1-catalog-content\") pod \"certified-operators-496fm\" (UID: \"cf7918b8-b504-4596-a5b1-8fa1fba433b1\") " pod="openshift-marketplace/certified-operators-496fm" Nov 22 10:33:53 crc kubenswrapper[4789]: I1122 10:33:53.008071 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cf7918b8-b504-4596-a5b1-8fa1fba433b1-catalog-content\") pod \"certified-operators-496fm\" (UID: \"cf7918b8-b504-4596-a5b1-8fa1fba433b1\") " pod="openshift-marketplace/certified-operators-496fm" Nov 22 10:33:53 crc kubenswrapper[4789]: I1122 10:33:53.008301 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cf7918b8-b504-4596-a5b1-8fa1fba433b1-utilities\") pod \"certified-operators-496fm\" (UID: \"cf7918b8-b504-4596-a5b1-8fa1fba433b1\") " pod="openshift-marketplace/certified-operators-496fm" Nov 22 10:33:53 crc kubenswrapper[4789]: I1122 10:33:53.033324 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hwrpz\" (UniqueName: \"kubernetes.io/projected/cf7918b8-b504-4596-a5b1-8fa1fba433b1-kube-api-access-hwrpz\") pod \"certified-operators-496fm\" (UID: \"cf7918b8-b504-4596-a5b1-8fa1fba433b1\") " pod="openshift-marketplace/certified-operators-496fm" Nov 22 10:33:53 crc kubenswrapper[4789]: I1122 10:33:53.172221 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-496fm" Nov 22 10:33:53 crc kubenswrapper[4789]: I1122 10:33:53.746666 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-496fm"] Nov 22 10:33:54 crc kubenswrapper[4789]: I1122 10:33:54.788952 4789 generic.go:334] "Generic (PLEG): container finished" podID="cf7918b8-b504-4596-a5b1-8fa1fba433b1" containerID="a8070ace2793e69cc310de182b1c76732376232ea208d6d2bda09b769d4b0c24" exitCode=0 Nov 22 10:33:54 crc kubenswrapper[4789]: I1122 10:33:54.789017 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-496fm" event={"ID":"cf7918b8-b504-4596-a5b1-8fa1fba433b1","Type":"ContainerDied","Data":"a8070ace2793e69cc310de182b1c76732376232ea208d6d2bda09b769d4b0c24"} Nov 22 10:33:54 crc kubenswrapper[4789]: I1122 10:33:54.789318 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-496fm" event={"ID":"cf7918b8-b504-4596-a5b1-8fa1fba433b1","Type":"ContainerStarted","Data":"8d0717b6677f3496ac060cdd1c050988a28ba6d171aff307505cf8f85ef05426"} Nov 22 10:33:55 crc kubenswrapper[4789]: I1122 10:33:55.965279 4789 scope.go:117] "RemoveContainer" containerID="3590e488179d593a3c8981249b3f907f5a30f296b7d9ba57504c20a09c6eb5bc" Nov 22 10:33:55 crc kubenswrapper[4789]: E1122 10:33:55.966399 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" Nov 22 10:33:56 crc kubenswrapper[4789]: I1122 10:33:56.817290 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-496fm" event={"ID":"cf7918b8-b504-4596-a5b1-8fa1fba433b1","Type":"ContainerStarted","Data":"a8638c77381181df74721ed5b2c417b5da6ae401335c2d38422d7b544fdaf9ce"} Nov 22 10:33:57 crc kubenswrapper[4789]: I1122 10:33:57.837671 4789 generic.go:334] "Generic (PLEG): container finished" podID="cf7918b8-b504-4596-a5b1-8fa1fba433b1" containerID="a8638c77381181df74721ed5b2c417b5da6ae401335c2d38422d7b544fdaf9ce" exitCode=0 Nov 22 10:33:57 crc kubenswrapper[4789]: I1122 10:33:57.837793 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-496fm" event={"ID":"cf7918b8-b504-4596-a5b1-8fa1fba433b1","Type":"ContainerDied","Data":"a8638c77381181df74721ed5b2c417b5da6ae401335c2d38422d7b544fdaf9ce"} Nov 22 10:33:58 crc kubenswrapper[4789]: I1122 10:33:58.851171 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-496fm" event={"ID":"cf7918b8-b504-4596-a5b1-8fa1fba433b1","Type":"ContainerStarted","Data":"6572a8d115cd0bae17be75ffe879fc9d55d93520166f1a05c6ae0bb5a5cd68ab"} Nov 22 10:33:58 crc kubenswrapper[4789]: I1122 10:33:58.892574 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-496fm" podStartSLOduration=3.43889676 podStartE2EDuration="6.892541827s" podCreationTimestamp="2025-11-22 10:33:52 +0000 UTC" firstStartedPulling="2025-11-22 10:33:54.792510024 +0000 UTC m=+9449.026910337" lastFinishedPulling="2025-11-22 10:33:58.246155091 +0000 UTC m=+9452.480555404" observedRunningTime="2025-11-22 10:33:58.87299302 +0000 UTC m=+9453.107393313" watchObservedRunningTime="2025-11-22 10:33:58.892541827 +0000 UTC m=+9453.126942130" Nov 22 10:34:03 crc kubenswrapper[4789]: I1122 10:34:03.172458 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-496fm" Nov 22 10:34:03 crc kubenswrapper[4789]: I1122 10:34:03.175556 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-496fm" Nov 22 10:34:03 crc kubenswrapper[4789]: I1122 10:34:03.269927 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-496fm" Nov 22 10:34:03 crc kubenswrapper[4789]: I1122 10:34:03.991944 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-496fm" Nov 22 10:34:04 crc kubenswrapper[4789]: I1122 10:34:04.064256 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-496fm"] Nov 22 10:34:05 crc kubenswrapper[4789]: I1122 10:34:05.936490 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-496fm" podUID="cf7918b8-b504-4596-a5b1-8fa1fba433b1" containerName="registry-server" containerID="cri-o://6572a8d115cd0bae17be75ffe879fc9d55d93520166f1a05c6ae0bb5a5cd68ab" gracePeriod=2 Nov 22 10:34:06 crc kubenswrapper[4789]: I1122 10:34:06.470096 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-496fm" Nov 22 10:34:06 crc kubenswrapper[4789]: I1122 10:34:06.484151 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cf7918b8-b504-4596-a5b1-8fa1fba433b1-utilities\") pod \"cf7918b8-b504-4596-a5b1-8fa1fba433b1\" (UID: \"cf7918b8-b504-4596-a5b1-8fa1fba433b1\") " Nov 22 10:34:06 crc kubenswrapper[4789]: I1122 10:34:06.484320 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hwrpz\" (UniqueName: \"kubernetes.io/projected/cf7918b8-b504-4596-a5b1-8fa1fba433b1-kube-api-access-hwrpz\") pod \"cf7918b8-b504-4596-a5b1-8fa1fba433b1\" (UID: \"cf7918b8-b504-4596-a5b1-8fa1fba433b1\") " Nov 22 10:34:06 crc kubenswrapper[4789]: I1122 10:34:06.484473 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cf7918b8-b504-4596-a5b1-8fa1fba433b1-catalog-content\") pod \"cf7918b8-b504-4596-a5b1-8fa1fba433b1\" (UID: \"cf7918b8-b504-4596-a5b1-8fa1fba433b1\") " Nov 22 10:34:06 crc kubenswrapper[4789]: I1122 10:34:06.489457 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cf7918b8-b504-4596-a5b1-8fa1fba433b1-utilities" (OuterVolumeSpecName: "utilities") pod "cf7918b8-b504-4596-a5b1-8fa1fba433b1" (UID: "cf7918b8-b504-4596-a5b1-8fa1fba433b1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 10:34:06 crc kubenswrapper[4789]: I1122 10:34:06.491943 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf7918b8-b504-4596-a5b1-8fa1fba433b1-kube-api-access-hwrpz" (OuterVolumeSpecName: "kube-api-access-hwrpz") pod "cf7918b8-b504-4596-a5b1-8fa1fba433b1" (UID: "cf7918b8-b504-4596-a5b1-8fa1fba433b1"). InnerVolumeSpecName "kube-api-access-hwrpz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 10:34:06 crc kubenswrapper[4789]: I1122 10:34:06.548596 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cf7918b8-b504-4596-a5b1-8fa1fba433b1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cf7918b8-b504-4596-a5b1-8fa1fba433b1" (UID: "cf7918b8-b504-4596-a5b1-8fa1fba433b1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 10:34:06 crc kubenswrapper[4789]: I1122 10:34:06.587198 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hwrpz\" (UniqueName: \"kubernetes.io/projected/cf7918b8-b504-4596-a5b1-8fa1fba433b1-kube-api-access-hwrpz\") on node \"crc\" DevicePath \"\"" Nov 22 10:34:06 crc kubenswrapper[4789]: I1122 10:34:06.587250 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cf7918b8-b504-4596-a5b1-8fa1fba433b1-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 10:34:06 crc kubenswrapper[4789]: I1122 10:34:06.587269 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cf7918b8-b504-4596-a5b1-8fa1fba433b1-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 10:34:06 crc kubenswrapper[4789]: I1122 10:34:06.953048 4789 generic.go:334] "Generic (PLEG): container finished" podID="cf7918b8-b504-4596-a5b1-8fa1fba433b1" containerID="6572a8d115cd0bae17be75ffe879fc9d55d93520166f1a05c6ae0bb5a5cd68ab" exitCode=0 Nov 22 10:34:06 crc kubenswrapper[4789]: I1122 10:34:06.953141 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-496fm" Nov 22 10:34:06 crc kubenswrapper[4789]: I1122 10:34:06.953140 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-496fm" event={"ID":"cf7918b8-b504-4596-a5b1-8fa1fba433b1","Type":"ContainerDied","Data":"6572a8d115cd0bae17be75ffe879fc9d55d93520166f1a05c6ae0bb5a5cd68ab"} Nov 22 10:34:06 crc kubenswrapper[4789]: I1122 10:34:06.953578 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-496fm" event={"ID":"cf7918b8-b504-4596-a5b1-8fa1fba433b1","Type":"ContainerDied","Data":"8d0717b6677f3496ac060cdd1c050988a28ba6d171aff307505cf8f85ef05426"} Nov 22 10:34:06 crc kubenswrapper[4789]: I1122 10:34:06.953643 4789 scope.go:117] "RemoveContainer" containerID="6572a8d115cd0bae17be75ffe879fc9d55d93520166f1a05c6ae0bb5a5cd68ab" Nov 22 10:34:06 crc kubenswrapper[4789]: I1122 10:34:06.987248 4789 scope.go:117] "RemoveContainer" containerID="a8638c77381181df74721ed5b2c417b5da6ae401335c2d38422d7b544fdaf9ce" Nov 22 10:34:07 crc kubenswrapper[4789]: I1122 10:34:07.004519 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-496fm"] Nov 22 10:34:07 crc kubenswrapper[4789]: I1122 10:34:07.013784 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-496fm"] Nov 22 10:34:07 crc kubenswrapper[4789]: I1122 10:34:07.020812 4789 scope.go:117] "RemoveContainer" containerID="a8070ace2793e69cc310de182b1c76732376232ea208d6d2bda09b769d4b0c24" Nov 22 10:34:07 crc kubenswrapper[4789]: I1122 10:34:07.064348 4789 scope.go:117] "RemoveContainer" containerID="6572a8d115cd0bae17be75ffe879fc9d55d93520166f1a05c6ae0bb5a5cd68ab" Nov 22 10:34:07 crc kubenswrapper[4789]: E1122 10:34:07.064981 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6572a8d115cd0bae17be75ffe879fc9d55d93520166f1a05c6ae0bb5a5cd68ab\": container with ID starting with 6572a8d115cd0bae17be75ffe879fc9d55d93520166f1a05c6ae0bb5a5cd68ab not found: ID does not exist" containerID="6572a8d115cd0bae17be75ffe879fc9d55d93520166f1a05c6ae0bb5a5cd68ab" Nov 22 10:34:07 crc kubenswrapper[4789]: I1122 10:34:07.065102 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6572a8d115cd0bae17be75ffe879fc9d55d93520166f1a05c6ae0bb5a5cd68ab"} err="failed to get container status \"6572a8d115cd0bae17be75ffe879fc9d55d93520166f1a05c6ae0bb5a5cd68ab\": rpc error: code = NotFound desc = could not find container \"6572a8d115cd0bae17be75ffe879fc9d55d93520166f1a05c6ae0bb5a5cd68ab\": container with ID starting with 6572a8d115cd0bae17be75ffe879fc9d55d93520166f1a05c6ae0bb5a5cd68ab not found: ID does not exist" Nov 22 10:34:07 crc kubenswrapper[4789]: I1122 10:34:07.065193 4789 scope.go:117] "RemoveContainer" containerID="a8638c77381181df74721ed5b2c417b5da6ae401335c2d38422d7b544fdaf9ce" Nov 22 10:34:07 crc kubenswrapper[4789]: E1122 10:34:07.065667 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a8638c77381181df74721ed5b2c417b5da6ae401335c2d38422d7b544fdaf9ce\": container with ID starting with a8638c77381181df74721ed5b2c417b5da6ae401335c2d38422d7b544fdaf9ce not found: ID does not exist" containerID="a8638c77381181df74721ed5b2c417b5da6ae401335c2d38422d7b544fdaf9ce" Nov 22 10:34:07 crc kubenswrapper[4789]: I1122 10:34:07.065706 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8638c77381181df74721ed5b2c417b5da6ae401335c2d38422d7b544fdaf9ce"} err="failed to get container status \"a8638c77381181df74721ed5b2c417b5da6ae401335c2d38422d7b544fdaf9ce\": rpc error: code = NotFound desc = could not find container \"a8638c77381181df74721ed5b2c417b5da6ae401335c2d38422d7b544fdaf9ce\": container with ID starting with a8638c77381181df74721ed5b2c417b5da6ae401335c2d38422d7b544fdaf9ce not found: ID does not exist" Nov 22 10:34:07 crc kubenswrapper[4789]: I1122 10:34:07.065732 4789 scope.go:117] "RemoveContainer" containerID="a8070ace2793e69cc310de182b1c76732376232ea208d6d2bda09b769d4b0c24" Nov 22 10:34:07 crc kubenswrapper[4789]: E1122 10:34:07.066063 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a8070ace2793e69cc310de182b1c76732376232ea208d6d2bda09b769d4b0c24\": container with ID starting with a8070ace2793e69cc310de182b1c76732376232ea208d6d2bda09b769d4b0c24 not found: ID does not exist" containerID="a8070ace2793e69cc310de182b1c76732376232ea208d6d2bda09b769d4b0c24" Nov 22 10:34:07 crc kubenswrapper[4789]: I1122 10:34:07.066114 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8070ace2793e69cc310de182b1c76732376232ea208d6d2bda09b769d4b0c24"} err="failed to get container status \"a8070ace2793e69cc310de182b1c76732376232ea208d6d2bda09b769d4b0c24\": rpc error: code = NotFound desc = could not find container \"a8070ace2793e69cc310de182b1c76732376232ea208d6d2bda09b769d4b0c24\": container with ID starting with a8070ace2793e69cc310de182b1c76732376232ea208d6d2bda09b769d4b0c24 not found: ID does not exist" Nov 22 10:34:07 crc kubenswrapper[4789]: I1122 10:34:07.977321 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cf7918b8-b504-4596-a5b1-8fa1fba433b1" path="/var/lib/kubelet/pods/cf7918b8-b504-4596-a5b1-8fa1fba433b1/volumes" Nov 22 10:34:10 crc kubenswrapper[4789]: I1122 10:34:10.966270 4789 scope.go:117] "RemoveContainer" containerID="3590e488179d593a3c8981249b3f907f5a30f296b7d9ba57504c20a09c6eb5bc" Nov 22 10:34:10 crc kubenswrapper[4789]: E1122 10:34:10.967612 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-pxvrm_openshift-machine-config-operator(58aa8071-7ecc-4692-8789-2db04ec70647)\"" pod="openshift-machine-config-operator/machine-config-daemon-pxvrm" podUID="58aa8071-7ecc-4692-8789-2db04ec70647" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515110310653024440 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015110310653017355 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015110265675016515 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015110265675015465 5ustar corecore